热点
"LLM压缩" 相关文章
From 2:4 to 8:16 sparsity patterns in LLMs for Outliers and Weights with Variance Correction
cs.AI updates on arXiv.org 2025-07-08T05:54:02.000000Z
70% 大小,100% 准确!完美压缩 LLM 性能 0 损失,推理速度最高飙升 39 倍
掘金 人工智能 2025-04-27T06:22:53.000000Z
SeedLM: A Post-Training Compression Method that Uses Pseudo-Random Generators to Efficiently Encode and Compress LLM Weights
MarkTechPost@AI 2024-10-16T03:51:07.000000Z