热点
"量化" 相关文章
微软1bit LLM新研究:原生4bit激活值量化,可充分利用新一代GPU对4bit计算的原生支持
智源社区 2025-04-30T12:44:24.000000Z
LLMs Can Now Retain High Accuracy at 2-Bit Precision: Researchers from UNC Chapel Hill Introduce TACQ, a Task-Aware Quantization Approach that Preserves Critical Weight Circuits for Compression Without Performance Loss
MarkTechPost@AI 2025-04-22T07:05:37.000000Z
Gemma 3支援QAT技術,消費級GPU就可執行270億參數大模型
AI & Big Data 2025-04-21T06:57:58.000000Z
25.4.17 赚了一天
漫漫游资路 2025-04-19T06:53:54.000000Z
不惧关税事件扰动,私募股票仓位创七个月新高,股票策略一季度领跑
深度财经头条 2025-04-13T10:45:45.000000Z
10倍吞吐提升无损性能:多模态适用的KV cache量化策略来了,即插即用无需改原模型
PaperWeekly 2025-04-10T13:47:21.000000Z
咱QwQ 32B就够了?Llama4还要2000B?!
算力百科 2025-04-09T11:12:39.000000Z
这个30岁后的顶级混世智慧,会得罪人
36kr 2025-04-09T08:33:46.000000Z
2月私募登记与产品备案环比减少,百亿量化私募备案更显积极
深度财经头条 2025-03-21T01:30:24.000000Z
热情再起?1月备案私募证券产品超700只,量化占近四成,宽德等百亿量化备案积极
深度财经头条 2025-02-20T13:45:02.000000Z
Meta AI Introduces ParetoQ: A Unified Machine Learning Framework for Sub-4-Bit Quantization in Large Language Models
MarkTechPost@AI 2025-02-09T00:20:11.000000Z
完整的671B R1塞进本地,详尽教程来了!
智源社区 2025-02-05T04:07:16.000000Z
Quantization Space Utilization Rate (QSUR): A Novel Post-Training Quantization Method Designed to Enhance the Efficiency of Large Language Models (LLMs)
MarkTechPost@AI 2025-01-30T06:19:40.000000Z
本地大模型之路(二):了解模型能力与性能需求,让硬件选购恰到好处
少数派-AI 2025-01-03T07:00:04.000000Z
4比特量化三倍加速不掉点!清华即插即用的SageAttention迎来升级
机器之心 2024-12-27T08:09:02.000000Z
A popular technique to make AI more efficient has drawbacks
TechCrunch News 2024-12-23T18:07:31.000000Z
苦寻绝对收益的私募也快撑不下去了
36kr-科技 2024-12-18T08:03:20.000000Z
微盘大跌,量化再成矛头指向,真相如何?
36kr 2024-12-18T00:46:16.000000Z
1-bit大模型还能再突破,新一代BitNet架构启用4位激活值
36氪 - 科技频道 2024-12-05T05:10:06.000000Z
社区供稿 | 探索 Ovis: 多模态大模型量化的实战指南
Hugging Face 2024-11-20T11:48:45.000000Z