热点
关于我们
xx
xx
"
模型效率
" 相关文章
A Technique of Pure Reason
少点错误
2025-06-04T19:17:31.000000Z
This AI paper from DeepSeek-AI Explores How DeepSeek-V3 Delivers High-Performance Language Modeling by Minimizing Hardware Overhead and Maximizing Computational Efficiency
MarkTechPost@AI
2025-05-17T06:35:43.000000Z
斯坦福 AI 指数报告发布:2024 年美国 AI 私人投资约为中国的 12 倍、GPT-3.5级别模型推理成本下降 280 倍
AI科技评论
2025-04-09T10:19:09.000000Z
Decoupling Tokenization: How Over-Tokenized Transformers Redefine Vocabulary Scaling in Language Models
MarkTechPost@AI
2025-01-30T17:46:58.000000Z
Autonomy-of-Experts (AoE): A Router-Free Paradigm for Efficient and Adaptive Mixture-of-Experts Models
MarkTechPost@AI
2025-01-27T01:04:01.000000Z
面壁、智谱,卷上新赛道
极客公园官网
2025-01-19T05:06:39.000000Z
Apple Researchers Introduce Instruction-Following Pruning (IFPruning): A Dynamic AI Approach to Efficient and Scalable LLM Optimization
MarkTechPost@AI
2025-01-14T02:42:50.000000Z
田渊栋:2024年年终总结
新智元
2025-01-02T06:31:00.000000Z
1-bit大模型还能再突破!新一代BitNet架构启用4位激活值
新智元
2024-12-05T05:26:26.000000Z
【NLP】Kaggle知识点:文本分类与LoRA
机器学习初学者
2024-12-04T05:36:18.000000Z
清华UCSD提出全新微调方法,8B小模型媲美GPT-4o!科学问题正确率提高28%
智源社区
2024-12-03T09:05:40.000000Z
Google AI Introduces LAuReL (Learned Augmented Residual Layer): Revolutionizing Neural Networks with Enhanced Residual Connections for Efficient Model Performance
MarkTechPost@AI
2024-11-17T07:50:15.000000Z
Heterogeneous Mixture of Experts (HMoE): Enhancing Model Efficiency and Performance with Diverse Expert Capacities
MarkTechPost@AI
2024-08-25T03:49:48.000000Z
成本降低90%!Claude上新Gemini同款提示词缓存,一次性记住整个代码库
智源社区
2024-08-19T04:22:42.000000Z
This AI Paper by Meta FAIR Introduces MoMa: A Modality-Aware Mixture-of-Experts Architecture for Efficient Multimodal Pre-training
MarkTechPost@AI
2024-08-04T02:49:37.000000Z