热点
"训练效率" 相关文章
人工智能的下个浪潮是物理AI
韭研公社 2025-07-21T02:40:52.000000Z
回应撤离中国市场原因,Manus首度披露技术侧经验教训
36氪 2025-07-20T02:20:52.000000Z
BootSeer: Analyzing and Mitigating Initialization Bottlenecks in Large-Scale LLM Training
cs.AI updates on arXiv.org 2025-07-18T04:13:52.000000Z
Learning to Reason at the Frontier of Learnability
cs.AI updates on arXiv.org 2025-07-17T04:14:46.000000Z
Inversion-DPO: Precise and Efficient Post-Training for Diffusion Models
cs.AI updates on arXiv.org 2025-07-17T04:14:17.000000Z
GHPO: Adaptive Guidance for Stable and Efficient LLM Reinforcement Learning
cs.AI updates on arXiv.org 2025-07-16T04:28:52.000000Z
DuetGraph: Coarse-to-Fine Knowledge Graph Reasoning with Dual-Pathway Global-Local Fusion
cs.AI updates on arXiv.org 2025-07-16T04:28:38.000000Z
MENTOR: Efficient Multimodal-Conditioned Tuning for Autoregressive Vision Generation Models
cs.AI updates on arXiv.org 2025-07-15T04:26:49.000000Z
Gradients as an Action: Towards Communication-Efficient Federated Recommender Systems via Adaptive Action Sharing
cs.AI updates on arXiv.org 2025-07-15T04:24:23.000000Z
LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization
cs.AI updates on arXiv.org 2025-07-08T05:54:02.000000Z
Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models
cs.AI updates on arXiv.org 2025-07-04T04:08:40.000000Z
万卡集群真实部署,已节省数百万 GPU 小时!MoE 通信优化技术 COMET 开源
字节跳动技术团队 2025-03-25T12:00:59.000000Z
Moonshot AI and UCLA Researchers Release Moonlight: A 3B/16B-Parameter Mixture-of-Expert (MoE) Model Trained with 5.7T Tokens Using Muon Optimizer
MarkTechPost@AI 2025-02-23T04:50:15.000000Z
出人意料!DeepSeek-R1用的GRPO其实非最优?规模化强化学习训练用PPO就够了
机器之心 2025-02-21T05:49:07.000000Z
DeepSeek新论文再次引发热议,它说了什么?
虎嗅 2025-02-19T08:18:29.000000Z
Max Tegmark组新工作:利用调和损失训练可解释的AI模型
集智俱乐部 2025-02-13T15:38:04.000000Z
Process Reinforcement through Implicit Rewards (PRIME): A Scalable Machine Learning Framework for Enhancing Reasoning Capabilities
MarkTechPost@AI 2025-02-08T03:49:59.000000Z
谢赛宁新作:表征学习有多重要?一个操作刷新SOTA,DiT训练速度暴涨18倍
硅星人Pro 2024-10-29T00:26:49.000000Z
关于 RWKV 架构的一些谣言
RWKV元始智能 2024-10-28T00:09:59.000000Z
谢赛宁新作:表征学习有多重要?一个操作刷新SOTA,DiT训练速度暴涨18倍
智源社区 2024-10-24T03:23:48.000000Z