热点
关于我们
xx
xx
"
持续预训练
" 相关文章
A Practice of Post-Training on Llama-3 70B with Optimal Selection of Additional Language Mixture Ratio
cs.AI updates on arXiv.org
2025-07-29T04:22:41.000000Z
Data Mixing Agent: Learning to Re-weight Domains for Continual Pre-training
cs.AI updates on arXiv.org
2025-07-22T04:44:31.000000Z
HanjaBridge: Resolving Semantic Ambiguity in Korean LLMs via Hanja-Augmented Pre-Training
cs.AI updates on arXiv.org
2025-07-16T04:29:02.000000Z
单向VLM变双向!人大斯坦福等提出MoCa框架:双向多模态编码器
新智元
2025-07-10T12:06:34.000000Z
Noteworthy AI Research Papers of 2024 (Part One)
Ahead of AI
2024-12-31T12:28:58.000000Z
Aquila-Med LLM:开创性的全流程开源医疗语言模型
智源研究院
2024-10-24T17:00:57.000000Z