热点
关于我们
xx
xx
"
预训练
" 相关文章
Positive-Augmented Contrastive Learning for Vision-and-Language Evaluation and Training
cs.AI updates on arXiv.org
2025-07-31T04:48:23.000000Z
Spike No More: Stabilizing the Pre-training of Large Language Models
cs.AI updates on arXiv.org
2025-07-28T04:43:03.000000Z
Can Synthetic Images Conquer Forgetting? Beyond Unexplored Doubts in Few-Shot Class-Incremental Learning
cs.AI updates on arXiv.org
2025-07-21T04:06:50.000000Z
Two-Stage Pretraining for Molecular Property Prediction in the Wild
cs.AI updates on arXiv.org
2025-07-21T04:06:49.000000Z
大模型新周期开始,25H2就是24H1
韭研公社
2025-07-20T10:32:18.000000Z
【华福计算机】KIMI K2——尚未到达的通用模型天花板 ,提升点在哪?
韭研公社
2025-07-18T01:05:33.000000Z
Compute Requirements for Algorithmic Innovation in Frontier AI Models
cs.AI updates on arXiv.org
2025-07-16T04:28:50.000000Z
B-cos LM: Efficiently Transforming Pre-trained Language Models for Improved Explainability
cs.AI updates on arXiv.org
2025-07-15T04:27:17.000000Z
TabDPT: Scaling Tabular Foundation Models on Real Data
cs.AI updates on arXiv.org
2025-07-15T04:27:14.000000Z
5分钟带你搞懂从0打造一个ChatGPT
掘金 人工智能
2025-07-14T10:38:58.000000Z
5分钟带你搞懂从0打造一个ChatGPT
掘金 人工智能
2025-07-14T00:43:29.000000Z
【周末特辑】7月第2周最火AI论文 | 长视频推理框架创新;内存操作系统提升AI性能
HuggingFace 每日AI论文速递
2025-07-13T06:08:01.000000Z
单向VLM变双向!人大斯坦福等提出MoCa框架:双向多模态编码器
智源社区
2025-07-11T08:54:02.000000Z
Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs
cs.AI updates on arXiv.org
2025-07-11T04:04:03.000000Z
大模型炼丹术(五):LLM自回归预训练过程详解
掘金 人工智能
2025-07-11T01:03:52.000000Z
大模型炼丹术(一):从tokenizer说起,为LLM自回归预训练准备数据集
掘金 人工智能
2025-07-09T01:13:16.000000Z
喝点VC|红杉美国对谈OpenAI前研究主管:预训练已经进入边际效益递减阶段,其真正杠杆在于架构的改进
Z Potentials
2025-07-08T06:14:09.000000Z
The Art and Science of Fine-Tuning LLMs for Domain-Specific Excellence
Cogito Tech
2025-07-07T06:34:40.000000Z
The Base Model Lens
少点错误
2025-07-07T00:17:24.000000Z
RoBERTa:优化训练流程,让 BERT 更强
掘金 人工智能
2025-06-30T08:33:11.000000Z