热点
"LLM性能" 相关文章
MArgE: Meshing Argumentative Evidence from Multiple Large Language Models for Justifiable Claim Verification
cs.AI updates on arXiv.org 2025-08-05T11:29:13.000000Z
GneissWeb: Preparing High Quality Data for LLMs at Scale
cs.AI updates on arXiv.org 2025-07-31T04:48:26.000000Z
LLM-Crowdsourced: A Benchmark-Free Paradigm for Mutual Evaluation of Large Language Models
cs.AI updates on arXiv.org 2025-07-31T04:47:53.000000Z
Is It Time To Treat Prompts As Code? A Multi-Use Case Study For Prompt Optimization Using DSPy
cs.AI updates on arXiv.org 2025-07-08T06:58:11.000000Z
Princeton University Researchers Introduce Self-MoA and Self-MoA-Seq: Optimizing LLM Performance with Single-Model Ensembles
MarkTechPost@AI 2025-02-07T17:35:11.000000Z
KAG开源了,知识增强掀翻RAG,性能翻倍
PaperAgent 2024-10-28T12:44:21.000000Z
Where do LLMs spend their FLOPS?
Artificial Fintelligence 2024-10-22T06:07:41.000000Z
Archon: A Machine Learning Framework for Large Language Model Enhancement Using Automated Inference-Time Architecture Search for Improved Task Performance
MarkTechPost@AI 2024-10-10T17:21:38.000000Z
MAGICORE: An AI Framework for Multi Agent Iteration for Coarse-to-fine Refinement
MarkTechPost@AI 2024-09-23T10:05:32.000000Z
CoT神话破灭,并非LLM标配!三大学府机构联手证实,CoT仅在数学符号推理有用
智源社区 2024-09-22T06:08:16.000000Z