热点
关于我们
xx
xx
"
事实准确性
" 相关文章
研究发现向聊天机器人询问简短的答案可能会增加幻觉
Cnbeta
2025-05-08T12:39:10.000000Z
多模态大模型事实正确性评估:o1最强,模型普遍过于自信,最擅长现代建筑/工程技术/科学
智源社区
2025-02-24T07:37:16.000000Z
让「幻觉」无处遁形!谷歌DeepMind全新基准,三代Gemini同台霸榜
新智元
2025-01-13T16:54:44.000000Z
让「幻觉」无处遁形,谷歌DeepMind全新基准,三代Gemini同台霸榜
36氪 - 科技频道
2025-01-13T10:43:05.000000Z
Meta AI Introduces EWE (Explicit Working Memory): A Novel Approach that Enhances Factuality in Long-Form Text Generation by Integrating a Working Memory
MarkTechPost@AI
2025-01-04T06:34:52.000000Z
谷歌发布 FACTS Grounding 基准:Gemini、GPT-4o、Claude 当评委,成 AI 大语言模型“幻觉照妖镜”
IT之家
2024-12-18T05:22:36.000000Z
NeurIPS 2024|杜克大学&谷歌提出SLED解码框架,无需外部数据与额外训练,有效缓解大语言模型幻觉,提高事实准确性
机器之心
2024-12-02T06:09:56.000000Z
OpenAI :最新研究发现AI模型在回答事实问题时表现拉胯 GPT-4o准确率才38.2%
互联网数据资讯网-199IT
2024-11-03T11:01:26.000000Z
OpenAI 开源 SimpleQA 新基准,专治大模型“胡言乱语”
IT之家
2024-10-30T23:37:55.000000Z
MMed-RAG: A Versatile Multimodal Retrieval-Augmented Generation System Transforming Factual Accuracy in Medical Vision-Language Models Across Multiple Domains
MarkTechPost@AI
2024-10-19T16:50:54.000000Z
FactAlign: A Novel Alignment AI Framework Designed to Enhance the Factuality of LLMs’ Long-Form Responses While Maintaining Their Helpfulness
MarkTechPost@AI
2024-10-06T14:05:12.000000Z
Rethinking QA Dataset Design: How Popular Knowledge Enhances LLM Accuracy?
MarkTechPost@AI
2024-07-04T20:01:42.000000Z
Unveiling the Shortcuts: How Retrieval Augmented Generation (RAG) Influences Language Model Behavior and Memory Utilization
MarkTechPost@AI
2024-06-20T10:01:45.000000Z