MarkTechPost@AI 2024年09月18日
An Extensible Open-Source AI Framework to Benchmark Attributable Information-Seeking Using Representative LLM-based Approaches
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

法国研究团队提出可扩展的AI框架,用于评估归因信息寻求,该框架支持多种LLM架构,不同方法在性能上有差异

🌐 该AI框架支持多种LLM架构,适应任意数据集,旨在对归因信息寻求任务进行基准测试,涵盖生成、检索后生成、生成后检索等方法

📈 性能评估使用HAGRID数据集,结果显示RTG方法在答案正确性和引用质量方面总体表现更好,其中RTG-查询生成场景的引用质量得分最高

🔍 实验发现增加支持文档数量对引用有混合影响,自动引用指标在开放信息寻求任务中效果欠佳,该框架为未来研究提供了有价值的见解和基准

🎉 此框架通过关注答案正确性和引用质量,有效解决了评估归因信息寻求场景中的差距,RTG-查询生成方法在引用准确性上有显著提升

With the success of LLMs in various tasks, search engines have begun using generative methods to provide accurate answers with in-line citations to user queries. However, generating reliable and attributable answers, especially in open-ended information-seeking scenarios, poses challenges due to the complexity of questions and the broad scope of candidate-attributed answers. Existing methods typically focus on attributed question-answering, which involves finding specific answers to precise queries without focusing on the more difficult problem of attributed information seeking. The primary issue is the potential for LLMs to generate incorrect or “hallucinated” information.

A team of researchers from France proposed a reproducible AI framework that supports various LLM architectures for attributed information seeking and is adaptable to any dataset. The proposed framework is designed to benchmark attributed information-seeking tasks with different LLM architectures. The Generate approach leverages LLMs to produce answers based solely on their pre-existing knowledge. In the Retrieve Then Generate approach, documents relevant to the query are first retrieved, and then the LLM generates answers with citations based on these documents. The framework includes variations of RTG, such as vanilla retrieval and query generation, where the latter involves generating subqueries to improve retrieval accuracy. In the Generate Then Retrieve approach, answers are initially generated without citations, followed by identifying relevant documents to support the generated statements.

Performance evaluations using the HAGRID dataset show that RTG approaches outperform other methods, demonstrating better overall performance in answer correctness and citation quality. Specifically, the RTG-query-gen scenario generates queries to guide document retrieval and achieves the highest citation quality scores. The analysis reveals that the quality of citations and the impact of retrieval methods are crucial factors in the effectiveness of attributed information-seeking systems. The framework also includes various metrics to evaluate both answer correctness and citation quality, with results indicating that RTG methods generally yield superior outcomes compared to GTR approaches.

Several other findings emerged from their experiments. Increasing the number of supporting documents had a mixed impact on citations. While it improved recall for citations, it also led to issues with over-citation. Furthermore, the study revealed that while automatic citation metrics such as AutoAIS and natural language inference (NLI)-based metrics correlate well with human judgments in question-answering scenarios, they perform less effectively in open-ended information-seeking tasks.

In conclusion, the proposed framework effectively addresses the gap in evaluating attributed information-seeking scenarios by proposing a comprehensive, open-source framework that supports various LLM architectures. By focusing on both answer correctness and citation quality, the framework offers valuable insights and benchmarks for future research. The RTG-query-gen approach demonstrates significant improvements in citation accuracy, highlighting the importance of effective document retrieval and query generation in attributed information-seeking tasks.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 50k+ ML SubReddit

FREE AI WEBINAR: ‘SAM 2 for Video: How to Fine-tune On Your Data’ (Wed, Sep 25, 4:00 AM – 4:45 AM EST)

The post An Extensible Open-Source AI Framework to Benchmark Attributable Information-Seeking Using Representative LLM-based Approaches appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI框架 归因信息寻求 性能评估 RTG方法
相关文章