A Geodyssey – Enterprise Search Discovery, Text Mining, Machine Learning 02月05日
Misconceptions of LLM Chatbots in Geoscience
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了LLM聊天机器人在科学和商业应用中的局限性。目前许多机构采用的检索增强生成(RAG)技术,本质上类似于在文本段落上进行“Google式”搜索,仅将最相关的段落附加到用户的问题中。这种方法虽然能提供看似准确的答案,但它只使用了部分文本语料库,容易忽略离群值,导致用户错过信息空间中最有用的信息。文章强调,如果追求数据驱动的科学和商业发现,需要超越有限的LLM聊天机器人技术,采用能够整合所有信息的技术,包括离群值。

🔍RAG技术:检索增强生成(RAG)技术是目前许多机构采用的,用于在LLM聊天机器人中提供尽可能准确的源材料和AI生成答案之间联系的技术。它类似于在文本段落上进行“Google式”搜索,并将最相关的段落附加到用户的问题中。

🎯信息检索驱动:RAG技术实际上是一种信息检索(IR)驱动的技术,提取/总结发生在最后。它不使用整个文本语料库或信息空间,因此存在与Google搜索相同的问题,即只呈现前n个结果,即在统计上与你的提示(问题)最相似的结果。

💡探索性搜索的局限性:对于探索性搜索目标,RAG技术很可能像Google搜索一样,将你引向最明显的内容。离群值会被忽略,你可能无法偏离既定的道路。这可能会导致人们误以为“AI”是某种特别的东西,而实际上它并没有利用整个文本语料库。

⚠️潜在的误导性:LLM聊天机器人生成的答案可能具有很强的说服力,但重要的是要知道,你可能没有得到信息空间中最有用和重要的信息,因为只使用了其中的一小部分。

Misconceptions of LLM Chatbots: For scientists and business professionals it is critical to know the source of any AI generated answer or assertion. If we cannot trace the sources accurately we are unlikely to trust the output. Imagine reading a literature review where no sources were cited.

The technique used to provide as accurate as possible linkage between source material and AI generated answers is called Retrieval Augmented Generation (RAG). This is what many organisations appear to be deploying.

Simplifying, this is essentially a ‘Google-like’ search on paragraphs of text rather than full documents, with the top n paragraphs (chunks) simply appended as content to your question (LLM prompt). These paragraphs determine the usefulness of the answer you get, the LLM just summarises this somewhat.

The implication is that this is effectively an Information Retrieval (IR) driven technique. The extraction/summarisation happens at the end. It does not use the whole text corpus or information space, so suffers the same issues raised with Google search, in that it only presents the top n results, what is statistically most similar (embedding) to your prompt (question).

Just like Google this lends itself to lookup/known item questions, where there is a right answer. For example, “What is the age of the Bathonian?”. For exploratory search goals, this technique will likely, just like Google search, drive you to what is most obvious. For example, “What do we know about host rocks for copper in this area?”. Outliers will be missed and you may not be taken off the well beaten path.

This can be an issue because we may think of “AI” in this sense as something special, more than it is. Something that utilises the whole text corpus. It absolutely does not in this case. We are using the statistical similarity of words to come up with a ranked set of paragraphs which go into a summariser. The obvious will always likely out-compete the less obvious.

This may be very important to know so people are not ‘fooled’ by the highly convincing AI generated answer. This is not a point about so called ‘hallucinations’. You may have been given an answer which is true for your exploratory search question. The point is that you may not have been given the most useful and significant information held within the information space, available because only a fraction of it is being used.

If we are interested in data driven discovery for science and business opportunities, we need to look beyond just limited ‘LLM driven chatbot’ style techniques. To include techniques which stack ‘everything’ including outliers from our information space. Language Models and other Natural Language Processing techniques offer so much more than a Chatbot…

I touch on some of these points in an ethics paper published late last year. Link in the comments.

https://www.journalofgeoethics.eu/index.php/jgsg/article/view/63

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

LLM聊天机器人 检索增强生成(RAG) 信息检索 数据驱动发现
相关文章