Unite.AI 06月10日 01:02
Stopping AI from Spinning Stories: A Guide to Preventing Hallucinations
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了人工智能(AI)在各行业中的应用,以及AI“幻觉”带来的挑战。AI幻觉指的是AI生成的错误信息,可能导致严重后果,如误导性回答和法律问题。文章分析了AI幻觉产生的原因,如输入数据的准确性、时效性和偏见等。文章强调了在选择AI工具时,企业需要谨慎,重视数据的质量和模型的训练。为了减少AI幻觉,文章建议采用严谨的测试、人类反馈以及关注动态语义理论。最终,文章呼吁企业重视AI的准确性,以提升客户体验。

🤔 AI幻觉指的是人工智能生成的错误信息,其频率不容忽视,现代大型语言模型(LLM)的错误率可能在1%到30%之间。

🗣️ AI的输出质量取决于输入数据的质量,不准确、过时或有偏见的数据会导致错误信息的产生。缺乏人为干预和监督也会增加幻觉的风险。

⚠️ AI幻觉可能损害企业声誉和客户忠诚度。如果客户从聊天机器人那里得到不准确的回复,或者员工需要花费时间核实聊天机器人的输出,可能会导致客户流失。

💡 减少AI幻觉的关键在于严谨的测试和训练。这包括使用特定于行业和企业的数据,进行基于转弯的对话测试,以及直接的人类反馈,特别是针对语气和语法的回应。

AI is revolutionizing the way nearly every industry operates. It’s making us more efficient, more productive, and – when implemented correctly – better at our jobs overall. But as our reliance on this novel technology increases rapidly, we have to remind ourselves of one simple fact: AI is not infallible. Its outputs should not be taken at face value because, just like humans, AI can make mistakes.

We call these mistakes “AI hallucinations.” Such mishaps range anywhere from answering a math problem incorrectly to providing inaccurate information on government policies. In highly regulated industries, hallucinations can lead to costly fines and legal trouble, not to mention dissatisfied customers.

The frequency of AI hallucinations should therefore be cause for concern: it’s estimated that modern large language models (LLMs) hallucinate anywhere from 1% to 30% of the time. This results in hundreds of false answers generated on a daily basis, which means businesses looking to leverage this technology must be painstakingly selective when choosing which tools to implement.

Let’s explore why AI hallucinations happen, what’s at stake, and how we can identify and correct them.

Garbage in, garbage out

Do you remember playing the game “telephone” as a child? How the starting phrase would get warped as it passed from player to player, resulting in a completely different statement by the time it made its way around the circle?

The way AI learns from its inputs is similar. The responses LLMs generate are only as good as the information they’re fed, which means incorrect context can lead to the generation and dissemination of false information. If an AI system is built on data that’s inaccurate, out of date, or biased, then its outputs will reflect that.

As such, an LLM is only as good as its inputs, especially when there’s a lack of human intervention or oversight. As more autonomous AI solutions proliferate, it’s critical that we provide tools with the correct data context to avoid causing hallucinations. We need rigorous training of this data, and/or the ability to guide LLMs in such a way that they respond only from the context they’re provided, rather than pulling information from anywhere on the internet.

Why do hallucinations matter?

For customer-facing businesses, accuracy is everything. If employees are relying on AI for tasks like synthesizing customer data or answering customer queries, they need to trust that the responses such tools generate are accurate.

Otherwise, businesses risk damage to their reputation and customer loyalty. If customers are fed insufficient or false answers by a chatbot, or if they’re left waiting while employees fact-check the chatbot’s outputs, they may take their business elsewhere. People shouldn’t have to worry about whether or not the businesses they interact with are feeding them false information – they want swift and reliable support, which means getting these interactions right is of the utmost importance.

Business leaders must do their due diligence when selecting the right AI tool for their employees. AI is supposed to free up time and energy for staff to focus on higher-value tasks; investing in a chatbot that requires constant human scrutiny defeats the whole purpose of adoption. But are the existence of hallucinations really so prominent or is the term simply over-used to identify with any response we assume to be incorrect?

Combating AI hallucinations

Take into consideration: Dynamic Meaning Theory (DMT), the concept that an understanding between two persons – in this case the user and the AI – are being exchanged. But, the limitations of language and knowledge of the subjects cause a misalignment in the interpretation of the response.

In the case of AI-generated responses, it is possible that the underlying algorithms are not yet fully equipped to accurately interpret or generate text in a way that aligns with the expectations we have as humans. This discrepancy can lead to responses that may seem accurate on the surface but ultimately lack the depth or nuance required for true understanding.

Furthermore, most general-purpose LLMs pull information only from content that’s publicly available on the internet. Enterprise applications of AI perform better when they’re informed by data and policies that are specific to individual industries and businesses. Models can also be improved with direct human feedback – particularly agentic solutions that are designed to respond to tone and syntax.

Such tools should also be stringently tested before they become consumer-facing. This is a critical part of preventing AI hallucinations. The entire flow should be tested using turn-based conversations with the LLM playing the role of a persona. This allows businesses to better assume the general success of conversations with an AI model before releasing it into the world.

It’s essential for both developers and users of AI technology to remain aware of dynamic meaning theory in the responses they receive, as well as the dynamics of the language being used in the input. Remember, context is key. And, as humans, most of our context is understood through unspoken means, whether that be through body language, societal trends — even our tone. As humans, we have the potential to hallucinate in response to questions. But, in our current iteration of AI, our human-to-human understanding isn’t so easily contextualized, so we need to be more critical of the context we provide in writing.

Suffice it to say – not all AI models are created equal. As the technology develops to complete increasingly complex tasks, it’s crucial for businesses eyeing implementation to identify tools that will improve customer interactions and experiences rather than detract from them.

The onus isn’t just on solutions providers to ensure they’ve done everything in their power to minimize the chance for hallucinations to occur. Potential buyers have their role to play too. By prioritizing solutions that are rigorously trained and tested and can learn from proprietary data (instead of anything and everything on the internet), businesses can make the most out of their AI investments to set employees and customers up for success.

The post Stopping AI from Spinning Stories: A Guide to Preventing Hallucinations appeared first on Unite.AI.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI幻觉 人工智能 数据质量 LLM
相关文章