Unite.AI 06月04日 01:27
If Your AI Is Hallucinating, Don’t Blame the AI
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了AI“幻觉”现象,即AI生成看似可信但实际上错误的答案。文章指出,在商业应用中,AI幻觉是一个严重问题,但解决之道并非责怪AI,而是确保为AI提供正确、相关的数据。作者分享了其团队在构建AI代理时采取的措施,包括确保AI提问正确问题、使用结构化的方法、构建高质量的数据提取工具、让AI展示其工作以及设置安全防护措施。文章强调,高质量的数据是减少AI幻觉的关键,并呼吁人们正确利用AI,投入时间和精力使其发挥作用。

🤔 AI“幻觉”是由于AI在缺乏相关数据时,为了给出答案而编造信息。这在商业应用中尤其危险,可能导致错误的决策。

💡 解决AI幻觉问题的关键在于正确使用AI。这意味着为AI提供最相关、最优质的数据,而不是责怪AI本身。

✅ 构建可靠AI代理的措施包括:设计代理提问正确问题并验证数据、使用结构化方法避免每次都重新创建方案、构建高质量的数据提取工具、让AI展示其数据来源以及设置安全防护措施。

🔍 确保AI展示其工作,例如引用数据来源,使用户可以验证信息,避免AI“凭空捏造”。

🛡️ 设置安全防护措施,例如当AI缺乏数据时,宁愿无法回答,也不要提供错误或误导性答案,这是至关重要的。

AI “hallucinations” – those convincing-sounding but false answers – draw a lot of media attention, as with the recent New York Times article, AI Is Getting More Powerful, But Its Hallucinations Are Getting Worse. Hallucinations are a real hazard when you’re dealing with a consumer chatbot. In the context of business applications of AI, it’s an even more serious concern. Fortunately, as a business technology leader I have more control over it as well. I can make sure the agent has the right data to produce a meaningful answer.

Because that’s the real problem. In business, there is no excuse for AI hallucinations. Stop blaming AI. Blame yourself for not using AI properly.

When generative AI tools hallucinate, they are doing what they are designed to do – provide the best answer they can based on the data they have available. When they make stuff up, producing an answer that is not based in reality, it’s because they’re missing the relevant data, can’t find it, or don’t understand the question. Yes, new models like OpenAI’s o3 and o4-mini are hallucinating more, acting even more “creative” when they don’t have a good answer to the question that’s been posed to them. Yes, more powerful tools can hallucinate more – but they can also produce more powerful and valuable results if we set them up for success.

If you don’t want your AI to hallucinate, don’t starve it for data. Feed the AI the best, most relevant data for the problem you want it to solve, and it won’t be tempted to go astray.

Even then, when working with any AI tool, I recommend keeping your critical thinking skills intact. The results AI agents deliver can be productive and delightful, but the point is not to unplug your brain and let the software do all the thinking for you. Keep asking questions. When an AI agent gives you an answer, question that answer to be sure it makes sense and is backed by data. If so, that should be an encouraging sign that it’s worth your time to ask follow up questions.

The more you question, the better insights you will get.

Why hallucinations happen

It’s not some mystery. The AI is not trying to lie to you. Every large language model (LLM) AI is essentially predicting the next word or number based on probability.

At a  high level, what’s happening here is that LLMs string together sentences and paragraphs one word at a time, predicting the next word that should occur in the sentence based on billions of other examples in its training data. The ancestors of LLMs (aside from Clippy) were autocomplete prompts for text messages and computer code, automated human language translation tools, and other probabilistic linguistic systems. With increased brute force compute power, plus training on internet-scale volumes of data, these systems got “smart” enough that they could carry on a full conversation over chat, as the world learned with the introduction of ChatGPT.

AI naysayers like to point out that this is not the same as real “intelligence,” only software that can distill and regurgitate the human intelligence that has been fed into it. Ask it to summarize data in a written report, and it imitates the way other writers have summarized similar data.

That strikes me as an academic argument as long as the data is correct and the analysis is useful.

What happens if the AI doesn’t have the data? It fills in the blanks. Sometimes it’s funny. Sometimes it’s a total mess.

When building AI agents, this is 10x the risk. Agents are supposed to provide actionable insights, but they make more decisions along the way. They executed multi-step tasks, where the result of step 1 informs steps 2, 3, 4, 5, … 10 … 20. If the results of step 1 are incorrect, the error will be amplified, making the output at step 20 that much worse. Especially, as agents can make decisions and skip steps.

Done right, agents accomplish more for the business that deploys them. Yet as AI product managers, we have to recognize the greater risk that goes along with the greater reward.

Which is what our team did. We saw the risk, and tackled it. We didn’t just build a fancy robot;  we made sure it runs on the right data. Here is what I think we did right:

We’ve incorporated these principles into our recent release of our three new agents, with more to follow. For example, our AI Meeting Prep Agent for salespeople doesn’t just ask for the name of the target company but details on the goal of the meeting and who it is with, priming it to provide a better answer. It doesn’t have to guess because it uses a wealth of company data, digital data, and executive profiles to inform its recommendations.

Are our agents perfect? No. Nobody is creating perfect AI yet, not even the biggest companies in the world. But facing the problem is a hell of a lot better than ignoring it.

Want fewer hallucinations? Give your AI a nice chunk of high quality data.

If it hallucinates, maybe it’s not the AI that needs fixing. Maybe it’s your approach to taking advantage of these powerful new capabilities without putting in the time and effort to get them right.

The post If Your AI Is Hallucinating, Don’t Blame the AI appeared first on Unite.AI.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI幻觉 数据质量 AI应用 商业 人工智能
相关文章