The Networking Nerd 2024年07月05日
Human Generated Questions About AI Assistants
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了构建AI虚拟助理时需要考虑的三个关键问题:LLM的选择、数据来源以及如何处理AI幻觉。作者建议企业在选择LLM时,应权衡使用服务型产品或自行构建系统的优劣;在数据来源方面,应考虑是否使用来自其他公司的匿名数据来建立基线,并注意潜在的风险;在处理AI幻觉方面,应采取措施确保AI能够准确回答问题,而不是编造答案。

🤔**LLM的选择:服务型产品vs自建系统** 构建AI虚拟助理的第一步是选择合适的LLM(大型语言模型)。目前,企业可以选择使用服务型产品,例如OpenAI,或者自行构建系统。服务型产品无需编程,只需将数据输入LLM即可,但需要付费使用,且使用成本会随着使用量的增加而增加。自建系统则可以完全控制数据处理方式,但需要进行维护和更新。 选择LLM时,企业需要根据自身需求和资源情况进行权衡。如果需要快速部署,并愿意支付一定的费用,可以选择服务型产品。如果需要更灵活的控制,并愿意投入人力和资源,可以选择自建系统。

🌎**数据来源:企业数据vs外部数据** AI虚拟助理的训练数据来源至关重要,它决定了AI虚拟助理的回答质量和准确性。企业可以选择使用自身数据进行训练,也可以使用来自其他公司的匿名数据来建立基线。使用自身数据可以确保AI虚拟助理能够准确地回答有关企业自身网络的问题,但需要时间进行学习和分析。使用外部数据可以帮助企业了解行业基准,但需要谨慎选择数据来源,避免引入错误信息。 企业在选择数据来源时,应考虑数据质量、数据安全和隐私问题。

👻**幻觉处理:识别与应对** AI幻觉是指AI虚拟助理编造答案的情况,这是AI技术发展过程中面临的一个重要挑战。企业需要采取措施识别和应对AI幻觉,确保AI虚拟助理能够准确地回答问题。 识别AI幻觉的方法包括:检查AI虚拟助理的答案是否符合逻辑、是否与已知信息相符、是否与其他数据源的信息一致。应对AI幻觉的方法包括:对AI虚拟助理进行更严格的训练、使用更可靠的数据源、开发更有效的幻觉检测算法。 企业在构建AI虚拟助理时,应将幻觉处理作为一项重要工作,确保AI虚拟助理能够提供准确可靠的信息。

I’ve taken a number of briefings in the last few months that all mention how companies are starting to get into AI by building an AI virtual assistant. In theory this is the easiest entry point into the technology. Your network already has a ton of information about usage patterns and trouble spots. Network operations and engineering teams have learned over the years to read that information and provide analysis and feedback.

If marketing is to be believed, no one in the modern world has time to learn how to read all that data. Instead, AI provides a natural language way to ask simple questions and have the system provide the data back to you with proper context. It will highlight areas of concern and help you grasp what’s going on. Only you don’t need to get a CCNA to get there. Or, more likely, it’s more useful for someone on the executive team to ask questions and get answers without the need to talk to the network team.

I have some questions that I always like to ask when companies start telling me about their new AI assistant that help me understand how it’s being built.

Question 1: Laying Out LLMs

My first question is always:

Which LLM are you using to power your system?

The reason is because there are only two real options. You’re either paying someone else to do it as a service, like OpenAI, or you’re pulling down your own large language model (LLM) and building your own system. Both have advantages and disadvantages.

The advantage of a service-based offering is that you don’t need to program anything. You just feed the data to the LLM and it takes off. No tuning needed. It’s fast and universally available.

The downside of a service based model is the fact that it costs money. And if you’re using it commercially it’s going to cost more than a simple monthly fee. The more you use it, the more expensive it gets. If your vendor is pulling thousands of daily requests from the LLM is that factored into the fee they’re charging you? What happens when the OpenAI prices go up?

The advantages of building your own system are that you have complete control over the way the data is being processed. You tune the LLM and you own the way it’s being used. No need to pay more to someone else to do all the work for you. You can also decide how and when features are implemented. If you’re updating the LLM on your schedule you can include new features when they’re ready and not when OpenAI pushes them live and makes them available for everyone.

The disadvantages of building your own system involves maintenance. You have to update and patch it. You have to figure out what features to develop. You have to put in the work. And if the model you use goes out of support or is no longer being maintained you have to swap to something new and hope that all your functions are going to work with the new one.

Question 2: Data Sources

My second question:

Where does the LLM data come from?

May seem simple at first, right? You’re training your LLM on your data so it gives you answers based on your environment. You’d want that to be the case so it’s more likely to tell you things about your network. But that insight doesn’t come out of thin air. If you want to feed your data to the LLM to get answers you’re going to have to wait while it studies the network and comes up with conclusions.

I often ask companies if they’re populating the system with anonymized data from other companies to provide baselines. I’ve seen this before from companies like Nyansa, which was bought by VMware, and Raza Networks, while is part of HPE Aruba. Both of those companies, which came out long before the current AI craze, collected data from customers and used it to build baselines for everyone. If you wanted to see how you compared to other high education or medical verticals the system could tell you what those types of environments looked like, with the names obscured of course.

Pre-populating the LLM with information from other companies is great if your stakeholders want to know how they fare against other companies. But it also runs the risk of populating data that shouldn’t be in the system. That could create situations where you’re acting on bad information or chasing phantoms in the organization. Worse yet, your own data could be used in ways you didn’t intend to feed other organizations. Even with the names obscured someone might be able to engineer a way to obtain knowledge about your environment you don’t want everyone to have.

Question 3: Are You Seeing That?

My third question:

How do you handle hallucinations?

Hallucination is the term for when the AI comes up with an answer that is false. That’s right, the super intelligent system just made up an answer instead of saying “I don’t know”. Which is great if you’re trying to convince someone you’re smart or useful. But if the entire reason why I’m using your service is accurate answers about my problems I’d rather have you say you don’t have an answer or you need to do research instead of giving me bad data that I use to make bad decisions.

If a company tells me they don’t really see hallucinations then I immediately get concerned, especially if they’re leveraging OpenAI for their LLM. I’ve talked before about how ChatGPT has a really bad habit of making up answers so it always looks like it knows everything. That’s great if you’re trying to get the system to write a term paper for you. It’s really bad if you try to reroute traffic in your network around a non-existent problem. I know there are many systems out there that can help reduce hallucinations, such as retrieval augmented generation (RAG), but I need that to be addressed up front instead of a simple “we don’t see hallucinations” because that makes me feel like something is being hidden or glossed over.


Tom’s Take

These aren’t the only questions you should be asking about AI and LLMs in your network but they’re not a bad start. They encompass the first big issues that people are likely to run into when evaluating an AI system. How do you do your analysis? What is happening with my data? What happens when the system doesn’t know what to do? Sure, there’s always going to be questions about cost and lock-in but I’d rather know the technology is sound before I ever try to deploy the system. You can always negotiate cost. You can’t negotiate with a flaw AI.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI虚拟助理 LLM 数据来源 幻觉处理
相关文章