AI News 04月30日 21:42
Meta beefs up AI security with new Llama tools 
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Meta 近期推出一系列 Llama AI 模型的新安全工具,旨在提升 AI 开发和使用的安全性。这些工具包括 Llama Guard 4,一个增强型安全过滤器,现已支持多模态,能够识别和应用安全规则于文本和图像。此外,Meta 还推出了 LlamaFirewall,一个用于管理 AI 系统安全性的控制中心,以及升级版的 Llama Prompt Guard。同时,Meta 还更新了 CyberSec Eval 4 评估套件,并启动了 Llama Defenders 项目,以帮助网络安全团队利用 AI 防御网络攻击。Meta 还分享了内部使用的 AI 安全工具,如自动敏感文档分类工具,以及用于检测 AI 生成音频的工具,如 Llama Generated Audio Detector 和 Llama Audio Watermark Detector。此外,Meta 还在 WhatsApp 上推出了 Private Processing 技术,以增强用户隐私。

🛡️ Llama Guard 4 是一款可定制的安全过滤器,现已支持多模态,能够处理文本和图像内容,并应用安全规则。它被集成到 Meta 的 Llama API 中,用于增强 AI 应用的安全性。

🚦 LlamaFirewall 作为 AI 系统的安全控制中心,旨在管理不同的安全模型,并与 Meta 的其他保护工具协同工作,以检测和阻止潜在的 AI 风险,如提示注入攻击、恶意代码生成和 AI 插件的风险行为。

🚀 Meta 还推出了 Prompt Guard 2 22M,这是一个更小、更快的版本,与大型模型相比,延迟和计算成本降低了 75%,同时保持了检测能力,适用于需要更快响应或预算有限的场景。

🛡️ CyberSec Eval 4 评估套件的更新,包括 CyberSOC Eval 和 AutoPatchBench,前者用于评估 AI 在安全运营中心(SOC)环境中的表现,后者用于测试 AI 自动修复代码漏洞的能力。

📢 Meta 启动了 Llama Defenders 项目,为合作伙伴和开发者提供 AI 安全解决方案,包括开源、早期访问和专有工具,以应对不同的安全挑战。Meta 还分享了内部使用的 AI 安全工具,如自动敏感文档分类工具和 AI 生成音频检测工具。

If you’re building with AI, or trying to defend against the less savoury side of the technology, Meta just dropped new Llama security tools.

The improved security tools for the Llama AI models arrive alongside fresh resources from Meta designed to help cybersecurity teams harness AI for defence. It’s all part of their push to make developing and using AI a bit safer for everyone involved.

Developers working with the Llama family of models now have some upgraded kit to play with. You can grab these latest Llama Protection tools directly from Meta’s own Llama Protections page, or find them where many developers live: Hugging Face and GitHub.

First up is Llama Guard 4. Think of it as an evolution of Meta’s customisable safety filter for AI. The big news here is that it’s now multimodal so it can understand and apply safety rules not just to text, but to images as well. That’s crucial as AI applications get more visual. This new version is also being baked into Meta’s brand-new Llama API, which is currently in a limited preview.

Then there’s LlamaFirewall. This is a new piece of the puzzle from Meta, designed to act like a security control centre for AI systems. It helps manage different safety models working together and hooks into Meta’s other protection tools. Its job? To spot and block the kind of risks that keep AI developers up at night – things like clever ‘prompt injection’ attacks designed to trick the AI, potentially dodgy code generation, or risky behaviour from AI plug-ins.

Meta has also given its Llama Prompt Guard a tune-up. The main Prompt Guard 2 (86M) model is now better at sniffing out those pesky jailbreak attempts and prompt injections. More interestingly, perhaps, is the introduction of Prompt Guard 2 22M.

Prompt Guard 2 22M is a much smaller, nippier version. Meta reckons it can slash latency and compute costs by up to 75% compared to the bigger model, without sacrificing too much detection power. For anyone needing faster responses or working on tighter budgets, that’s a welcome addition.

But Meta isn’t just focusing on the AI builders; they’re also looking at the cyber defenders on the front lines of digital security. They’ve heard the calls for better AI-powered tools to help in the fight against cyberattacks, and they’re sharing some updates aimed at just that.

The CyberSec Eval 4 benchmark suite has been updated. This open-source toolkit helps organisations figure out how good AI systems actually are at security tasks. This latest version includes two new tools:

To help get these kinds of tools into the hands of those who need them, Meta is kicking off the Llama Defenders Program. This seems to be about giving partner companies and developers special access to a mix of AI solutions – some open-source, some early-access, some perhaps proprietary – all geared towards different security challenges.

As part of this, Meta is sharing an AI security tool they use internally: the Automated Sensitive Doc Classification Tool. It automatically slaps security labels on documents inside an organisation. Why? To stop sensitive info from walking out the door, or to prevent it from being accidentally fed into an AI system (like in RAG setups) where it could be leaked.

They’re also tackling the problem of fake audio generated by AI, which is increasingly used in scams. The Llama Generated Audio Detector and Llama Audio Watermark Detector are being shared with partners to help them spot AI-generated voices in potential phishing calls or fraud attempts. Companies like ZenDesk, Bell Canada, and AT&T are already lined up to integrate these.

Finally, Meta gave a sneak peek at something potentially huge for user privacy: Private Processing. This is new tech they’re working on for WhatsApp. The idea is to let AI do helpful things like summarise your unread messages or help you draft replies, but without Meta or WhatsApp being able to read the content of those messages.

Meta is being quite open about the security side, even publishing their threat model and inviting security researchers to poke holes in the architecture before it ever goes live. It’s a sign they know they need to get the privacy aspect right.

Overall, it’s a broad set of AI security announcements from Meta. They’re clearly trying to put serious muscle behind securing the AI they build, while also giving the wider tech community better tools to build safely and defend effectively.

See also: Alarming rise in AI-powered scams: Microsoft reveals $4B in thwarted fraud

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Meta beefs up AI security with new Llama tools  appeared first on AI News.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Meta Llama AI安全 网络安全
相关文章