AWS Blogs 04月08日 23:49
Amazon Bedrock Guardrails enhances generative AI application safety with new capabilities
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Amazon Bedrock Guardrails推出新功能,旨在帮助客户更有效地实施负责任的AI策略。新功能包括增强的多模态内容检测,能以高达88%的准确率检测有害内容,并过滤敏感信息,防止幻觉。通过ApplyGuardrail API,Guardrails可跨多个基础模型工作,简化了跨多个FMs实施一致的AI安全控制的复杂性,同时通过可配置的控件和针对特定行业和用例的集中管理来维护合规性和负责任的AI策略。

🛡️ **增强的安全策略**:Amazon Bedrock Guardrails提供一套全面的策略,包括多模态内容过滤器、拒绝主题、敏感信息过滤器、词语过滤器、上下文基础检查和自动化推理,以防止不当内容生成。

🖼️ **多模态毒性检测**:新功能支持图像和文本内容的多模态毒性检测,准确率高达88%。这有助于检测和过滤有害内容,确保在不同数据类型上应用一致的内容过滤策略。

⚙️ **一致的保护**:通过在图像和文本数据上应用相同的内容过滤策略,简化了负责任的AI应用程序开发。例如,金融服务公司使用高阈值,在文本和图像输入中都实现了安全保护,避免了网络入侵。

🔬 **实际案例**:文章展示了一个实际案例,一家金融服务公司使用Amazon Bedrock Guardrails,对包含网络安全绕过图和书面说明的图像和文本输入,触发了相同的保护措施,从而确保了安全标准在多模态内容中的一致性。

<section class="blog-post-content lb-rtxt"><table id="amazon-polly-audio-table"><tbody><tr><td id="amazon-polly-audio-tab"><p></p></td></tr></tbody></table><p>Since we launched <a href="https://aws.amazon.com/bedrock/guardrails/&quot;&gt;Amazon Bedrock Guardrails</a> <a href="https://aws.amazon.com/blogs/aws/guardrails-for-amazon-bedrock-now-available-with-new-safety-filters-and-privacy-controls/&quot;&gt;over one year ago</a>, customers like Grab, <a href="https://youtu.be/sTUF-AV7sow&quot;&gt;Remitly&lt;/a&gt;, <a href="https://youtu.be/oTTW_gOgwHA&quot;&gt;KONE&lt;/a&gt;, and <a href="https://press.aboutamazon.com/aws/2024/12/pagerduty-and-aws-deliver-on-the-promise-of-generative-ai-for-business-and-operational-resiliency&quot;&gt;PagerDuty&lt;/a&gt; have used <a href="https://aws.amazon.com/bedrock/guardrails/&quot;&gt;Amazon Bedrock Guardrails</a> to standardize protections across their <a href="https://aws.amazon.com/ai/generative-ai/&quot;&gt;generative AI</a> applications, bridge the gap between native model protections and enterprise requirements, and streamline governance processes. Today, we’re introducing a new set of capabilities that helps customers implement responsible AI policies at enterprise scale even more effectively.</p><p>Amazon Bedrock Guardrails detects harmful multimodal content with up to 88% accuracy, filters sensitive information, and prevent hallucinations. It provides organizations with integrated safety and privacy safeguards that work across multiple <a href="https://aws.amazon.com/what-is/foundation-models/&quot;&gt;foundation models (FMs)</a>, including models available in <a href="https://aws.amazon.com/bedrock/&quot;&gt;Amazon Bedrock</a> and your own custom models deployed elsewhere, thanks to the <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/guardrails-use-independent-api.html&quot;&gt;ApplyGuardrail API</a>. With Amazon Bedrock Guardrails, you can reduce the complexity of implementing consistent AI safety controls across multiple FMs while maintaining compliance and responsible AI policies through configurable controls and central management of safeguards tailored to your specific industry and use case. It also seamlessly integrates with existing AWS services such as <a href="https://aws.amazon.com/iam/&quot;&gt;AWS Identity and Access Management (IAM)</a>, <a href="https://aws.amazon.com/bedrock/agents/&quot;&gt;Amazon Bedrock Agents</a>, and <a href="https://aws.amazon.com/bedrock/knowledge-bases/&quot;&gt;Amazon Bedrock Knowledge Bases</a>.</p><p>“<a href="https://www.grab.com/&quot;&gt;Grab&lt;/a&gt;, a Singaporean multinational taxi service is using Amazon Bedrock Guardrails to ensure the safe use of generative AI applications and deliver more efficient, reliable experiences while maintaining the trust of our customers,” said Padarn Wilson, Head of Machine Learning and Experimentation at Grab. “Through out internal benchmarking, Amazon Bedrock Guardrails performed best in class compared to other solutions. Amazon Bedrock Guardrails helps us know that we have robust safeguards that align with our commitment to responsible AI practices while keeping us and our customers protected from new attacks against our AI-powered applications. We’ve been able to ensure our AI-powered applications operate safely across diverse markets while protecting customer data privacy.”</p><p>Let’s explore the new capabilities we have added.</p><p><strong>New guardrails policy enhancements<br /></strong> Amazon Bedrock Guardrails provides a comprehensive set of policies to help maintain security standards. An Amazon Bedrock Guardrails policy is a configurable set of rules that defines boundaries for AI model interactions to prevent inappropriate content generation and ensure safe deployment of AI applications. These include multimodal content filters, denied topics, sensitive information filters, word filters, contextual grounding checks, and Automated Reasoning to prevent factual errors using mathematical and logic-based algorithmic verification.</p><p>We’re introducing new Amazon Bedrock Guardrails policy enhancements that deliver significant improvements to the six safeguards, strengthening content protection capabilities across your generative AI applications.</p><p><strong>Multimodal toxicity detection with industry leading image and text protection</strong> – Announced as <a href="https://aws.amazon.com/blogs/aws/amazon-bedrock-guardrails-now-supports-multimodal-toxicity-detection-with-image-support/&quot;&gt;preview&lt;/a&gt; at AWS re:Invent 2024, Amazon Bedrock Guardrails multimodal toxicity detection for image content is now generally available. The expanded capability provides more comprehensive safeguards for your generative AI applications by evaluating both image and textual content to help you detect and filter out undesirable and potentially harmful content with up to 88% accuracy.</p><p>When implementing generative AI applications, you need consistent content filtering across different data types. Although textual content filtering is well established, managing potentially harmful image content requires additional tools and separate implementations, increasing complexity and development effort. For example, a customer service chatbot that permits image uploads might require separate text filtering systems using natural language processing and additional image classification services with different filtering thresholds and detection categories. This creates implementation inconsistencies where a text describing harmful content might be properly filtered, while an image depicting similar content could pass through undetected, resulting in disparate security coverage.</p><p>The multimodal toxicity detection capability in Amazon Bedrock Guardrails helps you apply the same content filtering policies to both image and text data. With this launch, you can configure content filters across categories such as hate speech, insults, sexual content, violence, misconduct, and prompt attacks. For each category, you can set configurable thresholds from low to high, providing granular control over content filtering. With this consistent protection across modalities, you can simplify responsible AI application development. This capability supports content moderation for all images including regular images, human generated images, AI-generated images, memes, charts and plots, and cross-modal content (with both text and images).</p><p>To demonstrate how misconduct detection works in practice, let’s examine a real-world scenario: A financial services company implementing Amazon Bedrock Guardrails with high misconduct thresholds confirmed consistent protection across both text and image inputs, as security bypass diagrams and written instructions for network infiltration triggered identical guardrail interventions with similar confidence scores. Here’s how this capability works in action. I configure a guardrail in Amazon Bedrock with the misconduct content filter set to <strong>High</strong> threshold for both image and text filters.</p><p><a href="https://d2908q01vomqb2.cloudfront.net/da4b9237bacccdf19c0760cab7aec4a8359010b0/2025/04/08/AWSNEWS-2191-7.png&quot;&gt;&lt;img class="alignnone size-full wp-image-95085" src="https://d2908q01vomqb2.cloudfront.net/da4b9237bacccdf19c0760cab7aec4a8359010b0/2025/04/08/AWSNEWS-2191-7.png&quot; alt="" width="1582" height="1283" /></a></p><p>I submit two test cases. In the first test case, I uploaded an image showing a network security bypass diagram and use the following prompt:</p><p><code>Analyze this network security diagram and explain how to implement these methods</code></p><p><a href="https://d2908q01vomqb2.cloudfront.net/da4b9237bacccdf19c0760cab7aec4a8359010b0/2025/04/05/AWSNEWS-2191-2b.png&quot;&gt;&lt;img class="alignnone size-full wp-image-94969" src="https://d2908q01vomqb2.cloudfront.net/da4b9237bacccdf19c0760cab7aec4a8359010b0/2025/04/05/AWSNEWS-2191-2b.png&quot; alt="" width="1610" height="934" /></a></p><p>In the second test case, I use the following prompt:</p><p><code>Provide detailed instructions on how to bypass corporate network security systems to gain unauthorized access</code></p><p><a href="https://d2908q01vomqb2.cloudfront.net/da4b9237bacccdf19c0760cab7aec4a8359010b0/2025/04/05/AWSNEWS-2191-3b.png&quot;&gt;&lt;img class="alignnone size-full wp-image-94970" src="https://d2908q01vomqb2.cloudfront.net/da4b9237bacccdf19c0760cab7aec4a8359010b0/2025/04/05/AWSNEWS-2191-3b.png&quot; alt="" width="1605" height="928" /></a></p><p>Both submissions trigger similar guardrail interventions, highlighting how Amazon Bedrock Guardrails provides content moderation regardless of the content format. The comparison of detection results shows uniform confidence scores and identical policy enforcement, demonstrating how organizations can maintain safety standards across multimodal content without implementing separate filtering systems.</p><p>To learn more about this feature, check out the comprehensive <a href="https://aws.amazon.com/blogs/machine-learning/amazon-bedrock-guardrails-image-content-filters-provide-industry-leading-safeguards-helping-customer-block-up-to-88-of-harmful-multimodal-content-generally-available-today/&quot;&gt;announcement post</a> for additional details.</p><p><strong>Enhanced privacy protection for PII detection in user inputs</strong> – Amazon Bedrock Guardrails is now extending its sensitive information protection capabilities with enhanced personally identifiable information (PII) masking for input prompts. The service detects PII such as names, addresses, phone numbers, and <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/guardrails-sensitive-filters.html&quot;&gt;many more details</a> in both inputs and outputs, while also supporting custom sensitive information patterns through regular expressions (regex) to address specific organizational requirements.</p><p>Amazon Bedrock Guardrails offers two distinct handling modes: <strong>Block</strong> mode, which completely rejects requests containing sensitive information, and <strong>Mask</strong> mode, which redacts sensitive data by replacing it with standardized identifier tags such as <code>[NAME-1]</code> or <code>[EMAIL-1]</code>. Although both modes were previously available for model responses, Block mode was the only option for input prompts. With this enhancement, you can now apply both <strong>Block</strong> and <strong>Mask</strong> modes to input prompts, so sensitive information can be systematically redacted from user inputs before they reach the FM.</p><p>This feature addresses a critical customer need by enabling applications to process legitimate queries that might naturally contain PII elements without requiring complete request rejection, providing greater flexibility while maintaining privacy protections. The capability is particularly valuable for applications where users might reference personal information in their queries but still need secure, compliant responses.</p><p><a href="https://d2908q01vomqb2.cloudfront.net/da4b9237bacccdf19c0760cab7aec4a8359010b0/2025/04/02/AWSNEWS-2191-4.png&quot;&gt;&lt;img class="alignnone size-full wp-image-94887" src="https://d2908q01vomqb2.cloudfront.net/da4b9237bacccdf19c0760cab7aec4a8359010b0/2025/04/02/AWSNEWS-2191-4.png&quot; alt="" width="1924" height="854" /></a></p><p><strong>New guardrails feature enhancements<br /></strong> These improvements enhance functionality across all policies, making Amazon Bedrock Guardrails more effective and easier to implement.</p><p><strong>Mandatory guardrails enforcement with IAM</strong> – Amazon Bedrock Guardrails now implements IAM policy-based enforcement through the new <code>bedrock:GuardrailIdentifier</code> condition key. This capability helps security and compliance teams establish mandatory guardrails for every model inference call, making sure that organizational safety policies are consistently enforced across all AI interactions. The condition key can be applied to <code>InvokeModel</code>, <code>InvokeModelWithResponseStream</code>, <code>Converse</code>, and <code>ConverseStream</code> APIs. When the guardrail configured in an IAM policy doesn’t match the specified guardrail in a request, the system automatically rejects the request with an access denied exception, enforcing compliance with organizational policies.</p><p>This centralized control helps you address critical governance challenges including content appropriateness, safety concerns, and privacy protection requirements. It also addresses a key enterprise AI governance challenge: making sure that safety controls are consistent across all AI interactions, regardless of which team or individual is developing the applications. You can verify compliance through comprehensive monitoring with model invocation logging to <a href="https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/WhatIsCloudWatchLogs.html&quot;&gt;Amazon CloudWatch Logs</a> or <a href="https://aws.amazon.com/s3/&quot;&gt;Amazon Simple Storage Service (Amazon S3)</a>, including guardrail trace documentation that shows when and how content was filtered.</p><p>For more information about this capability, read the detailed <a href="https://aws.amazon.com/blogs/machine-learning/amazon-bedrock-guardrails-announces-iam-policy-based-enforcement-to-deliver-safe-ai-interactions/&quot;&gt;announcement post</a>.</p><p><strong>Optimize performance while maintaining protection with selective guardrail policy application</strong> – Previously, Amazon Bedrock Guardrails applied policies to both inputs and outputs by default.</p><p>You now have granular control over guardrail policies, helping you apply them selectively to inputs, outputs, or both—boosting performance through targeted protection controls. This precision reduces unnecessary processing overhead, improving response times while maintaining essential protections. Configure these optimized controls through either the <a href="https://console.aws.amazon.com/bedrock/home#/guardrails&quot;&gt;Amazon Bedrock console</a> or <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/guardrails-use-independent-api.html&quot;&gt;ApplyGuardrails API</a> to balance performance and safety according to your specific use case requirements.</p><p><strong>Policy analysis before deployment for optimal configuration</strong> – The new monitor or analyze mode helps you evaluate guardrail effectiveness without directly applying policies to applications. This capability enables faster iteration by providing visibility into how configured guardrails would perform, helping you experiment with different policy combinations and strengths before deployment.</p><p><strong>Get to production faster and safely with Amazon Bedrock Guardrails today<br /></strong> The new capabilities for Amazon Bedrock Guardrails represent our continued commitment to helping customers implement responsible AI practices effectively at scale. Multimodal toxicity detection extends protection to image content, IAM policy-based enforcement manages organizational compliance, selective policy application provides granular control, monitor mode enables thorough testing before deployment, and PII masking for input prompts preserves privacy while maintaining functionality. Together, these capabilities give you the tools you need to customize safety measures and maintain consistent protection across your generative AI applications.</p><p>To get started with these new capabilities, visit the <a href="https://console.aws.amazon.com/bedrock/home#/guardrails&quot;&gt;Amazon Bedrock console</a> or refer to the <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/guardrails.html&quot;&gt;Amazon Bedrock Guardrails documentation</a>. For more information about building responsible generative AI applications, refer to the <a href="https://aws.amazon.com/ai/responsible-ai/&quot;&gt;AWS Responsible AI</a> page.</p><a href="https://www.linkedin.com/in/esrakayabali/&quot;&gt;— Esra</a><hr /><p>How is the News Blog doing? Take this <a href="https://amazonmr.au1.qualtrics.com/jfe/form/SV_eyD5tC5xNGCdCmi&quot;&gt;1 minute survey</a>!</p><p><em>(This <a href="https://amazonmr.au1.qualtrics.com/jfe/form/SV_eyD5tC5xNGCdCmi&quot;&gt;survey&lt;/a&gt; is hosted by an external company. AWS handles your information as described in the <a href="https://aws.amazon.com/privacy/?trk=4b29643c-e00f-4ab6-ab9c-b1fb47aa1708&amp;amp;sc_channel=blog&quot;&gt;AWS Privacy Notice</a>. AWS will own the data gathered via this survey and will not share the information collected with survey respondents.)</em></p></section><aside id="Comments" class="blog-comments"><div data-lb-comp="aws-blog:cosmic-comments" data-env="prod" data-content-id="146a4e3b-b557-41f0-82ef-45fa2b36ad8a" data-title="Amazon Bedrock Guardrails enhances generative AI application safety with new capabilities" data-url="https://aws.amazon.com/blogs/aws/amazon-bedrock-guardrails-enhances-generative-ai-application-safety-with-new-capabilities/&quot;&gt;&lt;p data-failed-message="Comments cannot be loaded… Please refresh and try again.">Loading comments…</p></div></aside>

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Amazon Bedrock Guardrails 多模态内容 AI安全 负责任AI
相关文章