热点
"对抗攻击" 相关文章
Theoretical Analysis of Relative Errors in Gradient Computations for Adversarial Attacks with CE Loss
cs.AI updates on arXiv.org 2025-07-31T04:48:07.000000Z
RCR-AF: Enhancing Model Generalization via Rademacher Complexity Reduction Activation Function
cs.AI updates on arXiv.org 2025-07-31T04:48:07.000000Z
智能体安全与可信AI:防护机制与伦理考量
掘金 人工智能 2025-07-29T04:21:08.000000Z
Reinforced Embodied Active Defense: Exploiting Adaptive Interaction for Robust Visual Perception in Adversarial 3D Environments
cs.AI updates on arXiv.org 2025-07-25T04:28:54.000000Z
Evaluating the Performance of AI Text Detectors, Few-Shot and Chain-of-Thought Prompting Using DeepSeek Generated Text
cs.AI updates on arXiv.org 2025-07-25T04:28:41.000000Z
From Seed to Harvest: Augmenting Human Creativity with AI for Red-teaming Text-to-Image Models
cs.AI updates on arXiv.org 2025-07-25T04:28:40.000000Z
Advancing Robustness in Deep Reinforcement Learning with an Ensemble Defense Approach
cs.AI updates on arXiv.org 2025-07-24T05:31:11.000000Z
VTarbel: Targeted Label Attack with Minimal Knowledge on Detector-enhanced Vertical Federated Learning
cs.AI updates on arXiv.org 2025-07-22T04:44:37.000000Z
Breaking the Illusion of Security via Interpretation: Interpretable Vision Transformer Systems under Attack
cs.AI updates on arXiv.org 2025-07-22T04:34:48.000000Z
Scaling Decentralized Learning with FLock
cs.AI updates on arXiv.org 2025-07-22T04:34:25.000000Z
Non-Adaptive Adversarial Face Generation
cs.AI updates on arXiv.org 2025-07-17T04:14:47.000000Z
Crafting Imperceptible On-Manifold Adversarial Attacks for Tabular Data
cs.AI updates on arXiv.org 2025-07-16T04:29:04.000000Z
False Alarms, Real Damage: Adversarial Attacks Using LLM-based Models on Text-based Cyber Threat Intelligence Systems
cs.AI updates on arXiv.org 2025-07-10T04:05:35.000000Z
DATABench: Evaluating Dataset Auditing in Deep Learning from an Adversarial Perspective
cs.AI updates on arXiv.org 2025-07-09T04:01:45.000000Z
Evaluating the Evaluators: Trust in Adversarial Robustness Tests
cs.AI updates on arXiv.org 2025-07-08T06:58:07.000000Z
Concept-based Adversarial Attack: a Probabilistic Perspective
cs.AI updates on arXiv.org 2025-07-08T05:53:52.000000Z
Probing Latent Subspaces in LLM for AI Security: Identifying and Manipulating Adversarial States
cs.AI updates on arXiv.org 2025-07-08T04:33:43.000000Z
Smaller = Weaker? Benchmarking Robustness of Quantized LLMs in Code Generation
cs.AI updates on arXiv.org 2025-07-01T04:13:55.000000Z
Vulnerability in Trusted Monitoring and Mitigations
少点错误 2025-06-07T15:07:36.000000Z
深度研究 | Agentic AI系统安全防护实施指南
安全牛 2025-04-29T10:20:29.000000Z