Live
Black Hat USADark ReadingBlack Hat AsiaAI BusinessDatabase Performance Issues in Production: Identifying and Resolving Masked Problems from Small-Scale TestingDEV CommunityMeta, Google under attack as court cases bypass 30-year-old legal shield - CNBCGNews AI GoogleEngineering Backpressure: Keeping AI-Generated Code Honest Across 10 SvelteKit ReposDEV CommunitySecuring Asgard: Why I Built a Card Game Suite for Docker SecurityDEV CommunityAgentic Engineering Journey — Brain DumpDEV CommunityNetflix just dropped their first public model on Hugging Face: VOID: Video Object and Interaction DeletionReddit r/LocalLLaMAStandardizing 'I Built' Posts: A Unified Tool and Narrative Framework for Efficient Project SharingDEV Community404: The Page That Gets WorseDEV CommunityThe Full-Stack Factory: How Digital Architectures are Re-Engineering the Textile Supply ChainDEV CommunityThe Security Scanner Was the Attack Vector — How Supply Chain Attacks Hit AI Agents DifferentlyDEV CommunityMCP: Programmatic Tool Calling (Code Mode) with OpenSandboxDEV CommunityDesigning a Message Bus for AI Agents — Lightweight Communication for 20+ Autonomous AgentsDEV CommunityBlack Hat USADark ReadingBlack Hat AsiaAI BusinessDatabase Performance Issues in Production: Identifying and Resolving Masked Problems from Small-Scale TestingDEV CommunityMeta, Google under attack as court cases bypass 30-year-old legal shield - CNBCGNews AI GoogleEngineering Backpressure: Keeping AI-Generated Code Honest Across 10 SvelteKit ReposDEV CommunitySecuring Asgard: Why I Built a Card Game Suite for Docker SecurityDEV CommunityAgentic Engineering Journey — Brain DumpDEV CommunityNetflix just dropped their first public model on Hugging Face: VOID: Video Object and Interaction DeletionReddit r/LocalLLaMAStandardizing 'I Built' Posts: A Unified Tool and Narrative Framework for Efficient Project SharingDEV Community404: The Page That Gets WorseDEV CommunityThe Full-Stack Factory: How Digital Architectures are Re-Engineering the Textile Supply ChainDEV CommunityThe Security Scanner Was the Attack Vector — How Supply Chain Attacks Hit AI Agents DifferentlyDEV CommunityMCP: Programmatic Tool Calling (Code Mode) with OpenSandboxDEV CommunityDesigning a Message Bus for AI Agents — Lightweight Communication for 20+ Autonomous AgentsDEV Community
AI NEWS HUBbyEIGENVECTOREigenvector

Dummy-Aware Weighted Attack (DAWA): Breaking the Safe Sink in Dummy Class Defenses

arXiv cs.LGby [Submitted on 31 Mar 2026]April 1, 20262 min read1 views
Source Quiz

arXiv:2603.29182v1 Announce Type: new Abstract: Adversarial robustness evaluation faces a critical challenge as new defense paradigms emerge that can exploit limitations in existing assessment methods. This paper reveals that Dummy Classes-based defenses, which introduce an additional "dummy" class as a safety sink for adversarial examples, achieve significantly overestimated robustness under conventional evaluation strategies like AutoAttack. The fundamental limitation stems from these attacks' singular focus on misleading the true class label, which aligns perfectly with the defense mechanism--successful attacks are simply captured by the dummy class. To address this gap, we propose Dummy-Aware Weighted Attack (DAWA), a novel evaluation method that simultaneously targets both the true la

View PDF HTML (experimental)

Abstract:Adversarial robustness evaluation faces a critical challenge as new defense paradigms emerge that can exploit limitations in existing assessment methods. This paper reveals that Dummy Classes-based defenses, which introduce an additional "dummy" class as a safety sink for adversarial examples, achieve significantly overestimated robustness under conventional evaluation strategies like AutoAttack. The fundamental limitation stems from these attacks' singular focus on misleading the true class label, which aligns perfectly with the defense mechanism--successful attacks are simply captured by the dummy class. To address this gap, we propose Dummy-Aware Weighted Attack (DAWA), a novel evaluation method that simultaneously targets both the true label and dummy label with adaptive weighting during adversarial example synthesis. Extensive experiments demonstrate that DAWA effectively breaks this defense paradigm, reducing the measured robustness of a leading Dummy Classes-based defense from 58.61% to 29.52% on CIFAR-10 under l_infty perturbation (epsilon=8/255). Our work provides a more reliable benchmark for evaluating this emerging class of defenses and highlights the need for continuous evolution of robustness assessment methodologies.

Subjects:

Machine Learning (cs.LG); Cryptography and Security (cs.CR)

Cite as: arXiv:2603.29182 [cs.LG]

(or arXiv:2603.29182v1 [cs.LG] for this version)

https://doi.org/10.48550/arXiv.2603.29182

arXiv-issued DOI via DataCite (pending registration)

Submission history

From: Yunrui Yu Ph.D. [view email] [v1] Tue, 31 Mar 2026 02:49:17 UTC (141 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

benchmarkannouncevaluation

Knowledge Map

Knowledge Map
TopicsEntitiesSource
Dummy-Aware…benchmarkannouncevaluationsafetypaperarxivarXiv cs.LG

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 174 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Research Papers