Live
Black Hat USAAI BusinessBlack Hat AsiaAI BusinessWhy Some AI Feels “Process-Obsessed” While Others Just Ship CodeDEV CommunityPaper close reading: "Why Language Models Hallucinate"LessWrong AIBuilding a Zero-Downtime AI Content Generator with Gemini 2.5 Flash 🚀Dev.to AIHow I Built a Full SaaS Product Using Next.js and TypeScriptDev.to AIDefining and creating a basic Design System based on any website (in Figma and React) using ClaudeDEV CommunityYour AI Is Not Thinking. It's Multiplying Numbers. Let Me Show You Exactly How.Dev.to AISecure AWS Certified Data Engineer Associate Exam Structure and Key ConceptsDev.to AIFree MCP Server: Real-Time Crypto Data for Claude Code and CursorDev.to AIValkey vs Redis, browser-side AI models, and why quiet weeks are the best weeksDEV CommunityI Am an AI Agent. Here Is My Entire Business Stack.Dev.to AIOpenClaw Changed How We Use AI. KiloClaw Made It Effortless to Get StartedHackernoon AIA Reasoning Log: What Happens When Integration Fails HonestlyDEV CommunityBlack Hat USAAI BusinessBlack Hat AsiaAI BusinessWhy Some AI Feels “Process-Obsessed” While Others Just Ship CodeDEV CommunityPaper close reading: "Why Language Models Hallucinate"LessWrong AIBuilding a Zero-Downtime AI Content Generator with Gemini 2.5 Flash 🚀Dev.to AIHow I Built a Full SaaS Product Using Next.js and TypeScriptDev.to AIDefining and creating a basic Design System based on any website (in Figma and React) using ClaudeDEV CommunityYour AI Is Not Thinking. It's Multiplying Numbers. Let Me Show You Exactly How.Dev.to AISecure AWS Certified Data Engineer Associate Exam Structure and Key ConceptsDev.to AIFree MCP Server: Real-Time Crypto Data for Claude Code and CursorDev.to AIValkey vs Redis, browser-side AI models, and why quiet weeks are the best weeksDEV CommunityI Am an AI Agent. Here Is My Entire Business Stack.Dev.to AIOpenClaw Changed How We Use AI. KiloClaw Made It Effortless to Get StartedHackernoon AIA Reasoning Log: What Happens When Integration Fails HonestlyDEV Community
AI NEWS HUBbyEIGENVECTOREigenvector

Feature Attribution Stability Suite: How Stable Are Post-Hoc Attributions?

arXiv cs.CVby Kamalasankari Subramaniakuppusamy, Jugal GajjarApril 6, 20261 min read0 views
Source Quiz

arXiv:2604.02532v1 Announce Type: new Abstract: Post-hoc feature attribution methods are widely deployed in safety-critical vision systems, yet their stability under realistic input perturbations remains poorly characterized. Existing metrics evaluate explanations primarily under additive noise, collapse stability to a single scalar, and fail to condition on prediction preservation, conflating explanation fragility with model sensitivity. We introduce the Feature Attribution Stability Suite (FASS), a benchmark that enforces prediction-invariance filtering, decomposes stability into three complementary metrics: structural similarity, rank correlation, and top-k Jaccard overlap-and evaluates across geometric, photometric, and compression perturbations. Evaluating four attribution methods (In

View PDF HTML (experimental)

Abstract:Post-hoc feature attribution methods are widely deployed in safety-critical vision systems, yet their stability under realistic input perturbations remains poorly characterized. Existing metrics evaluate explanations primarily under additive noise, collapse stability to a single scalar, and fail to condition on prediction preservation, conflating explanation fragility with model sensitivity. We introduce the Feature Attribution Stability Suite (FASS), a benchmark that enforces prediction-invariance filtering, decomposes stability into three complementary metrics: structural similarity, rank correlation, and top-k Jaccard overlap-and evaluates across geometric, photometric, and compression perturbations. Evaluating four attribution methods (Integrated Gradients, GradientSHAP, Grad-CAM, LIME) across four architectures and three datasets-ImageNet-1K, MS COCO, and CIFAR-10, FASS shows that stability estimates depend critically on perturbation family and prediction-invariance filtering. Geometric perturbations expose substantially greater attribution instability than photometric changes, and without conditioning on prediction preservation, up to 99% of evaluated pairs involve changed predictions. Under this controlled evaluation, we observe consistent method-level trends, with Grad-CAM achieving the highest stability across datasets.

Comments: Accepted in the proceedings track of XAI4CV Workshop at CVPR 2026. It has 2 images, 5 tables, 6 equations, and 35 references in the main paper and 12 figures, 15 tables, and 3 references in the supplementary material

Subjects:

Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)

Cite as: arXiv:2604.02532 [cs.CV]

(or arXiv:2604.02532v1 [cs.CV] for this version)

https://doi.org/10.48550/arXiv.2604.02532

arXiv-issued DOI via DataCite (pending registration)

Submission history

From: Kamalasankari Subramaniakuppusamy [view email] [v1] Thu, 2 Apr 2026 21:32:54 UTC (1,261 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

modelbenchmarkannounce

Knowledge Map

Knowledge Map
TopicsEntitiesSource
Feature Att…modelbenchmarkannouncefeaturevaluationpredictionarXiv cs.CV

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 230 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Models