Live
Black Hat USADark ReadingBlack Hat AsiaAI BusinessBillionaire Philippe Laffont Sold CoreWeave and Bought This Artificial Intelligence (AI) Stock Instead - The Motley FoolGoogle News: AITSMC vs. Nvidia: Which AI Supercycle Growth Stock Is the Better Long-Term Buy? - AOL.comGNews AI NVIDIAGenerative AI Business Use Cases 2026: The 11 Applications Delivering Real ROI - BBN TimesGoogle News: Generative AITSMC vs. Nvidia: Which AI Supercycle Growth Stock Is the Better Long-Term Buy? - Yahoo FinanceGNews AI NVIDIA[P] MCGrad: fix calibration of your ML model in subgroupsReddit r/MachineLearningTSMC vs. Nvidia: Which AI Supercycle Growth Stock Is the Better Long-Term Buy? - The Motley FoolGNews AI NVIDIANVIDIA’s US$2b Marvell Bet Extends AI Platform Into Telecom Networks - simplywall.stGNews AI NVIDIAGemma4 26B A4B runs easily on 16GB MacsReddit r/LocalLLaMAMACHINE LEARNING EXPERT WITNESS TESTIMONY FOR ML & AI TRIAL TESTIFYING NEEDS - futuristsspeakers.comGoogle News: Machine LearningThe AI Doc Explores Promise and Peril of Artificial Intelligence - Metro WeeklyGNews AI AGIIran says Iraqi ships are allowed to use Strait of HormuzFortune TechRyan Coogler Is Shifting Animorphs Into a Disney+ ShowGizmodoBlack Hat USADark ReadingBlack Hat AsiaAI BusinessBillionaire Philippe Laffont Sold CoreWeave and Bought This Artificial Intelligence (AI) Stock Instead - The Motley FoolGoogle News: AITSMC vs. Nvidia: Which AI Supercycle Growth Stock Is the Better Long-Term Buy? - AOL.comGNews AI NVIDIAGenerative AI Business Use Cases 2026: The 11 Applications Delivering Real ROI - BBN TimesGoogle News: Generative AITSMC vs. Nvidia: Which AI Supercycle Growth Stock Is the Better Long-Term Buy? - Yahoo FinanceGNews AI NVIDIA[P] MCGrad: fix calibration of your ML model in subgroupsReddit r/MachineLearningTSMC vs. Nvidia: Which AI Supercycle Growth Stock Is the Better Long-Term Buy? - The Motley FoolGNews AI NVIDIANVIDIA’s US$2b Marvell Bet Extends AI Platform Into Telecom Networks - simplywall.stGNews AI NVIDIAGemma4 26B A4B runs easily on 16GB MacsReddit r/LocalLLaMAMACHINE LEARNING EXPERT WITNESS TESTIMONY FOR ML & AI TRIAL TESTIFYING NEEDS - futuristsspeakers.comGoogle News: Machine LearningThe AI Doc Explores Promise and Peril of Artificial Intelligence - Metro WeeklyGNews AI AGIIran says Iraqi ships are allowed to use Strait of HormuzFortune TechRyan Coogler Is Shifting Animorphs Into a Disney+ ShowGizmodo
AI NEWS HUBbyEIGENVECTOREigenvector

Detecting Call Graph Unsoundness without Ground Truth

arXiv cs.SEby Fangtian Zhong, Ollie Wold, Joseph WindmannApril 2, 20261 min read0 views
Source Quiz

arXiv:2604.00885v1 Announce Type: new Abstract: Java static analysis frameworks are commonly compared under the assumption that analysis algorithms and configurations compose monotonically and yield semantically comparable results across tools. In this work, we show that this assumption is fundamentally flawed. We present a large-scale empirical study of semantic consistency within and across four widely used Java static analysis frameworks: Soot, SootUp, WALA, and Doop. Using precision partial orders over analysis algorithms and configurations, we systematically identify violations where increased precision introduces new call-graph edges or amplifies inconsistencies. Our results reveal three key findings. First, algorithmic precision orders frequently break within frameworks due to moder

View PDF HTML (experimental)

Abstract:Java static analysis frameworks are commonly compared under the assumption that analysis algorithms and configurations compose monotonically and yield semantically comparable results across tools. In this work, we show that this assumption is fundamentally flawed. We present a large-scale empirical study of semantic consistency within and across four widely used Java static analysis frameworks: Soot, SootUp, WALA, and Doop. Using precision partial orders over analysis algorithms and configurations, we systematically identify violations where increased precision introduces new call-graph edges or amplifies inconsistencies. Our results reveal three key findings. First, algorithmic precision orders frequently break within frameworks due to modern language features such as lambdas, reflection, and native modeling. Second, configuration choices strongly interact with analysis algorithms, producing synergistic failures that exceed the effects of algorithm or configuration changes alone. Third, cross-framework comparisons expose irreconcilable semantic gaps, demonstrating that different frameworks operate over incompatible notions of call-graph ground truth. These findings challenge prevailing evaluation practices in static analysis and highlight the need to reason jointly about algorithms, configurations, and framework semantics when assessing precision and soundness.

Subjects:

Software Engineering (cs.SE)

Cite as: arXiv:2604.00885 [cs.SE]

(or arXiv:2604.00885v1 [cs.SE] for this version)

https://doi.org/10.48550/arXiv.2604.00885

arXiv-issued DOI via DataCite (pending registration)

Submission history

From: Fangtian Zhong [view email] [v1] Wed, 1 Apr 2026 13:32:53 UTC (209 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

Knowledge Map

Knowledge Map
TopicsEntitiesSource
Detecting C…modelannouncefeaturevaluationanalysisstudyarXiv cs.SE

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 184 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Research Papers