Live
Black Hat USADark ReadingBlack Hat AsiaAI BusinessAI, Price Theory, and the Future of Economics ResearchHacker News AI TopShow HN: EU Compliance SaaS for Sale ($4K Each) – CBAM, AI Act, Public TendersHacker News AI TopShow HN: Filoxenia – open protocol for human-AI companionshipHacker News AI TopShow HN: AI agent skills for affiliate marketing (Markdown, works with any LLM)Hacker News AI TopMeta Pauses Work With Mercor After Data Breach Puts AI Industry Secrets at RiskWired AI"Cognitive surrender" leads AI users to abandon logical thinking, research findsHacker News AI TopUsing multiple AI agents as an architectural review councilHacker News AI TopAES Maximo robot installs 100 megawatts of solar capacityThe Robot Reportb8653llama.cpp Releases5 Backend Concepts You Shouldn’t IgnoreTowards AIAI's Next Frontier: Insights from Jeff Dean and Bill Dally InHacker News AI TopGoogle launches Gemma 4, an enterprise-grade open source AI model set - CIO DiveGNews AI GemmaBlack Hat USADark ReadingBlack Hat AsiaAI BusinessAI, Price Theory, and the Future of Economics ResearchHacker News AI TopShow HN: EU Compliance SaaS for Sale ($4K Each) – CBAM, AI Act, Public TendersHacker News AI TopShow HN: Filoxenia – open protocol for human-AI companionshipHacker News AI TopShow HN: AI agent skills for affiliate marketing (Markdown, works with any LLM)Hacker News AI TopMeta Pauses Work With Mercor After Data Breach Puts AI Industry Secrets at RiskWired AI"Cognitive surrender" leads AI users to abandon logical thinking, research findsHacker News AI TopUsing multiple AI agents as an architectural review councilHacker News AI TopAES Maximo robot installs 100 megawatts of solar capacityThe Robot Reportb8653llama.cpp Releases5 Backend Concepts You Shouldn’t IgnoreTowards AIAI's Next Frontier: Insights from Jeff Dean and Bill Dally InHacker News AI TopGoogle launches Gemma 4, an enterprise-grade open source AI model set - CIO DiveGNews AI Gemma
AI NEWS HUBbyEIGENVECTOREigenvector

An Isotropic Approach to Efficient Uncertainty Quantification with Gradient Norms

arXiv cs.LGby Nils Gr\"unefeld, Jes Frellsen, Christian HardmeierApril 1, 20262 min read0 views
Source Quiz

arXiv:2603.29466v1 Announce Type: new Abstract: Existing methods for quantifying predictive uncertainty in neural networks are either computationally intractable for large language models or require access to training data that is typically unavailable. We derive a lightweight alternative through two approximations: a first-order Taylor expansion that expresses uncertainty in terms of the gradient of the prediction and the parameter covariance, and an isotropy assumption on the parameter covariance. Together, these yield epistemic uncertainty as the squared gradient norm and aleatoric uncertainty as the Bernoulli variance of the point prediction, from a single forward-backward pass through an unmodified pretrained model. We justify the isotropy assumption by showing that covariance estimat

View PDF HTML (experimental)

Abstract:Existing methods for quantifying predictive uncertainty in neural networks are either computationally intractable for large language models or require access to training data that is typically unavailable. We derive a lightweight alternative through two approximations: a first-order Taylor expansion that expresses uncertainty in terms of the gradient of the prediction and the parameter covariance, and an isotropy assumption on the parameter covariance. Together, these yield epistemic uncertainty as the squared gradient norm and aleatoric uncertainty as the Bernoulli variance of the point prediction, from a single forward-backward pass through an unmodified pretrained model. We justify the isotropy assumption by showing that covariance estimates built from non-training data introduce structured distortions that isotropic covariance avoids, and that theoretical results on the spectral properties of large networks support the approximation at scale. Validation against reference Markov Chain Monte Carlo estimates on synthetic problems shows strong correspondence that improves with model size. We then use the estimates to investigate when each uncertainty type carries useful signal for predicting answer correctness in question answering with large language models, revealing a benchmark-dependent divergence: the combined estimate achieves the highest mean AUROC on TruthfulQA, where questions involve genuine conflict between plausible answers, but falls to near chance on TriviaQA's factual recall, suggesting that parameter-level uncertainty captures a fundamentally different signal than self-assessment methods.

Subjects:

Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL)

Cite as: arXiv:2603.29466 [cs.LG]

(or arXiv:2603.29466v1 [cs.LG] for this version)

https://doi.org/10.48550/arXiv.2603.29466

arXiv-issued DOI via DataCite (pending registration)

Submission history

From: Nils Grünefeld [view email] [v1] Tue, 31 Mar 2026 09:13:09 UTC (939 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

Knowledge Map

Knowledge Map
TopicsEntitiesSource
An Isotropi…modellanguage mo…neural netw…benchmarktrainingannouncearXiv cs.LG

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 127 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Models