Live
Black Hat USADark ReadingBlack Hat AsiaAI BusinessI tested Gemini on Android Auto and now I can't stop talking to it: 5 tasks it nailsZDNet AIBuilding Intelligent Search with Amazon Bedrock and Amazon OpenSearch for hybrid RAG solutionsAWS Machine Learning BlogFrom isolated alerts to contextual intelligence: Agentic maritime anomaly analysis with generative AIAWS Machine Learning Blog🔥 ggml-org/llama.cppGitHub Trending🔥 ollama/ollamaGitHub Trending🔥 sponsors/kepanoGitHub Trending🔥 KeygraphHQ/shannonGitHub Trending🔥 sponsors/abhigyanpatwariGitHub TrendingOpenAI Releases Policy Recommendations for AI AgeBloomberg TechnologyBeware the Magical 2-Person, $1 Billion AI-Driven StartupForrester AI Blog[D] ICML 26 - What to do with the zero follow-up questionsReddit r/MachineLearningStop Writing Mega-Prompts: Use These 5 Anthropic Design Patterns InsteadMedium AIBlack Hat USADark ReadingBlack Hat AsiaAI BusinessI tested Gemini on Android Auto and now I can't stop talking to it: 5 tasks it nailsZDNet AIBuilding Intelligent Search with Amazon Bedrock and Amazon OpenSearch for hybrid RAG solutionsAWS Machine Learning BlogFrom isolated alerts to contextual intelligence: Agentic maritime anomaly analysis with generative AIAWS Machine Learning Blog🔥 ggml-org/llama.cppGitHub Trending🔥 ollama/ollamaGitHub Trending🔥 sponsors/kepanoGitHub Trending🔥 KeygraphHQ/shannonGitHub Trending🔥 sponsors/abhigyanpatwariGitHub TrendingOpenAI Releases Policy Recommendations for AI AgeBloomberg TechnologyBeware the Magical 2-Person, $1 Billion AI-Driven StartupForrester AI Blog[D] ICML 26 - What to do with the zero follow-up questionsReddit r/MachineLearningStop Writing Mega-Prompts: Use These 5 Anthropic Design Patterns InsteadMedium AI
AI NEWS HUBbyEIGENVECTOREigenvector

Inversion-Free Natural Gradient Descent on Riemannian Manifolds

arXiv stat.MLby [Submitted on 3 Apr 2026]April 6, 20262 min read1 views
Source Quiz

arXiv:2604.02969v1 Announce Type: new Abstract: The natural gradient method is widely used in statistical optimization, but its standard formulation assumes a Euclidean parameter space. This paper proposes an inversion-free stochastic natural gradient method for probability distributions whose parameters lie on a Riemannian manifold. The manifold setting offers several advantages: one can implicitly enforce parameter constraints such as positive definiteness and orthogonality, ensure parameters are identifiable, or guarantee regularity properties of the objective like geodesic convexity. Building on an intrinsic formulation of the Fisher information matrix (FIM) on a manifold, our method maintains an online approximation of the inverse FIM, which is efficiently updated at quadratic cost us

View PDF HTML (experimental)

Abstract:The natural gradient method is widely used in statistical optimization, but its standard formulation assumes a Euclidean parameter space. This paper proposes an inversion-free stochastic natural gradient method for probability distributions whose parameters lie on a Riemannian manifold. The manifold setting offers several advantages: one can implicitly enforce parameter constraints such as positive definiteness and orthogonality, ensure parameters are identifiable, or guarantee regularity properties of the objective like geodesic convexity. Building on an intrinsic formulation of the Fisher information matrix (FIM) on a manifold, our method maintains an online approximation of the inverse FIM, which is efficiently updated at quadratic cost using score vectors sampled at successive iterates. In the Riemannian setting, these score vectors belong to different tangent spaces and must be combined using transport operations. We prove almost-sure convergence rates of $O(\log{s}/s^\alpha)$ for the squared distance to the minimizer when the step size exponent $\alpha >2/3$. We also establish almost-sure rates for the approximate FIM, which now accumulates transport-based errors. A limited-memory variant of the algorithm with sub-quadratic storage complexity is proposed. Finally, we demonstrate the effectiveness of our method relative to its Euclidean counterparts on variational Bayes with Gaussian approximations and normalizing flows.

Comments: 73 pages, 3 figures

Subjects:

Machine Learning (stat.ML); Machine Learning (cs.LG); Computation (stat.CO); Methodology (stat.ME)

Cite as: arXiv:2604.02969 [stat.ML]

(or arXiv:2604.02969v1 [stat.ML] for this version)

https://doi.org/10.48550/arXiv.2604.02969

arXiv-issued DOI via DataCite (pending registration)

Submission history

From: Minh-Ngoc Tran [view email] [v1] Fri, 3 Apr 2026 11:08:59 UTC (917 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

announceversionupdate

Knowledge Map

Knowledge Map
TopicsEntitiesSource
Inversion-F…announceversionupdatepaperarxivarXiv stat.…

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 224 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Releases