Live
Black Hat USADark ReadingBlack Hat AsiaAI BusinessRAG Web Browser: Give Your AI Real-Time Web Access Without HallucinationsDEV CommunityWhat Nobody Tells You About Building a Protocol for AI AgentsDEV CommunityThe Evidence Is in the Phone. Most of It Never Makes It Into the Case.DEV CommunityIt's Not Smarter Models — It's Cheaper Memory: TurboQuant's Real Impact, Wall Street Panic & Academic StormDEV CommunityYour Production Code Is Training AI Models Right Now (And How to Audit Your Stack)DEV CommunitySetting Up Your Databricks Account (Free Trial + First Look at the UI)DEV CommunityHow to Use Augmented Coding to Build a Web AppDEV CommunitySetting up a hugo static site hosted with PorkbunDEV CommunityWhy Domain Knowledge Is the Core Architecture of Fine-Tuning and RAG — Not an AfterthoughtDEV CommunityComo o Ataque à Cadeia de Suprimentos do NPM Axios Acontece (E Como Proteger Seus Projetos de API)DEV CommunityAI Impact: Focus on Clarity, Results & Sophistication - FTI ConsultingGoogle News: AIClaude Code Source Code "Rebranded" Amid Wild Web Cloning, Anthropic's Blocking Attempt Fails - 36 KrGoogle News: ClaudeBlack Hat USADark ReadingBlack Hat AsiaAI BusinessRAG Web Browser: Give Your AI Real-Time Web Access Without HallucinationsDEV CommunityWhat Nobody Tells You About Building a Protocol for AI AgentsDEV CommunityThe Evidence Is in the Phone. Most of It Never Makes It Into the Case.DEV CommunityIt's Not Smarter Models — It's Cheaper Memory: TurboQuant's Real Impact, Wall Street Panic & Academic StormDEV CommunityYour Production Code Is Training AI Models Right Now (And How to Audit Your Stack)DEV CommunitySetting Up Your Databricks Account (Free Trial + First Look at the UI)DEV CommunityHow to Use Augmented Coding to Build a Web AppDEV CommunitySetting up a hugo static site hosted with PorkbunDEV CommunityWhy Domain Knowledge Is the Core Architecture of Fine-Tuning and RAG — Not an AfterthoughtDEV CommunityComo o Ataque à Cadeia de Suprimentos do NPM Axios Acontece (E Como Proteger Seus Projetos de API)DEV CommunityAI Impact: Focus on Clarity, Results & Sophistication - FTI ConsultingGoogle News: AIClaude Code Source Code "Rebranded" Amid Wild Web Cloning, Anthropic's Blocking Attempt Fails - 36 KrGoogle News: Claude

Extending Mean-Field Variational Inference via Entropic Regularization: Theory and Computation

JMLRby Bohan Wu, David M. BleiJanuary 1, 20261 min read0 views
Source Quiz

Variational inference (VI) has emerged as a popular method for approximate inference for high-dimensional Bayesian models. In this paper, we propose a novel VI method that extends the naive mean field via entropic regularization, referred to as $\Xi$-variational inference ($\Xi$-VI). $\Xi$-VI has a close connection to the entropic optimal transport problem and benefits from the computationally efficient Sinkhorn algorithm. We show that $\Xi$-variational posteriors effectively recover the true posterior dependency, where the likelihood function is downweighted by a regularization parameter. We analyze the role of dimensionality of the parameter space on the accuracy of $\Xi$-variational approximation and the computational complexity of computing the approximate distribution, providing a rou

Bohan Wu, David M. Blei; 27(7):1−68, 2026.

Abstract

Variational inference (VI) has emerged as a popular method for approximate inference for high-dimensional Bayesian models. In this paper, we propose a novel VI method that extends the naive mean field via entropic regularization, referred to as $\Xi$-variational inference ($\Xi$-VI). $\Xi$-VI has a close connection to the entropic optimal transport problem and benefits from the computationally efficient Sinkhorn algorithm. We show that $\Xi$-variational posteriors effectively recover the true posterior dependency, where the likelihood function is downweighted by a regularization parameter. We analyze the role of dimensionality of the parameter space on the accuracy of $\Xi$-variational approximation and the computational complexity of computing the approximate distribution, providing a rough characterization of the statistical-computational trade-off in $\Xi$-VI, where higher statistical accuracy requires greater computational effort. We also investigate the frequentist properties of $\Xi$-VI and establish results on consistency, asymptotic normality, high-dimensional asymptotics, and algorithmic stability. We provide sufficient criteria for our algorithm to achieve polynomial-time convergence. Finally, we show the inferential benefits of using $\Xi$-VI over mean-field VI and other competing methods, such as normalizing flow, on simulated and real datasets.

[abs][pdf][bib]

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by AI News Hub · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

modelpaper

Knowledge Map

Knowledge Map
TopicsEntitiesSource
Extending M…modelpaperJMLR

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 171 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Models