Error Analysis for Deep ReLU Feedforward Density-Ratio Estimation with Bregman Divergence
We consider the problem of density-ratio estimation using Bregman Divergence with Deep ReLU feedforward neural networks (BDD). We establish non-asymptotic error bounds for BDD density-ratio estimators, which are minimax optimal up to a logarithmic factor when the data distribution has finite support. As an application of our theoretical findings, we propose an estimator for the KL-divergence that is asymptotically normal, leveraging our convergence results for the deep density-ratio estimator and a data-splitting method. We also extend our results to cases with unbounded support and unbounded density ratios. Furthermore, we show that the BDD density-ratio estimator can mitigate the curse of dimensionality when data distributions are supported on an approximately low-dimensional manifold. O
Siming Zheng, Guohao Shen, Yuanyuan Lin, Jian Huang; 27(15):1−60, 2026.
Abstract
We consider the problem of density-ratio estimation using Bregman Divergence with Deep ReLU feedforward neural networks (BDD). We establish non-asymptotic error bounds for BDD density-ratio estimators, which are minimax optimal up to a logarithmic factor when the data distribution has finite support. As an application of our theoretical findings, we propose an estimator for the KL-divergence that is asymptotically normal, leveraging our convergence results for the deep density-ratio estimator and a data-splitting method. We also extend our results to cases with unbounded support and unbounded density ratios. Furthermore, we show that the BDD density-ratio estimator can mitigate the curse of dimensionality when data distributions are supported on an approximately low-dimensional manifold. Our results are applied to investigate the convergence properties of the telescoping density-ratio estimator proposed by Rhodes (2020). We provide sufficient conditions under which it achieves a lower error bound than a single-ratio estimator. Moreover, we conduct simulation studies to validate our main theoretical results and assess the performance of the BDD density-ratio estimator.
[abs][pdf][bib]
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Products

🚀 The Developer Who Survives 2026 Is NOT the One You Think
⚠️ The Hard Truth In 2026, being a “good developer” is no longer enough. You can: Write clean code ✅ Know Docker, Kubernetes ✅ Grind LeetCode daily ✅ …and still get replaced. Not by another developer. But by someone who knows how to use AI better than you. 🤖 The New Battlefield: AI-Augmented Developers Let’s be clear: AI is NOT replacing developers. But developers using AI are replacing those who don’t. The game has changed from: “How well can you code?” to: “How well can you THINK, DESIGN, and ORCHESTRATE?” 🧠 The 3 Skills That Actually Matter Now 1. 🧩 AI Orchestration (The Hidden Superpower) Most devs use one tool. Top devs use systems of tools : GPT → for architecture Claude → for reasoning large codebases Copilot/Cursor → for execution Local LLM → for privacy 👉 The magic is not in t

Mura: The Source of Uneven Flow
In part 1, we explored the eight wastes ( Muda ) as the visible symptoms of inefficiency in software delivery. We saw how waste shows up in unfinished work, handoffs, long waits, rework, and lost talent. Those are the effects we can observe and feel. Those wastes are almost always the result of Mura (斑), a Japanese term from the Toyota Production System meaning "unevenness" or "inconsistency" in how work flows. It is the "hurry up and wait" cycle: periods of low activity followed by periods of frantic catch-up, that make delivery unpredictable and unsustainable. This post examines in depth how to identify uneven flow, and how modern software delivery practices work together to reduce inconsistency and create predictability. The Detection Kit The principles of Lean have been empirically val
b8668
server : fix logging of build + system info ( #21460 ) This PR changes the logging that occurs at startup of llama-server. Currently, it is redundant (including CPU information twice) and it is missing the build + commit info. macOS/iOS: macOS Apple Silicon (arm64) macOS Intel (x64) iOS XCFramework Linux: Ubuntu x64 (CPU) Ubuntu arm64 (CPU) Ubuntu s390x (CPU) Ubuntu x64 (Vulkan) Ubuntu arm64 (Vulkan) Ubuntu x64 (ROCm 7.2) Ubuntu x64 (OpenVINO) Windows: Windows x64 (CPU) Windows arm64 (CPU) Windows x64 (CUDA 12) - CUDA 12.4 DLLs Windows x64 (CUDA 13) - CUDA 13.1 DLLs Windows x64 (Vulkan) Windows x64 (SYCL) Windows x64 (HIP) openEuler: openEuler x86 (310p) openEuler x86 (910b, ACL Graph) openEuler aarch64 (310p) openEuler aarch64 (910b, ACL Graph)



![Program analysis using random interpretation (2005) [pdf]](https://d2xsxph8kpxj0f.cloudfront.net/310419663032563854/konzwo8nGf8Z4uZsMefwMr/default-img-ai-chip-closeup-KMZ5N5zRxP2NRiYJ8TB9TM.webp)


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!