DeepEye: A Steerable Self-driving Data Agent System
arXiv:2603.28889v1 Announce Type: new Abstract: Large Language Models (LLMs) have revolutionized natural language interaction with data. The "holy grail" of data analytics is to build autonomous Data Agents that can self-drive complex data analysis workflows. However, current implementations are still limited to linear "ChatBI" systems. These systems struggle with joint analysis across heterogeneous data sources (e.g., databases, documents, and data files) and often encounter "context explosion" in complex and iterative data analysis workflows. To address these challenges, we present DeepEye, a production-ready data agent system that adopts a workflow-centric architecture to ensure scalability and trustworthiness. DeepEye introduces a Unified Multimodal Orchestration protocol, enabling sea
View PDF HTML (experimental)
Abstract:Large Language Models (LLMs) have revolutionized natural language interaction with data. The "holy grail" of data analytics is to build autonomous Data Agents that can self-drive complex data analysis workflows. However, current implementations are still limited to linear "ChatBI" systems. These systems struggle with joint analysis across heterogeneous data sources (e.g., databases, documents, and data files) and often encounter "context explosion" in complex and iterative data analysis workflows. To address these challenges, we present DeepEye, a production-ready data agent system that adopts a workflow-centric architecture to ensure scalability and trustworthiness. DeepEye introduces a Unified Multimodal Orchestration protocol, enabling seamless integration of structured and unstructured data sources. To mitigate hallucinations, it employs Hierarchical Reasoning with context isolation, decomposing complex intents into autonomous AgentNodes and deterministic ToolNodes. Furthermore, DeepEye incorporates a database-inspired Workflow Engine (comprising a Compiler, Validator, Optimizer, and Executor) that guarantees structural correctness and accelerates execution via runtime topological optimization. In this demonstration, we showcase DeepEye's ability to orchestrate complex workflows to generate diverse multimodal outputs -- including Data Videos, Dashboards, and Analytical Reports -- highlighting its advantages in transparent execution, automated optimization, and human-in-the-loop reliability.
Comments: SIGMOD Demo (2026)
Subjects:
Databases (cs.DB)
Cite as: arXiv:2603.28889 [cs.DB]
(or arXiv:2603.28889v1 [cs.DB] for this version)
https://doi.org/10.48550/arXiv.2603.28889
arXiv-issued DOI via DataCite
Related DOI:
https://doi.org/10.1145/3788853.3801612
DOI(s) linking to related resources
Submission history
From: Boyan Li [view email] [v1] Mon, 30 Mar 2026 18:14:28 UTC (4,751 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modellanguage modelannounceHate Speech Detection Still Cooks (Even in 2026)
The failure case you didn’t see coming In late 2025, a major social platform quietly rolled back parts of its LLM-based moderation pipeline after internal audits revealed a systematic pattern: posts in African American Vernacular English (AAVE) were flagged at nearly three times the rate of semantically equivalent Standard American English content. The LLM reasoner, a fine-tuned GPT-4-class model had learned to treat certain phonetic spellings and grammatical constructions as proxies for “informal aggression.” A linguist reviewing the flagged corpus found no aggression whatsoever. The failure wasn’t adversarial. It was architectural: the model had no representation of dialect as a legitimate register. Simultaneously, coordinated hate communities on adjacent platforms were having a producti

What Is New In Helm 4 And How It Improves Over Helm 3
<p>The release of <strong>Helm 4</strong> marks a massive milestone in the <strong>Kubernetes</strong> ecosystem. For years developers and system administrators have relied on this robust package manager to template deploy and manage complex cloud native applications. When the maintainers transitioned from the second version to <strong>Helm 3</strong> the community rejoiced because it completely removed <strong>Tiller</strong>. That removal drastically simplified cluster security models and streamlined deployment pipelines. Now the highly anticipated <strong>Helm 4</strong> is stepping into the spotlight to address the modern challenges of <strong>DevOps</strong> workflows. This comprehensive blog post will explore exactly what is new in <strong>Helm 4</strong> and how it provides a vastly

Developers Are Designing for AI Before Users Now
<p>A quiet shift is happening in modern web development.</p> <p>For years, developers designed applications with one priority: users.</p> <p>UI came first.<br><br> User flows came first.<br><br> User experience came first. </p> <p>Backend, APIs, and integrations were built around that experience.</p> <p>But today, something has changed.</p> <p>Developers are increasingly designing systems with <strong>AI in mind before users</strong>, and this is reshaping how frontend, UX, and fullstack engineering work.</p> <h2> The Old Way of Building Applications </h2> <p>Traditional product development looked like this:</p> <p>Design UI → Build frontend → Connect backend → Launch</p> <p>The focus was simple:</p> <ul> <li>What does the user need?</li> <li>How will they interact?</li> <li>What is the ea
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models
Hate Speech Detection Still Cooks (Even in 2026)
The failure case you didn’t see coming In late 2025, a major social platform quietly rolled back parts of its LLM-based moderation pipeline after internal audits revealed a systematic pattern: posts in African American Vernacular English (AAVE) were flagged at nearly three times the rate of semantically equivalent Standard American English content. The LLM reasoner, a fine-tuned GPT-4-class model had learned to treat certain phonetic spellings and grammatical constructions as proxies for “informal aggression.” A linguist reviewing the flagged corpus found no aggression whatsoever. The failure wasn’t adversarial. It was architectural: the model had no representation of dialect as a legitimate register. Simultaneously, coordinated hate communities on adjacent platforms were having a producti
The 5th Agent Orchestration Pattern: Market-Based Task Allocation
<p>Most conversations about agent orchestration patterns settle on the same four: pipeline, supervisor, router, blackboard. Each solves coordination differently. Pipelines chain steps linearly. Supervisors centralize control. Routers classify and dispatch. Blackboards let agents coordinate through shared state without direct communication.</p> <p>These four cover a lot of ground. But there is a fifth pattern that comes from an older field, and it solves a problem the other four handle poorly: dynamic allocation across heterogeneous agents when cost matters.</p> <h2> The Pattern: Auction-Based Task Allocation </h2> <p>Instead of a supervisor deciding which agent handles a task, you let agents bid on it.</p> <p>The mechanism works like this. A task enters the system. It gets broadcast to all
The Stages of AI Grief
<blockquote> <p><strong>Assumed audience:</strong> People who work with AI daily — or are starting to — and have complicated feelings about it.</p> </blockquote> <p>I don't think I've ever had so much fun in my programming career as I do now. Which is strange, because a few weeks ago I was in a very different place. I was watching - in horror - as the machine on my desk was taking over my craft. Like most people I guess, I derive quite a lot of my identity from that craft; hence the horror. (Let's ignore for now whether that's a good thing or not.)</p> <p>I just watched it melt away. Like a block of ice in the sun; inexorable. In that moment it felt like I was witnessing an emerging god: an uncontrollable force in the sky asserts its influence over all it touches, and every day, it touches
Unlock the Power of Private AI: Build a Local RAG Pipeline with LangGraph, Ollama & Vector Databases
<blockquote> <p>I created a new website: <a href="https://programmingcentral.vercel.app/books/typescript/" rel="noopener noreferrer">Free Access to the 8 Volumes on Typescript & AI Masterclass</a>, no registration required. Choose Volume and chapter on the menu on the left. 160 Chapters and hundreds of quizzes at the end of chapters.</p> </blockquote> <p>Retrieval-Augmented Generation (RAG) is revolutionizing how we interact with AI, allowing models to provide more informed and contextually relevant answers. But what if you need to keep your data private and secure? This guide dives into building a <strong>Private RAG pipeline</strong> – a self-contained AI system that operates entirely on your machine, leveraging local embeddings, vector stores, and Large Language Models (LLMs). We'll
Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!