I Turned helix-agent into helix-agents: One MCP Server for Ollama, Codex, and OpenAI-Compatible Models
<p>If you use Claude Code heavily, you eventually hit the same wall:</p> <ul> <li>some tasks are cheap enough for local models</li> <li>some tasks want a stronger coding agent</li> <li>some tasks are better sent to an API model</li> </ul> <p>But many MCP servers still force one provider and one execution style.</p> <p>So I evolved <code>helix-agent</code> into <strong>helix-agents</strong>.</p> <p>It now lets Claude Code delegate work across:</p> <ul> <li><code>ollama</code></li> <li><code>codex</code></li> <li><code>openai-compatible</code></li> </ul> <p>from one MCP server.</p> <h2> What changed </h2> <p>The original project was focused on one thing: sending routine work to local Ollama models with automatic routing.</p> <p>The new version keeps that path, but adds:</p> <ul> <li>multi-pr
If you use Claude Code heavily, you eventually hit the same wall:
-
some tasks are cheap enough for local models
-
some tasks want a stronger coding agent
-
some tasks are better sent to an API model
But many MCP servers still force one provider and one execution style.
So I evolved helix-agent into helix-agents.
It now lets Claude Code delegate work across:
-
ollama
-
codex
-
openai-compatible
from one MCP server.
What changed
The original project was focused on one thing: sending routine work to local Ollama models with automatic routing.
The new version keeps that path, but adds:
-
multi-provider switching
-
Codex-backed code delegation
-
OpenAI-compatible chat API support
-
Claude Code-style background agents
Under the hood, the runtime now supports two different delegation styles:
-
a built-in ReAct loop for ollama and openai-compatible
-
an autonomous Codex-backed path for repo-heavy work
That means the workflow is no longer:
Claude Code -> one tool call -> one reply
Enter fullscreen mode
Exit fullscreen mode
It can now be:
Claude Code -> spawn a worker -> send follow-up instructions -> wait for completion -> inspect and closeClaude Code -> spawn a worker -> send follow-up instructions -> wait for completion -> inspect and closeEnter fullscreen mode
Exit fullscreen mode
Why this matters
Different providers are good at different things.
-
ollama: local reasoning, low-cost drafts, vision
-
codex: code-heavy implementation and repo work
-
openai-compatible: hosted chat models behind standard APIs
Instead of wiring three separate MCP servers with different interaction models, I wanted one consistent runtime.
New tools
Core tools:
-
think
-
agent_task
-
see
-
providers
-
models
-
config
Background agent tools:
-
spawn_agent
-
send_agent_input
-
wait_agent
-
list_agents
-
close_agent
Example flows
1. Code review via Codex
think( task="Review this diff for regressions", provider="codex", cwd="/repo" )think( task="Review this diff for regressions", provider="codex", cwd="/repo" )Enter fullscreen mode
Exit fullscreen mode
2. Local summarization via Ollama
think( task="Summarize this build log", provider="ollama" )think( task="Summarize this build log", provider="ollama" )Enter fullscreen mode
Exit fullscreen mode
3. Persistent investigation worker
spawn_agent( description="Investigate flaky tests", provider="codex", agent_type="explorer" )spawn_agent( description="Investigate flaky tests", provider="codex", agent_type="explorer" )Enter fullscreen mode
Exit fullscreen mode
Then:
send_agent_input(...) wait_agent(...) close_agent(...)send_agent_input(...) wait_agent(...) close_agent(...)Enter fullscreen mode
Exit fullscreen mode
Setup
git clone https://github.com/tsunamayo7/helix-agent.git cd helix-agent uv sync uv run python server.pygit clone https://github.com/tsunamayo7/helix-agent.git cd helix-agent uv sync uv run python server.pyEnter fullscreen mode
Exit fullscreen mode
Add to Claude Code:
{ "mcpServers": { "helix-agents": { "command": "uv", "args": ["run", "--directory", "/path/to/helix-agent", "python", "server.py"] } } }{ "mcpServers": { "helix-agents": { "command": "uv", "args": ["run", "--directory", "/path/to/helix-agent", "python", "server.py"] } } }Enter fullscreen mode
Exit fullscreen mode
Notes
-
Codex requires codex on PATH
-
OpenAI-compatible mode requires an API key
-
The generic OpenAI-compatible path is currently text-first
-
Vision is currently centered on the Ollama path
GitHub: helix-agent
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
claudellamamodelAnthropic's AI Agent Claude leak that it termed 'human error' has exposed commercially … - The Times of India
<a href="https://news.google.com/rss/articles/CBMi-wFBVV95cUxOOFUwN202VWFBejRMM1VaQjQzTGl5cGJkSDF5WmpxeUFHTzFXWVB4TW5PS25hRTlNMFZfMGZoR0hHSUYwNEpEbUdaX1lqdE9aRGpMWnk4ZDc0TDJjeDluQ045TFlycU5zcVZSV1gxUWt0QlpKc3lseEV1ZnlpMHJNcF9DdFN5V3hQZENxdWxHVDkyYXZzOGFPWDhrV2Rhb25lRXN4LUxGSkx4dkJhcDNpdG9IZ2ZTbk1JMHVFWXJmTUtiVDNEaC1zVTRXY1RON3QzNXVuNlRaVFpyUm5Yd3NGdGZrbzgyMWdhTTB2YlhiY0ZEUTRfYndBZjNIZ9IBgAJBVV95cUxPUW9tczBFazI4ZTJYMXlOU1lWQWIwUkJmMFBtZ0VRYk5waFJWU0IxWXNWMzJzQWQzYlVDUVVZbkdOX245eDFZRXNzMzRiWWNOMzZzMk1ELS1MR3hMYnd5YjZ0VnVVb1ZXd29wX2JPUjJSOHJIWGJuQ0xnYUMtdk5CUnNSVVZUSE5GUUVKS2NqTVRsbF9aWUFDZWttSDF4cjY5Nk1nMlpwS3lKZWNtT2M5eDRxTHNDX1J0TkxXd0tsMDNjQlF2NlpzdHBMRk9jMV9zTVdmZnlRSWxCMEppQmJ4UjcxTGJNTGF1UElzaDNYR0VUMkJQdk9jSjFsOFcwMVZI?oc=5" target="_blank">Anthropic's AI Agent Claude leak that it te
Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - WSJ
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxPNFo5MVV0R29iUUFrY2xQZzhoNWJqTmlucXBEd1E4dXQwbmp1ZnA3cklGUGQtUkxwQTBpMXNKMUpkNEFiVmZMb0ZzU01LVmRNYUxUTXlVZ19oVmVzb3dOWFJrM2NlVlU2X0duTlkxT2lNaFdDTzg1Um1WQUxicmdPTDdkVkoyWkZfQ3NjRmFSZ3VvYlRjWF9IV3F2d0hpMmVYZmZUcVRoVTNtS3Z2VFlvSGhPLTgteHZyZ3pXRnZNbjI5UHlhWjJtT3NJS1BnVlRib3BVV3dJY04zYXR3TW1vem5mOUJuZXpDTHJMWWs5WERMdXozVWFVQUVCU2tPVGQ5M0tsSmlmblA0RG91c2RsUW1IUVJ1SEh6emtrb0ItZkNUWmJQLXp4cmNGdjdiOTVmNGdEVEZ2bUk2bXp1RFNSeFNieFc2MGl6aEJVblNyMXQ3UDlmSUdrVm1TZFNXZjJJZGozZHhvdk9mcTN5LTY3dEctaVNvaGFCX0lJbklxamRjX1VMVkpWbmVzVWVnd1NqUnVwZzhCQVJfOXhRNVA2RGtGNm8ycXp6OXRqUzNzMDNPX25zZUlmRg?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">WSJ</font>

Unlocking the Future: Sourcing Essential Components like the LM317 & ATtiny85 Online for Your Projects
<h1> Unlocking the Future: Sourcing Essential Components like the LM317 & ATtiny85 Online for Your Projects </h1> <p><em>Supply chain strategy from electronics production engineering, 500–50k units/year</em></p> <h2> Introduction </h2> <p>"Order from Digi-Key" is a prototyping strategy, not a production strategy. The 2020–2023 IC shortage demonstrated that supply chain resilience must be designed in — not improvised when lead times hit 52 weeks.</p> <h2> The Sourcing Tier Structure </h2> <div class="table-wrapper-paragraph"><table> <thead> <tr> <th>Tier</th> <th>Examples</th> <th>MOQ</th> <th>Price Premium</th> <th>Lead Time</th> <th>Risk</th> </tr> </thead> <tbody> <tr> <td>Authorized dist.</td> <td>Digi-Key, Mouser, Newark</td> <td>1 pc</td> <td>+25–40%</td> <td>1–3 days (stock)</td>
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models
ChatGPT Maker OpenAI Valued at $852B After Record $122B Funding Round - Bitcoin.com News
<a href="https://news.google.com/rss/articles/CBMimAFBVV95cUxNYl9RSVpUWDFpREp2N2JJbHVvWGVhaFRlRzBOcHl1RGxoYlpWVnZWSWlUWUo1NUNNUDZEbGR1RGl6VGZQa0hWdGlVbTlYYm9UM0U3ajc1UHREcmR0WjJIbXRBdHZjblVjREdTMXJZZ1ZVeGFVNHJ6T3A3b2JSN2pLbGlNaENEeXVkNXhjRmNPSTFQeWxKaG1rNA?oc=5" target="_blank">ChatGPT Maker OpenAI Valued at $852B After Record $122B Funding Round</a> <font color="#6f6f6f">Bitcoin.com News</font>
Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT - WSJ
<a href="https://news.google.com/rss/articles/CBMiogNBVV95cUxONW1odmhMalVDSUtBdXVwcjIxczdYWXdmamk2MDZlTVBDV28wcVVVUUtpNTFoZkt0MXFPdzZOZHBLUmNWSmRwclhaTTlvbXNvcFEwS2U0NmpyTmd1c0RYUjdSWDFXUmNKUlowQkdoV0l2c2tHQkl4cHJvZENOa2hBRVlRT0hWOVdZblVIWXRJTWNibjVsSlJZbjV2aUh6bHR1ZXRkUTIwNGtYVXBsWDQ0U3BfRXVFZkhTSWM0T1g3blRLaTl1eEZUR29XWXgwQVBrVXNDSTB1OVJ4aXMzbUJ0MXJWNDBLZW5OdzNoRG5sSEtjT0hRMlBoa3Q2TktzQjZWX1FSbWhhc25XSktZYUlJNGxqYjUxbXZPVWlOR2x1QTZzMjNMMVdVZUR5UjNwcTBCcFhnbDNyeWd0S3U4V2xWMzlMN3p2elMyenl6a0gzZ19GTXdVUDZTaWhCc1hjZ3pnRFowVFdYYWMxOWhkRFprLXkyY1hHWVNxSUtOenpUQTlFX2I1bnM0a09JNXNCTWphcXJmWW9FdXV6elU5bkxaNUF3?oc=5" target="_blank">Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT</a> <font color="#6f6f6f">WSJ</font>
Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT - WSJ
<a href="https://news.google.com/rss/articles/CBMiogNBVV95cUxPOVdTTGIzVFBDS0wzUDJTVDlfWkNGZVlLd3NBMjFZdjBQOEpQenhXYXQydUpidUdTeGVzeUljcmxaT2FXa21jYVNDTDJFMnc5OFhfRkpOUzFiY21WVVUxLU5MWkRqVHpUSFVxaEVKREZ4UkNUNkNKRGRlSHJwNEJaM3hvYlRjZlZ4QnVpbW1TY1Jua3Q3a09RQ2hIQlRIWk9nQmR1YmJCSUtuTGhIV2YzSk5mbmdRdVd1d2FKamU1M3VHdEduNWlzbU1VdlBMcXFhcjR6S243MGJFTDFFNVJYNEtGR2MzWnRPdFVVUG9vWnRLdXBhMngwME5JQk5HQkxpLWZVTl9LMThLMENYOU45bnNkTHBkZ0JZblMtdDE5ckxtZE95anV4Y25uRUxlZTdHY0JIUWw0dmg4MTh0cjBPLWZvcUtGeGtMVGJlWFh4ZkJ4SGpiY2FBU3lFUU9WNGFvbUNzWFl1QkhfWjJLN3JveTRxVU5GYm1pc00tVHQ5dW9fTEd1Qm9mTWhscEdJLURvSENrd2pUVDg2cmJXS0dKWWt3?oc=5" target="_blank">Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT</a> <font color="#6f6f6f">WSJ</font>
Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - WSJ
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxPNFo5MVV0R29iUUFrY2xQZzhoNWJqTmlucXBEd1E4dXQwbmp1ZnA3cklGUGQtUkxwQTBpMXNKMUpkNEFiVmZMb0ZzU01LVmRNYUxUTXlVZ19oVmVzb3dOWFJrM2NlVlU2X0duTlkxT2lNaFdDTzg1Um1WQUxicmdPTDdkVkoyWkZfQ3NjRmFSZ3VvYlRjWF9IV3F2d0hpMmVYZmZUcVRoVTNtS3Z2VFlvSGhPLTgteHZyZ3pXRnZNbjI5UHlhWjJtT3NJS1BnVlRib3BVV3dJY04zYXR3TW1vem5mOUJuZXpDTHJMWWs5WERMdXozVWFVQUVCU2tPVGQ5M0tsSmlmblA0RG91c2RsUW1IUVJ1SEh6emtrb0ItZkNUWmJQLXp4cmNGdjdiOTVmNGdEVEZ2bUk2bXp1RFNSeFNieFc2MGl6aEJVblNyMXQ3UDlmSUdrVm1TZFNXZjJJZGozZHhvdk9mcTN5LTY3dEctaVNvaGFCX0lJbklxamRjX1VMVkpWbmVzVWVnd1NqUnVwZzhCQVJfOXhRNVA2RGtGNm8ycXp6OXRqUzNzMDNPX25zZUlmRg?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">WSJ</font>
Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!