MemFactory: Unified Inference & Training Framework for Agent Memory
arXiv:2603.29493v1 Announce Type: new Abstract: Memory-augmented Large Language Models (LLMs) are essential for developing capable, long-term AI agents. Recently, applying Reinforcement Learning (RL) to optimize memory operations, such as extraction, updating, and retrieval, has emerged as a highly promising research direction. However, existing implementations remain highly fragmented and task-specific, lacking a unified infrastructure to streamline the integration, training, and evaluation of these complex pipelines. To address this gap, we present MemFactory, the first unified, highly modular training and inference framework specifically designed for memory-augmented agents. Inspired by the success of unified fine-tuning frameworks like LLaMA-Factory, MemFactory abstracts the memory lif
View PDF HTML (experimental)
Abstract:Memory-augmented Large Language Models (LLMs) are essential for developing capable, long-term AI agents. Recently, applying Reinforcement Learning (RL) to optimize memory operations, such as extraction, updating, and retrieval, has emerged as a highly promising research direction. However, existing implementations remain highly fragmented and task-specific, lacking a unified infrastructure to streamline the integration, training, and evaluation of these complex pipelines. To address this gap, we present MemFactory, the first unified, highly modular training and inference framework specifically designed for memory-augmented agents. Inspired by the success of unified fine-tuning frameworks like LLaMA-Factory, MemFactory abstracts the memory lifecycle into atomic, plug-and-play components, enabling researchers to seamlessly construct custom memory agents via a "Lego-like" architecture. Furthermore, the framework natively integrates Group Relative Policy Optimization (GRPO) to fine-tune internal memory management policies driven by multi-dimensional environmental rewards. MemFactory provides out-of-the-box support for recent cutting-edge paradigms, including Memory-R1, RMM, and MemAgent. We empirically validate MemFactory on the open-source MemAgent architecture using its publicly available training and evaluation data. Across both in-domain and out-of-distribution evaluation sets, MemFactory consistently improves performance over the corresponding base models, with relative gains of up to 14.8%. By providing a standardized, extensible, and easy-to-use infrastructure, MemFactory significantly lowers the barrier to entry, paving the way for future innovations in memory-driven AI agents.
Comments: 10 pages, Code: this https URL
Subjects:
Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
Cite as: arXiv:2603.29493 [cs.CL]
(or arXiv:2603.29493v1 [cs.CL] for this version)
https://doi.org/10.48550/arXiv.2603.29493
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Zi Liang Guo [view email] [v1] Tue, 31 Mar 2026 09:38:21 UTC (105 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
llamamodellanguage modelExclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - WSJ
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxPZ2pNbEQyT1dhaDJRWllyZHVUQnBJd0d4WGZnMTg3RnRWYUpvOHJYOGNLMUc5NTdFU1J2dFJrdW5UejdtSF9zeXlVa1l3V09ValkxS1BwdlhzR2ZKLUR2QktrdDhiNlh1RVZxTjI3aVVpWVpJWWI0NjN2Q3d0ekdrS2YtVmc4MEN6ZjRQN3BWUTU0ZzJpT0Y1N01GN1UyT1ROeDJCb0gxR2xNYkNBZ0dHazdmeXlCQ2p0Tk8zR3RyM0lHVmc4QlRLVDRGeFptNXJ2WGR0bHR0QlJIb2psZjBsNzhhSnZaOFVqMnhQVUFoRzltLTFlMUdVQWl5WUJRX3NQSW1yOW1pTFpURkEzd2otMHFxRmtyNDEyZ2NTOVBkVHZCcGh1aEpURjFQQUNrNFBQX3ozUk4yV2xCejQ5RHY0elNibEtXSEhBZ1NDVWhRQzFieXNrMjRxb085RUtSY2pleHhCZ2UyWU1SdVZZcFo5U0JES01yQmtuUzFySWl3MW9iako4X3FYWXFuUGN0SUc2MXJUWUx6OE8zbW1BMm5YNXZSYTduUHNPazZ2QlgwZlNBdFNEX2RKWA?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">WSJ</font>
Introducing LIMBO: Maintaining Optimal P(DOOM) (and a call for funding)
We are excited to publicly introduce the Laboratory for Importance-sampled Measure and Bayesian Observation (LIMBO), a small research group working at the intersection of cosmological theory, probability, and existential risk. We believe that the mechanisms by which observers continue to exist in the universe are important, neglected, and tractable to study and influence. Since our founding in October 2024, we have developed a mathematical framework for doing anthropic reasoning about rare-event estimation, and we have obtained significant empirical evidence which validates this framework. This empirical evidence was not cherry-picked: at LIMBO, we believe in putting our money where our mouth is, and we have a strong track record of success in financial and prediction markets downstream of
I Turned helix-agent into helix-agents: One MCP Server for Ollama, Codex, and OpenAI-Compatible Models
<p>If you use Claude Code heavily, you eventually hit the same wall:</p> <ul> <li>some tasks are cheap enough for local models</li> <li>some tasks want a stronger coding agent</li> <li>some tasks are better sent to an API model</li> </ul> <p>But many MCP servers still force one provider and one execution style.</p> <p>So I evolved <code>helix-agent</code> into <strong>helix-agents</strong>.</p> <p>It now lets Claude Code delegate work across:</p> <ul> <li><code>ollama</code></li> <li><code>codex</code></li> <li><code>openai-compatible</code></li> </ul> <p>from one MCP server.</p> <h2> What changed </h2> <p>The original project was focused on one thing: sending routine work to local Ollama models with automatic routing.</p> <p>The new version keeps that path, but adds:</p> <ul> <li>multi-pr
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models
The Inside Story of the Greatest Deal Google Ever Made: Buying DeepMind - WSJ
<a href="https://news.google.com/rss/articles/CBMi-wJBVV95cUxNV1B1aUozNUlaZl92aVNjcE9TMFFrWWNybVgxbDlpR1NGSjNxVlBzTS1JNU9hOGh2dmlzbk53S2M4X3pDdzR1eThNUGpJUFRHdmJvSEN2TEhSczBMX3ZGVFBBelpUMFpvNm5vSVF6SU1TNE5RZDdveVVsUEFuUzZZb3Y3T2ZOd0N6TDVoQmNzZGllM2RSUWNFVTdKT0RHQ1FfSnlEaXdTUXJpSHVyQU90Y0xyLW1sR0dKRnNrdmJwcDd5YXN6eXJwYXFGLU9LYk5KdmlVV0o5MEpHRFJWWTZfaXFwTUV2Y3FyYWVuS1hSajUtSGtWMWxrNzgwTGZuMllmNHR1LUZZTU5udXNDM0NRaDluV0pJaXhoNktGZGpDMGZTT1VnQUVMWkpTaFZaN1FQQ1RXeXBadjVTc1cyVm1RUTFFUUZxQjU3bE1tU2d1WVh6QWs1eUIyT0ZRSE0wMjZFSkFYaHZTS1AtaUlNLWtQbnV3eHFhaUVZZERn?oc=5" target="_blank">The Inside Story of the Greatest Deal Google Ever Made: Buying DeepMind</a> <font color="#6f6f6f">WSJ</font>
Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT - WSJ
<a href="https://news.google.com/rss/articles/CBMiogNBVV95cUxPanNzNi1fZkhrSDRYdmY3SmRJU3MxdlhnUWZWOUhkOTB1QnpHakNrX3BMc0RjU2VoTHZEeU8zcmE2TDB2LWRLZzhiVlBtdm1obHB2dm9CVHhqNUNzNGJUZGNnUDVCdmI3OXJYanRUVDVyd29ueEF6eGhiVDZZRm5ndUdfaWxwRXBHWXRNZlZJSzJqclNYaUJMdEMzWURMVXM1R2p6SVJucHpkd3M5dG1PUTZ2QlJ2Q1o5dlFUSlFTSnlnRG5pUHFJTmFSM193Y1NlXzRlZUIzRHBXSkYzbU9LSC1CWVJqZGRwUWVDWEJEc3Y4aTRocWhlTHowMzBhY0JCenhRUHhDZ2t2UURGUVphVUdZMjdrc0hIUVYybUJTVXl5Vm5iQVJ3eEEwRHV5MmEzU1Bqb0xOTXUxR0xQZWJ3S09SVWNOUDFtbGtCYnE0aWVneUV2Y3BSbUUyckJZekdaTWFqRXUxYU1qN0hOSzJ3NjBDeDJvMjZKYlQxejdkWFE2MEgwQUlhdms5c2ZMOW9JSVhTdFRBMXpPVHBTYzhWLWFn?oc=5" target="_blank">Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT</a> <font color="#6f6f6f">WSJ</font>
Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - WSJ
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxPZ2pNbEQyT1dhaDJRWllyZHVUQnBJd0d4WGZnMTg3RnRWYUpvOHJYOGNLMUc5NTdFU1J2dFJrdW5UejdtSF9zeXlVa1l3V09ValkxS1BwdlhzR2ZKLUR2QktrdDhiNlh1RVZxTjI3aVVpWVpJWWI0NjN2Q3d0ekdrS2YtVmc4MEN6ZjRQN3BWUTU0ZzJpT0Y1N01GN1UyT1ROeDJCb0gxR2xNYkNBZ0dHazdmeXlCQ2p0Tk8zR3RyM0lHVmc4QlRLVDRGeFptNXJ2WGR0bHR0QlJIb2psZjBsNzhhSnZaOFVqMnhQVUFoRzltLTFlMUdVQWl5WUJRX3NQSW1yOW1pTFpURkEzd2otMHFxRmtyNDEyZ2NTOVBkVHZCcGh1aEpURjFQQUNrNFBQX3ozUk4yV2xCejQ5RHY0elNibEtXSEhBZ1NDVWhRQzFieXNrMjRxb085RUtSY2pleHhCZ2UyWU1SdVZZcFo5U0JES01yQmtuUzFySWl3MW9iako4X3FYWXFuUGN0SUc2MXJUWUx6OE8zbW1BMm5YNXZSYTduUHNPazZ2QlgwZlNBdFNEX2RKWA?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">WSJ</font>
Anthropic Races to Contain Leak of Code Behind Claude AI Agent - WSJ
<a href="https://news.google.com/rss/articles/CBMipgNBVV95cUxPVElWTlAxY0pFb05VZ0E1NjlfNWdZaG5SWTFJVDBkcWZvY2Y2eWFibTNWbXROSnRJX05yZHA3djFJWmxJMjVhRlRCQmtHRG5nLTYxZ2tjMmZNcjFVYlhWMXh5aXRVTWJFTFRsQTBHRnpyVnVsN2JNclNtTTNqdTcyN3EySGIySjhnQXdJWEVqQzNzcm1ybnRoYnZJeWZ4TVhkU3ZYR1Z2ZTg1UThQRjhlODEtUGNOakJKWndwZ0xtaWw4aHNiV2tRal9TVTYzeUVENzExM3g2bVkzNS1OOWhsQTBfbE4xYnk1VGNXZU5zcGJENEtFVFpEYUZlYjJlNW0ydFVLQXpRTTNXR2dZOHd6UVJBMlBZeFZfenBObXp0RWJQZ21YNEhNY3Bvc0J2MnBaZE1IS0NaRWl4dklzMzRveFo1QnJQd01ESGx1TXFOcHA4QXNWZmFMRWhhLXRGN1NtZDFrYkdoOXFPNklDejQzcWxyX1Q4c2thdmhYOE4xQUZKUjhaNlh5UUUwODlLcF9yaENsVW5Cb1FVS0NndHoza21zSVVxZw?oc=5" target="_blank">Anthropic Races to Contain Leak of Code Behind Claude AI Agent</a> <font color="#6f6f6f">WSJ</font>

Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!