UnWeaving the knots of GraphRAG -- turns out VectorRAG is almost enough
arXiv:2603.29875v1 Announce Type: new Abstract: One of the key problems in Retrieval-augmented generation (RAG) systems is that chunk-based retrieval pipelines represent the source chunks as atomic objects, mixing the information contained within such a chunk into a single vector. These vector representations are then fundamentally treated as isolated, independent and self-sufficient, with no attempt to represent possible relations between them. Such an approach has no dedicated mechanisms for handling multi-hop questions. Graph-based RAG systems aimed to ameliorate this problem by modeling information as knowledge-graphs, with entities represented by nodes being connected by robust relations, and forming hierarchical communities. This approach however suffers from its own issues with some
View PDF HTML (experimental)
Abstract:One of the key problems in Retrieval-augmented generation (RAG) systems is that chunk-based retrieval pipelines represent the source chunks as atomic objects, mixing the information contained within such a chunk into a single vector. These vector representations are then fundamentally treated as isolated, independent and self-sufficient, with no attempt to represent possible relations between them. Such an approach has no dedicated mechanisms for handling multi-hop questions. Graph-based RAG systems aimed to ameliorate this problem by modeling information as knowledge-graphs, with entities represented by nodes being connected by robust relations, and forming hierarchical communities. This approach however suffers from its own issues with some of them being: orders of magnitude increased componential complexity in order to create graph-based indices, and reliance on heuristics for performing retrieval. We propose UnWeaver, a novel RAG framework simplifying the idea of GraphRAG. UnWeaver disentangles the contents of the documents into entities which can occur across multiple chunks using an LLM. In the retrieval process entities are used as an intermediate way of recovering original text chunks hence preserving fidelity to the source material. We argue that entity-based decomposition yields a more distilled representation of original information, and additionally serves to reduce noise in the indexing, and generation process.
Subjects:
Information Retrieval (cs.IR); Artificial Intelligence (cs.AI); Computation and Language (cs.CL)
ACM classes: I.2.4; H.3.3; I.2.7
Cite as: arXiv:2603.29875 [cs.IR]
(or arXiv:2603.29875v1 [cs.IR] for this version)
https://doi.org/10.48550/arXiv.2603.29875
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Mateusz Galiński [view email] [v1] Fri, 6 Feb 2026 11:37:10 UTC (80 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelannouncecomponentLimitations of large language models in clinical problem-solving arising from inflexible reasoning - Nature
<a href="https://news.google.com/rss/articles/CBMiX0FVX3lxTE5XaHQ4a25IRlZYZVo1QjVFTFRVdFNJM1piUTZ6bWs1bTRPLVlBLVF3eGhxclI1WWNkRUxsTV9fMFVtM0dkSkhvWDdRUVN3dm4yUUtVbEMweklWUGpCZEQ4?oc=5" target="_blank">Limitations of large language models in clinical problem-solving arising from inflexible reasoning</a> <font color="#6f6f6f">Nature</font>
Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - wsj.com
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxPV0Z4UERPZU5fUFY4QXBHeXRud2w2ZWN1WTRaUVFSQ081QnM0YXdKLVFtclZTb2l5SE5QNUxzQTE4eV9HMjhqYkE0RE1HN1hITFVOMFU1c1FSZjcxR3F0Y2w3NHVrVndjUERpS1FTX3JQS0Y2MjF6U1dpY1J0elBFUVN2VEZULXVqUUxGeHVjYUFNUERVVUdlb1F4cGxQQmRKY3poVVlVTVphbDV5SlU0X0ZZVHlUTmFlUWRFNHFfbm9nODVBV3pWZjVGOUZyOFlSZlBlLVNnS2o2eXJxWEtTUEJOaUlnTUlIN29tWTJoWXcwWGxXOXlzU1BLWkNQUzBURWxqRXk5RWVOQ0JDT1FoMVMtUVJfdk9LYlpLb3RSeGpRb2poUlYyZUZRNk54cFBKYTVqcXI1SkdKQlBGRGZKLWcwNDNjTHRaSGtVV1o0dHItTnpvNzJzZ3Qteks2NVFKSndoOVZhdFFHbXdCQXlkbFlSLUJfeFphTmlKN1FOcXVUQVNYcFBaMEk2WU5CSjM3M01lbUZROFlWbHZvaFpkR1I5ZGVlazdxSVJuNmVoN3hQai1GTDBkZw?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">wsj.com</font>
Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - WSJ
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxNNWh0OTV4cnNDLVdHdHVUdE02cWRiaE03VENfdWJFbFlyaXZmbWtJNm9OdU05TXVsRjd4dVFwUzl0WkRfLVJoVzlaNkhKRVl4S0Y0Um5jN2QzZzhsb0twMElFOEpSZjdjX1pZZzNacXIxU2U4Ulloam5nR1hQeXg3TWhoMEE1ZzFzQmdiSjktRG1rUEs2YVVhZ0VMMk0wS3J6SWNJdTdJZlAtTEE0SUdaaFl5QWFUWS05NGFDN1FudnNRN2ZpcnFmM0N1bGVpSjNYZmZ6MUJKSkpMWk5tRWFSN2s4V0tEdi1EVVBuUTdnZm92Sjk0MEVYZWRieTkxNWMwRzRiQmxWVHpvaGEwUnpEZGJ1UVFhQmoydGxSTW93XzFVR1ZHeG5mMTZOLWthOVVKZTZMeGdsS0dDaUROelpWc1l4QmJLNWkzRkhGUGdua3hnOHFWYUpXQWp3RktyemZiN0VBTFhfNGFZUHpNaV9jX2U0Sk9Fb2k1dXhOZHdENWpPc2dRU2ZQeHZoMnBZNEN6RHJnNU1YYk9SSzRYNzZrbXRnQ3VOdE0ydGFCM3ZBVTdHeFJLY29Feg?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">WSJ</font>
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models
Limitations of large language models in clinical problem-solving arising from inflexible reasoning - Nature
<a href="https://news.google.com/rss/articles/CBMiX0FVX3lxTE5XaHQ4a25IRlZYZVo1QjVFTFRVdFNJM1piUTZ6bWs1bTRPLVlBLVF3eGhxclI1WWNkRUxsTV9fMFVtM0dkSkhvWDdRUVN3dm4yUUtVbEMweklWUGpCZEQ4?oc=5" target="_blank">Limitations of large language models in clinical problem-solving arising from inflexible reasoning</a> <font color="#6f6f6f">Nature</font>
Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - wsj.com
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxPV0Z4UERPZU5fUFY4QXBHeXRud2w2ZWN1WTRaUVFSQ081QnM0YXdKLVFtclZTb2l5SE5QNUxzQTE4eV9HMjhqYkE0RE1HN1hITFVOMFU1c1FSZjcxR3F0Y2w3NHVrVndjUERpS1FTX3JQS0Y2MjF6U1dpY1J0elBFUVN2VEZULXVqUUxGeHVjYUFNUERVVUdlb1F4cGxQQmRKY3poVVlVTVphbDV5SlU0X0ZZVHlUTmFlUWRFNHFfbm9nODVBV3pWZjVGOUZyOFlSZlBlLVNnS2o2eXJxWEtTUEJOaUlnTUlIN29tWTJoWXcwWGxXOXlzU1BLWkNQUzBURWxqRXk5RWVOQ0JDT1FoMVMtUVJfdk9LYlpLb3RSeGpRb2poUlYyZUZRNk54cFBKYTVqcXI1SkdKQlBGRGZKLWcwNDNjTHRaSGtVV1o0dHItTnpvNzJzZ3Qteks2NVFKSndoOVZhdFFHbXdCQXlkbFlSLUJfeFphTmlKN1FOcXVUQVNYcFBaMEk2WU5CSjM3M01lbUZROFlWbHZvaFpkR1I5ZGVlazdxSVJuNmVoN3hQai1GTDBkZw?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">wsj.com</font>
Claude Code Leak Reveals Anthropic’s Internal Playbook - Startup Fortune
<a href="https://news.google.com/rss/articles/CBMiiAFBVV95cUxQT29vMFNyN0gzZVA0VXMwdDJEcVVTRXJIQlNRU1NzMHhpM1dGaDNNQTdIR3VBdjRuTFkxQTZVZXNKUUZkR0F3eWRJWENLa1BVU0tlbnJ0djdTekN0WGduc0FpWFRfcGN5am5BMWlUblZGX1V3QmtMWF9pZzYxQ0l2X1M0VGtNU0do?oc=5" target="_blank">Claude Code Leak Reveals Anthropic’s Internal Playbook</a> <font color="#6f6f6f">Startup Fortune</font>
I use these 5 simple ‘ChatGPT codes’ every day — and they instantly improve my results - TechRadar
<a href="https://news.google.com/rss/articles/CBMi1wFBVV95cUxNbzhUTm9iTHh5bGJJMVRZcDRBeDdZUXZZWFhvdFF3cHhocGhSV0JXLTdBWDNab21mdFFDdzBWVGpyY1VUZmN1aFI3YWtnQW9mQ2dRUW1ndG9iZjBsZ3lqdE1ZQTZDaUVYZV9rLWNmamNCQTJEUXFkOFNMTTA2WUY0bmpWalBjY1hBRms3M2ZjOEIwRHFWM0F4YlJCMGVHUms3REJ6ZlpkWjU1OUU3VWNWU09JTnBIYVpYNzhaclRoZW5aMUp1djAxZ2hKTmJZQ0tuMUFXNXpkNA?oc=5" target="_blank">I use these 5 simple ‘ChatGPT codes’ every day — and they instantly improve my results</a> <font color="#6f6f6f">TechRadar</font>
Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!