MemRerank: Preference Memory for Personalized Product Reranking
arXiv:2603.29247v1 Announce Type: new Abstract: LLM-based shopping agents increasingly rely on long purchase histories and multi-turn interactions for personalization, yet naively appending raw history to prompts is often ineffective due to noise, length, and relevance mismatch. We propose MemRerank, a preference memory framework that distills user purchase history into concise, query-independent signals for personalized product reranking. To study this problem, we build an end-to-end benchmark and evaluation framework centered on an LLM-based \textbf{1-in-5} selection task, which measures both memory quality and downstream reranking utility. We further train the memory extractor with reinforcement learning (RL), using downstream reranking performance as supervision. Experiments with two L
View PDF HTML (experimental)
Abstract:LLM-based shopping agents increasingly rely on long purchase histories and multi-turn interactions for personalization, yet naively appending raw history to prompts is often ineffective due to noise, length, and relevance mismatch. We propose MemRerank, a preference memory framework that distills user purchase history into concise, query-independent signals for personalized product reranking. To study this problem, we build an end-to-end benchmark and evaluation framework centered on an LLM-based \textbf{1-in-5} selection task, which measures both memory quality and downstream reranking utility. We further train the memory extractor with reinforcement learning (RL), using downstream reranking performance as supervision. Experiments with two LLM-based rerankers show that MemRerank consistently outperforms no-memory, raw-history, and off-the-shelf memory baselines, yielding up to \textbf{+10.61} absolute points in 1-in-5 accuracy. These results suggest that explicit preference memory is a practical and effective building block for personalization in agentic e-commerce systems.
Comments: correct author name in metadata
Subjects:
Computation and Language (cs.CL); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)
Cite as: arXiv:2603.29247 [cs.CL]
(or arXiv:2603.29247v2 [cs.CL] for this version)
https://doi.org/10.48550/arXiv.2603.29247
arXiv-issued DOI via DataCite
Submission history
From: Zhiyuan Peng [view email] [v1] Tue, 31 Mar 2026 04:16:18 UTC (1,548 KB) [v2] Thu, 2 Apr 2026 07:36:51 UTC (1,548 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
benchmarkannounceproduct
I open-sourced a tool that compiles raw documents into an AI-navigable wiki with persistent memory; runs 100% locally
After seeing Karpathy's tweet about using LLMs to build personal wikis from research documents, I realized I'd already been using something similar like this internally for our R D. So I cleaned it up and open-sourced it. What it does: You drop a folder of raw documents (PDFs, papers, notes, code, 60+ formats) and the LLM compiles them into a structured markdown wiki with backlinked articles, concept pages, and a master index. It then compresses everything into a .aura archive optimized for RAG retrieval (~97% smaller than raw source data). How it works: pip install aura-research research init my-project # copy docs into raw/ research ingest raw/ research compile research query "your question" Key design decisions: No embeddings, no vector databases. Uses SimHash + Bloom Filters instead. Z

Gemma4:26b's reasoning capabilities are crazy.
Been experimenting with it, first on my buddy's compute he let me borrow, and then with the Gemini SDK so that I don't need to keep stealing his macbook from 600 miles away. Originally my home agent was run through Gemini-3-Flash because no other model I've tried has been able to match it's reasoning ability. The script(s) I have it running through are a re-implementation of a multi-speaker smart home speaker setup, with several rasperry pi zeroes functioning as speaker satellites for a central LLM hub, right now a raspberry pi 5, soon to be an M4 mac mini prepped for full local operation. It also has a dedicated discord bot I use to interact with it from my phone and PC for more complicated tasks, and those requiring information from an image, like connector pinouts I want help with. I've
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

I open-sourced a tool that compiles raw documents into an AI-navigable wiki with persistent memory; runs 100% locally
After seeing Karpathy's tweet about using LLMs to build personal wikis from research documents, I realized I'd already been using something similar like this internally for our R D. So I cleaned it up and open-sourced it. What it does: You drop a folder of raw documents (PDFs, papers, notes, code, 60+ formats) and the LLM compiles them into a structured markdown wiki with backlinked articles, concept pages, and a master index. It then compresses everything into a .aura archive optimized for RAG retrieval (~97% smaller than raw source data). How it works: pip install aura-research research init my-project # copy docs into raw/ research ingest raw/ research compile research query "your question" Key design decisions: No embeddings, no vector databases. Uses SimHash + Bloom Filters instead. Z

Gemma4:26b's reasoning capabilities are crazy.
Been experimenting with it, first on my buddy's compute he let me borrow, and then with the Gemini SDK so that I don't need to keep stealing his macbook from 600 miles away. Originally my home agent was run through Gemini-3-Flash because no other model I've tried has been able to match it's reasoning ability. The script(s) I have it running through are a re-implementation of a multi-speaker smart home speaker setup, with several rasperry pi zeroes functioning as speaker satellites for a central LLM hub, right now a raspberry pi 5, soon to be an M4 mac mini prepped for full local operation. It also has a dedicated discord bot I use to interact with it from my phone and PC for more complicated tasks, and those requiring information from an image, like connector pinouts I want help with. I've

‘We got massively outgunned overnight’: Indian startup founder shares how Google, ChatGPT crushed his AI startup - The Financial Express
‘We got massively outgunned overnight’: Indian startup founder shares how Google, ChatGPT crushed his AI startup The Financial Express




Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!