Aligning Multimodal Sequential Recommendations via Robust Direct Preference Optimization with Sparse MoE
arXiv:2603.29259v1 Announce Type: new Abstract: Preference-based alignment objectives have been widely adopted, from RLHF-style pairwise learning in large language models to emerging applications in recommender systems. Yet, existing work rarely examines how Direct Preference Optimization (DPO) behaves under implicit feedback, where unobserved items are not reliable negatives. We conduct systematic experiments on multimodal sequential recommendation to compare common negative-selection strategies and their interaction with DPO training. Our central finding is that a simple modification, replacing deterministic hard negatives with stochastic sampling from a dynamic top-K candidate pool, consistently improves ranking performance. We attribute its effectiveness to two factors: (1) reducing er
View PDF HTML (experimental)
Abstract:Preference-based alignment objectives have been widely adopted, from RLHF-style pairwise learning in large language models to emerging applications in recommender systems. Yet, existing work rarely examines how Direct Preference Optimization (DPO) behaves under implicit feedback, where unobserved items are not reliable negatives. We conduct systematic experiments on multimodal sequential recommendation to compare common negative-selection strategies and their interaction with DPO training. Our central finding is that a simple modification, replacing deterministic hard negatives with stochastic sampling from a dynamic top-K candidate pool, consistently improves ranking performance. We attribute its effectiveness to two factors: (1) reducing erroneous suppressive gradients caused by false negatives, and (2) retaining informative hard signals while smoothing optimization via controlled stochasticity. With an optional sparse Mixture-of-Experts encoder for efficient capacity scaling, RoDPO achieves up to 5.25% NDCG@5 on three Amazon benchmarks, with nearly unchanged inference cost.
Subjects:
Information Retrieval (cs.IR); Computation and Language (cs.CL)
Cite as: arXiv:2603.29259 [cs.IR]
(or arXiv:2603.29259v1 [cs.IR] for this version)
https://doi.org/10.48550/arXiv.2603.29259
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Hejin Huang [view email] [v1] Tue, 31 Mar 2026 04:49:32 UTC (656 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modellanguage modelbenchmark
Tried Designing a 8 bytes PDAP BINARY based on JSON 404 and TOON 171, T-TOON 130 and A TOKENIZED T-TOON 112 not sure if it works in Real World Applications?
TOTAL: 8 bytes, zero parsing overhead ``` **What we eliminated:** Field names (`“disk”`, `“byte”`, `“value”`) Length prefixes Token tables / dictionaries Schema metadata Repetition & redundancy **What we kept:** Fixed positional meaning (byte 0 = disk0, byte 1 = disk1, etc.) Pre-agreed protocol between sender/receiver Direct memory mapping → CPU can load in 1–2 instructions -– ## Working Code: PDAP Binary Encoder/Decoder (JavaScript) ```javascript // PDAP Binary: 8-byte ultra-compact format class PDAPBinary { // Encode: 32-bit value + 4 disk bytes → 8-byte Buffer static encode(value32, diskBytes) { if (diskBytes.length !== 4) throw new Error(‘Exactly 4 disk bytes required’); const buffer = Buffer.alloc(8); // Bytes 0-3: 32-bit value (big-endian) buffer.writeUInt32BE(value32 >>> 0, 0); // B

Smallest.ai launches Lightning V3, a new text-to-speech model that beats OpenAI, Cartesia, and ElevenLabs on key voice quality benchmarks - The Tribune
Smallest.ai launches Lightning V3, a new text-to-speech model that beats OpenAI, Cartesia, and ElevenLabs on key voice quality benchmarks The Tribune
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

Smallest.ai launches Lightning V3, a new text-to-speech model that beats OpenAI, Cartesia, and ElevenLabs on key voice quality benchmarks - The Tribune
Smallest.ai launches Lightning V3, a new text-to-speech model that beats OpenAI, Cartesia, and ElevenLabs on key voice quality benchmarks The Tribune
langchain-core==1.2.26
Changes since langchain-core==1.2.25 release(core): 1.2.26 ( #36511 ) fix(core): add init validator and serialization mappings for Bedrock models ( #34510 ) feat(core): add ChatBaseten to serializable mapping ( #36510 ) chore(core): drop gpt-3.5-turbo from docstrings ( #36497 ) fix(core): correct parameter names in filter_messages docstring example ( #36462 )


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!