Live
Black Hat USADark ReadingBlack Hat AsiaAI BusinessHow We Built an EdTech Platform That Scaled to 250K Daily UsersDEV CommunityRoguelike Devlog: Redesigning a Game UI With an AI 2D Game MakerDEV CommunityI spent days debugging a cron job that was "working fine"DEV CommunityLLM Agents Need a Nervous System, Not Just a BrainDEV CommunityThe 22,000 Token Tax: Why I Killed My MCP ServerDEV CommunityOpenSpec (Spec-Driven Development) Failed My Experiment — Instructions.md Was Simpler and FasterDEV CommunityI Asked AI to Do Agile Sprint Planning (GitHub Copilot Test)DEV Community🌪️ Proof of Work: The To-Do List of Infinite RegretDEV CommunityShaping the UAE’s Digital Destiny: Building Sovereignty, Trust, and Resilience in the Cyber EraEE TimesA new dating app, Sonder, has a deliberately annoying sign-up process (and it’s working)TechCrunchPromoting late-gameplay BG3 composition contracts in the TD2 SDL portDEV CommunityArtificial Intelligence Is Facing a Crisis of Control—and the Industry Knows It - Council on Foreign RelationsGoogle News: AI SafetyBlack Hat USADark ReadingBlack Hat AsiaAI BusinessHow We Built an EdTech Platform That Scaled to 250K Daily UsersDEV CommunityRoguelike Devlog: Redesigning a Game UI With an AI 2D Game MakerDEV CommunityI spent days debugging a cron job that was "working fine"DEV CommunityLLM Agents Need a Nervous System, Not Just a BrainDEV CommunityThe 22,000 Token Tax: Why I Killed My MCP ServerDEV CommunityOpenSpec (Spec-Driven Development) Failed My Experiment — Instructions.md Was Simpler and FasterDEV CommunityI Asked AI to Do Agile Sprint Planning (GitHub Copilot Test)DEV Community🌪️ Proof of Work: The To-Do List of Infinite RegretDEV CommunityShaping the UAE’s Digital Destiny: Building Sovereignty, Trust, and Resilience in the Cyber EraEE TimesA new dating app, Sonder, has a deliberately annoying sign-up process (and it’s working)TechCrunchPromoting late-gameplay BG3 composition contracts in the TD2 SDL portDEV CommunityArtificial Intelligence Is Facing a Crisis of Control—and the Industry Knows It - Council on Foreign RelationsGoogle News: AI Safety

KARMA: Knowledge-Action Regularized Multimodal Alignment for Personalized Search at Taobao

arXiv cs.IRby Zhi Sun, Wenming Zhang, Yi Wei, Liren Yu, Zhixuan Zhang, Dan Ou, Haihong TangApril 1, 20262 min read0 views
Source Quiz

arXiv:2603.22779v2 Announce Type: replace Abstract: Large Language Models (LLMs) are equipped with profound semantic knowledge, making them a natural choice for injecting semantic generalization into personalized search systems. However, in practice we find that directly fine-tuning LLMs on industrial personalized tasks (e.g. next item prediction) often yields suboptimal results. We attribute this bottleneck to a critical Knowledge--Action Gap: the inherent conflict between preserving pre-trained semantic knowledge and aligning with specific personalized actions by discriminative objectives. Empirically, action-only training objectives induce Semantic Collapse, such as attention "sinks". This degradation severely cripples the LLM's generalization, failing to bring improvements to personali

View PDF HTML (experimental)

Abstract:Large Language Models (LLMs) are equipped with profound semantic knowledge, making them a natural choice for injecting semantic generalization into personalized search systems. However, in practice we find that directly fine-tuning LLMs on industrial personalized tasks (e.g. next item prediction) often yields suboptimal results. We attribute this bottleneck to a critical Knowledge--Action Gap: the inherent conflict between preserving pre-trained semantic knowledge and aligning with specific personalized actions by discriminative objectives. Empirically, action-only training objectives induce Semantic Collapse, such as attention "sinks". This degradation severely cripples the LLM's generalization, failing to bring improvements to personalized search systems. We propose KARMA (Knowledge--Action Regularized Multimodal Alignment), a unified framework that treats semantic reconstruction as a train-only regularizer. KARMA optimizes a next-interest embedding for retrieval (Action) while enforcing semantic decodability (Knowledge) through two complementary objectives: (i) history-conditioned semantic generation, which anchors optimization to the LLM's native next-token distribution, and (ii) embedding-conditioned semantic reconstruction, which constrains the interest embedding to remain semantically recoverable. On Taobao search system, KARMA mitigates semantic collapse (attention-sink analysis) and improves both action metrics and semantic fidelity. In ablations, semantic decodability yields up to +22.5 HR@200. With KARMA, we achieve +0.25 CTR AUC in ranking, +1.86 HR in pre-ranking and +2.51 HR in recalling. Deployed online with low inference overhead at ranking & pre-ranking stage, KARMA drives +0.9% increase in GMV.

Subjects:

Information Retrieval (cs.IR); Artificial Intelligence (cs.AI); Machine Learning (cs.LG)

Cite as: arXiv:2603.22779 [cs.IR]

(or arXiv:2603.22779v2 [cs.IR] for this version)

https://doi.org/10.48550/arXiv.2603.22779

arXiv-issued DOI via DataCite

Submission history

From: Zhi Sun [view email] [v1] Tue, 24 Mar 2026 04:13:30 UTC (1,195 KB) [v2] Tue, 31 Mar 2026 09:40:53 UTC (1,195 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by AI News Hub · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

modellanguage modeltraining

Knowledge Map

Knowledge Map
TopicsEntitiesSource
KARMA: Know…modellanguage mo…trainingannounceanalysispredictionarXiv cs.IR

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 199 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Models