Live
🔥 google-research/timesfmGitHub Trending🔥 aliasrobotics/caiGitHub Trending🔥 ComposioHQ/awesome-claude-skillsGitHub Trending🔥 SkyworkAI/Matrix-GameGitHub Trending🔥 sponsors/vas3kGitHub Trending🔥 sponsors/khoj-aiGitHub Trending🔥 PaddlePaddle/PaddleOCRGitHub TrendingTest: 15% of Americans say they would work for AI bossTechCrunch AIAutoMS: Multi-Agent Evolutionary Search for Cross-Physics Inverse Microstructure DesignarXivMultiverse: Language-Conditioned Multi-Game Level Blending via Shared RepresentationarXivMediHive: A Decentralized Agent Collective for Medical ReasoningarXivBitboard version of Tetris AIarXivThe Price of Meaning: Why Every Semantic Memory System ForgetsarXivWhen Verification Hurts: Asymmetric Effects of Multi-Agent Feedback in Logic Proof TutoringarXivQuantification of Credal Uncertainty: A Distance-Based ApproacharXiv🔥 google-research/timesfmGitHub Trending🔥 aliasrobotics/caiGitHub Trending🔥 ComposioHQ/awesome-claude-skillsGitHub Trending🔥 SkyworkAI/Matrix-GameGitHub Trending🔥 sponsors/vas3kGitHub Trending🔥 sponsors/khoj-aiGitHub Trending🔥 PaddlePaddle/PaddleOCRGitHub TrendingTest: 15% of Americans say they would work for AI bossTechCrunch AIAutoMS: Multi-Agent Evolutionary Search for Cross-Physics Inverse Microstructure DesignarXivMultiverse: Language-Conditioned Multi-Game Level Blending via Shared RepresentationarXivMediHive: A Decentralized Agent Collective for Medical ReasoningarXivBitboard version of Tetris AIarXivThe Price of Meaning: Why Every Semantic Memory System ForgetsarXivWhen Verification Hurts: Asymmetric Effects of Multi-Agent Feedback in Logic Proof TutoringarXivQuantification of Credal Uncertainty: A Distance-Based ApproacharXiv

TAPS: Task Aware Proposal Distributions for Speculative Sampling

HuggingFace PapersMarch 27, 20262 min read0 views
Source Quiz

Speculative decoding effectiveness depends on draft model training data alignment with downstream tasks, with specialized drafters performing better when combined through confidence-based routing rather than simple averaging. (2 upvotes on HuggingFace)

Published on Mar 27

Authors:

,

,

,

Abstract

Speculative decoding effectiveness depends on draft model training data alignment with downstream tasks, with specialized drafters performing better when combined through confidence-based routing rather than simple averaging.

AI-generated summary

Speculative decoding accelerates autoregressive generation by letting a lightweight draft model propose future tokens that a larger target model then verifies in parallel. In practice, however, draft models are usually trained on broad generic corpora, which leaves it unclear how much speculative decoding quality depends on the draft training distribution. We study this question with lightweight HASS and EAGLE-2 drafters trained on MathInstruct, ShareGPT, and mixed-data variants, evaluated on MT-Bench, GSM8K, MATH-500, and SVAMP. Measured by acceptance length, task-specific training yields clear specialization: MathInstruct-trained drafts are strongest on reasoning benchmarks, while ShareGPT-trained drafts are strongest on MT-Bench. Mixed-data training improves robustness, but larger mixtures do not dominate across decoding temperatures. We also study how to combine specialized drafters at inference time. Naive checkpoint averaging performs poorly, whereas confidence-based routing improves over single-domain drafts and merged-tree verification yields the highest acceptance length overall for both backbones. Finally, confidence is a more useful routing signal than entropy: rejected tokens tend to have higher entropy, but confidence produces much clearer benchmark-level routing decisions. These results show that speculative decoding quality depends not only on draft architecture, but also on the match between draft training data and downstream workload, and that specialized drafters are better combined at inference time than in weight space.

View arXiv page View PDF GitHub 0 Add to collection

Models citing this paper 10

Browse 10 models citing this paper

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2603.27027 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.

Original source

HuggingFace Papers

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by AI News Hub · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

Knowledge Map

Knowledge Map
TopicsEntitiesSource
TAPS: Task …researchpaperarxivspeculative…draft modelautoregress…HuggingFace…

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 336 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Research Papers