Live
Black Hat USAAI BusinessBlack Hat AsiaAI BusinessEarly Career Award recipient Aleksandra Ćiprijanović aims to create universal AI analysis framework - Fermilab (.gov)Google News: AIExclusive: Miravoice, Builder Of An AI ‘Interviewer’ To Conduct Phone Surveys, Raises $6.3MCrunchbase NewsMoltbook risks: The dangers of AI-to-AI interactions in health carePhys.org AIMaul: Shadow Lord Will Return for Season 2GizmodoMicrosoft Aims to Create Large Cutting-Edge AI Models By 2027Bloomberg TechnologyHow Disney Imagineers are using AI and robotics to reshape the company’s theme parksFast Company TechA jury says Meta and Google hurt a kid. What now?The Verge AII have always seen myself as ‘progressive’ – but with AI it’s time to hit the brakes - The GuardianGoogle News: AIOpenAI Teams Up with Smartly to Create Chatty Ads Inside ChatGPT - TipRanksGoogle News: ChatGPTDOJ to Appeal Court Order Halting Trump’s Ban on Anthropic AIBloomberg TechnologyCapacity and speed: why TikTok shelved its second Irish data centreSilicon RepublicBlack Hat USAAI BusinessBlack Hat AsiaAI BusinessEarly Career Award recipient Aleksandra Ćiprijanović aims to create universal AI analysis framework - Fermilab (.gov)Google News: AIExclusive: Miravoice, Builder Of An AI ‘Interviewer’ To Conduct Phone Surveys, Raises $6.3MCrunchbase NewsMoltbook risks: The dangers of AI-to-AI interactions in health carePhys.org AIMaul: Shadow Lord Will Return for Season 2GizmodoMicrosoft Aims to Create Large Cutting-Edge AI Models By 2027Bloomberg TechnologyHow Disney Imagineers are using AI and robotics to reshape the company’s theme parksFast Company TechA jury says Meta and Google hurt a kid. What now?The Verge AII have always seen myself as ‘progressive’ – but with AI it’s time to hit the brakes - The GuardianGoogle News: AIOpenAI Teams Up with Smartly to Create Chatty Ads Inside ChatGPT - TipRanksGoogle News: ChatGPTDOJ to Appeal Court Order Halting Trump’s Ban on Anthropic AIBloomberg TechnologyCapacity and speed: why TikTok shelved its second Irish data centreSilicon Republic
AI NEWS HUBbyEIGENVECTOREigenvector

QuadRank: Engineering a High Throughput Rank

arXiv cs.DSby R. Groot KoerkampApril 2, 20262 min read0 views
Source Quiz

arXiv:2602.04103v2 Announce Type: replace Abstract: Given a text, a query $\mathsf{rank}(q, c)$ counts the number of occurrences of character $c$ among the first $q$ characters of the text. Space-efficient methods to answer these rank queries form an important building block in many succinct data structures. For example, the FM-index is a widely used data structure that uses rank queries to locate all occurrences of a pattern in a text. In bioinformatics applications, the goal is usually to process a given input as fast as possible. Thus, data structures should have high throughput when used with many threads. Contributions. For the binary alphabet, we develop BiRank with 3.28% space overhead. It merges the central ideas of two recent papers: (1) we interleave (inline) offsets in each cach

View PDF HTML (experimental)

Abstract:Given a text, a query $\mathsf{rank}(q, c)$ counts the number of occurrences of character $c$ among the first $q$ characters of the text. Space-efficient methods to answer these rank queries form an important building block in many succinct data structures. For example, the FM-index is a widely used data structure that uses rank queries to locate all occurrences of a pattern in a text. In bioinformatics applications, the goal is usually to process a given input as fast as possible. Thus, data structures should have high throughput when used with many threads. Contributions. For the binary alphabet, we develop BiRank with 3.28% space overhead. It merges the central ideas of two recent papers: (1) we interleave (inline) offsets in each cache line of the underlying bit vector [Laws et al., 2024], reducing cache-misses, and (2) these offsets are to the middle of each block so that only half of them need popcounting [Gottlieb and Reinert, 2025]. In QuadRank (14.4% space overhead), we extend these techniques to the $\sigma=4$ (DNA) alphabet. Both data structures require only a single cache miss per query, making them highly suitable for high-throughput and memory-bound settings. To enable efficient batch-processing, we support prefetching the cache lines required to answer upcoming queries. Results. BiRank and QuadRank are around $1.5\times$ and $2\times$ faster than similar-overhead methods that do not use inlining. Prefetching gives an additional $2\times$ speedup, at which point the dual-channel DDR4 RAM bandwidth becomes a hard limit on the total throughput. With prefetching, both methods outperform all other methods apart from SPIDER [Laws et al., 2024] by $2\times$. When using QuadRank with prefetching in a toy count-only FM-index, QuadFm, this results in a smaller size and up to $4\times$ speedup over Genedex, a state-of-the-art batching FM-index implementation.

Comments: SEA-2026; 23 pages; 8 figures

Subjects:

Data Structures and Algorithms (cs.DS)

Cite as: arXiv:2602.04103 [cs.DS]

(or arXiv:2602.04103v2 [cs.DS] for this version)

https://doi.org/10.48550/arXiv.2602.04103

arXiv-issued DOI via DataCite

Submission history

From: Ragnar Groot Koerkamp [view email] [v1] Wed, 4 Feb 2026 00:41:00 UTC (4,089 KB) [v2] Wed, 1 Apr 2026 02:21:33 UTC (1,860 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

announceapplicationpaper

Knowledge Map

Knowledge Map
TopicsEntitiesSource
QuadRank: E…announceapplicationpaperarxivarXiv cs.DS

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 222 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Products