Real-Time and Scalable Zak-OTFS Receiver Processing on GPUs
arXiv:2604.02266v1 Announce Type: new Abstract: Orthogonal time frequency space (OTFS) modulation offers superior robustness to high-mobility channels compared to conventional orthogonal frequency-division multiplexing (OFDM) waveforms. However, its explicit delay-Doppler (DD) domain representation incurs substantial signal processing complexity, especially with increased DD domain grid sizes. To address this challenge, we present a scalable, real-time Zak-OTFS receiver architecture on GPUs through hardware--algorithm co-design that exploits DD-domain channel sparsity. Our design leverages compact matrix operations for key processing stages, a branchless iterative equalizer, and a structured sparse channel matrix of the DD domain channel matrix to significantly reduce computational and mem
View PDF HTML (experimental)
Abstract:Orthogonal time frequency space (OTFS) modulation offers superior robustness to high-mobility channels compared to conventional orthogonal frequency-division multiplexing (OFDM) waveforms. However, its explicit delay-Doppler (DD) domain representation incurs substantial signal processing complexity, especially with increased DD domain grid sizes. To address this challenge, we present a scalable, real-time Zak-OTFS receiver architecture on GPUs through hardware--algorithm co-design that exploits DD-domain channel sparsity. Our design leverages compact matrix operations for key processing stages, a branchless iterative equalizer, and a structured sparse channel matrix of the DD domain channel matrix to significantly reduce computational and memory overhead. These optimizations enable low-latency processing that consistently meets the 99.9-th percentile real-time processing deadline. The proposed system achieves up to 906.52 Mbps throughput with a DD grid size of (16384,32) using 16QAM modulation over 245.76 MHz bandwidth. Extensive evaluations under a Vehicular-A channel model demonstrate strong scalability and robust performance across CPU (Intel Xeon) and multiple GPU platforms (NVIDIA Jetson Orin, RTX 6000 Ada, A100, and H200), highlighting the effectiveness of compute-aware Zak-OTFS receiver design for next-generation (NextG) high-mobility communication systems.
Comments: This work has been submitted to the IEEE for possible publication
Subjects:
Signal Processing (eess.SP); Networking and Internet Architecture (cs.NI)
Cite as: arXiv:2604.02266 [eess.SP]
(or arXiv:2604.02266v1 [eess.SP] for this version)
https://doi.org/10.48550/arXiv.2604.02266
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Chung-Hsuan Tung [view email] [v1] Thu, 2 Apr 2026 16:59:32 UTC (552 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelannounceplatform
Production RAG: From Anti-Patterns to Platform Engineering
RAG is a distributed system . It becomes clear when moving beyond demos into production. It consists of independent services such as ingestion, retrieval, inference, orchestration, and observability. Each component introduces its own latency, scaling characteristics, and failure modes, making coordination, observability, and fault tolerance essential. RAG flowchart In regulated environments such as banking, these systems must also satisfy strict governance, auditability, and change-control requirements aligned with standards like SOX and PCI DSS. This article builds on existing frameworks like 12 Factor Agents (Dex Horthy)¹ and Google’s 16 Factor App² by exploring key anti-patterns and introducing the pillars required to take a typical RAG pipeline to production. I’ve included code snippet

Word2Vec Explained: The Moment Words Became Relations
How models first learned meaning from context — and why that changed everything In the first post, we built the base layer: Text → Tokens → Numbers → (lots of math) → Tokens → Text In the second post, we stayed with the deeper question: Once words become numbers, how does meaning not disappear? We saw that the answer is not “because numbers are magical.” The answer is this: the numbers are learned in a space that preserves relationships. That was the real story of embeddings. Now we are ready for the next step. Because once you accept that words can become numbers without losing meaning, the next question becomes unavoidable: How are those numbers actually learned? This is where Word2Vec enters the story. And Word2Vec matters for more than historical reasons. It was not just a clever neura

Chinese AI rivals clash over Anthropic’s OpenClaw exit amid global token crunch
Chinese tech companies are engaged in a public war of words as they compete to capitalise on US start-up Anthropic’s decision to pull its industry-leading Claude models from open-source AI agent tool OpenClaw. The development comes as AI agents have triggered a huge increase in demand for AI tokens – the core metric of AI usage – raising questions about the long-term ability of industry players to meet this demand amid a growing global crunch in computational power. On Sunday, Anthropic...
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

TurboQuant Explained: Extreme AI Compression for Faster, Cheaper LLM Inference and Vector Search
If you’ve been following the “long-context” wave in AI, you’ve probably heard the same story: bigger context windows feel magical… until… Continue reading on Towards AI »

Word2Vec Explained: The Moment Words Became Relations
How models first learned meaning from context — and why that changed everything In the first post, we built the base layer: Text → Tokens → Numbers → (lots of math) → Tokens → Text In the second post, we stayed with the deeper question: Once words become numbers, how does meaning not disappear? We saw that the answer is not “because numbers are magical.” The answer is this: the numbers are learned in a space that preserves relationships. That was the real story of embeddings. Now we are ready for the next step. Because once you accept that words can become numbers without losing meaning, the next question becomes unavoidable: How are those numbers actually learned? This is where Word2Vec enters the story. And Word2Vec matters for more than historical reasons. It was not just a clever neura

Chinese AI rivals clash over Anthropic’s OpenClaw exit amid global token crunch
Chinese tech companies are engaged in a public war of words as they compete to capitalise on US start-up Anthropic’s decision to pull its industry-leading Claude models from open-source AI agent tool OpenClaw. The development comes as AI agents have triggered a huge increase in demand for AI tokens – the core metric of AI usage – raising questions about the long-term ability of industry players to meet this demand amid a growing global crunch in computational power. On Sunday, Anthropic...


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!