Average-Case Reductions for $k$-XOR and Tensor PCA
Hi there, little explorer! Imagine you have a big puzzle, but some pieces are hidden, and some are tricky.
Scientists are like detectives trying to find the hidden pieces in two super-duper hard puzzles called "k-XOR" and "Tensor PCA." They're like treasure hunts!
This news says they found a special secret map! This map shows how to turn one hard puzzle into another, maybe a little easier one. It's like finding out that if you can solve one type of treasure hunt, you can also solve another!
This helps them understand which puzzles are the most tricky for our smart computers to solve. It's like learning which treasure chests need the strongest keys! Yay for clever detectives!
arXiv:2601.19016v2 Announce Type: replace-cross Abstract: We study the computational properties of two canonical planted average-case problems -- noisy planted $k$-XOR and Tensor PCA -- by formally unifying them into a family of planted problems parametrized by tensor order $k$, number of entries $m$, and noise level $\delta$. We build a wide range of poly-time average-case reductions within this family, across all regimes $m \in [1, n^k]$. In the denser $m \geq n^{k/2}$ regime, our reductions preserve proximity to the computational threshold, and, as a central application, reduce conjectured-hard $k$-XOR instances with $m \approx n^{k/2}$ to conjectured-hard instances of Tensor PCA. Additionally, we give new order-reducing maps at fixed densities (e.g., $5\to 4$ for $k$-XOR with $m \appro
View PDF HTML (experimental)
Abstract:We study the computational properties of two canonical planted average-case problems -- noisy planted $k$-XOR and Tensor PCA -- by formally unifying them into a family of planted problems parametrized by tensor order $k$, number of entries $m$, and noise level $\delta$. We build a wide range of poly-time average-case reductions within this family, across all regimes $m \in [1, n^k]$. In the denser $m \geq n^{k/2}$ regime, our reductions preserve proximity to the computational threshold, and, as a central application, reduce conjectured-hard $k$-XOR instances with $m \approx n^{k/2}$ to conjectured-hard instances of Tensor PCA. Additionally, we give new order-reducing maps at fixed densities (e.g., $5\to 4$ for $k$-XOR with $m \approx n^{k/2}$ entries and $7\to 4$ for Tensor PCA). In the sparser $m \leq n^{k/2}$ regime, we relate instances of different orders, reducing, for example, $7$-XOR with $m = n^{3.4}$ to the classical setting of $3$-XOR with $m = \widetilde\Theta(n^{1.4})$. Taken together, these results establish a hardness partial order in the space of planted tensor models.
Comments: 112 pages, 6 figures
Subjects:
Computational Complexity (cs.CC); Cryptography and Security (cs.CR); Probability (math.PR); Statistics Theory (math.ST)
Cite as: arXiv:2601.19016 [cs.CC]
(or arXiv:2601.19016v2 [cs.CC] for this version)
https://doi.org/10.48550/arXiv.2601.19016
arXiv-issued DOI via DataCite
Submission history
From: Alina Harbuzova [view email] [v1] Mon, 26 Jan 2026 23:05:54 UTC (175 KB) [v2] Thu, 2 Apr 2026 16:45:31 UTC (179 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelannounceapplication
Curia-2: Scaling Self-Supervised Learning for Radiology Foundation Models
The rapid growth of medical imaging has fueled the development of Foundation Models (FMs) to reduce the growing, unsustainable workload on radiologists. While recent FMs have shown the power of large-scale pre-training to CT and MRI analysis, there remains significant room to optimize how these models learn from complex radiological volumes. Building upon the Curia framework, this work introduces Curia-2, which significantly improves the original pre-training strategy and representation quality to better capture the specificities of radiological data. The proposed methodology enables scaling t — Antoine Saporta, Baptiste Callard, Corentin Dancette

Demographic Parity Tails for Regression
Demographic parity (DP) is a widely studied fairness criterion in regression, enforcing independence between the predictions and sensitive attributes. However, constraining the entire distribution can degrade predictive accuracy and may be unnecessary for many applications, where fairness concerns are localized to specific regions of the distribution. To overcome this issue, we propose a new framework for regression under DP that focuses on the tails of target distribution across sensitive groups. Our methodology builds on optimal transport theory. By enforcing fairness constraints only over t — Naht Sinh Le, Christophe Denis, Mohamed Hebiri

Apriel-Reasoner: RL Post-Training for General-Purpose and Efficient Reasoning
Building general-purpose reasoning models using reinforcement learning with verifiable rewards (RLVR) across diverse domains has been widely adopted by frontier open-weight models. However, their training recipes and domain mixtures are often not disclosed. Joint optimization across domains poses significant challenges: domains vary widely in rollout length, problem difficulty and sample efficiency. Further, models with long chain-of-thought traces increase inference cost and latency, making efficiency critical for practical deployment. We present Apriel-Reasoner, trained with a fully reproduc — Rafael Pardinas, Ehsan Kamalloo, David Vazquez
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Products

🦀 Rust Foundations — The Stuff That Finally Made Things Click
"Rust compiler and Clippy are the biggest tsunderes — they'll shout at you for every small mistake, but in the end… they just want your code to be perfect." Why I Even Started Rust I didn't pick Rust out of curiosity or hype. I had to. I'm working as a Rust dev at Garden Finance , where I built part of a Wallet-as-a-Service infrastructure. Along with an Axum backend, we had this core Rust crate ( standard-rs ) handling signing and broadcasting transactions across: Bitcoin EVM chains Sui Solana Starknet And suddenly… memory safety wasn't "nice to have" anymore. It was everything. Rust wasn't just a language — it was a guarantee . But yeah… in the beginning? It felt like the compiler hated me :( So I'm writing this to explain Rust foundations in the simplest way possible — from my personal n

Why Standard HTTP Libraries Are Dead for Web Scraping (And How to Fix It)
If you are building a data extraction pipeline in 2026 and your core network request looks like Ruby’s Net::HTTP.get(URI(url)) or Python's requests.get(url) , you are already blocked. The era of bypassing bot detection by rotating datacenter IPs and pasting a fake Mozilla/5.0 User-Agent string is long gone. Modern Web Application Firewalls (WAFs) like Cloudflare, Akamai, and DataDome don’t just read your headers anymore—they interrogate the cryptographic foundation of your connection. Here is a deep dive into why standard HTTP libraries actively sabotage your scraping infrastructure, and how I built a polyglot sidecar architecture to bypass Layer 4–7 fingerprinting entirely. The Fingerprint You Didn’t Know You Had When your code opens a secure connection to a server, long before the first

Tired of Zillow Blocking Scrapers — Here's What Actually Works in 2026
If you've ever tried scraping Zillow with BeautifulSoup or Selenium, you know the pain. CAPTCHAs, IP bans, constantly changing HTML selectors, headless browser detection — it's an arms race you're not going to win. I spent way too long fighting anti-bot systems before switching to an API-based approach. This post walks through how to pull Zillow property data, search listings, get Zestimates, and export everything to CSV/Excel — all with plain Python and zero browser automation. What You'll Need Python 3.7+ The requests library ( pip install requests ) A free API key from RealtyAPI That's it. No Selenium. No Playwright. No proxy rotation. Getting Started: Your First Property Lookup Let's start simple — get full property details for a single address: import requests url = " https://zillow.r

Why Gaussian Diffusion Models Fail on Discrete Data?
arXiv:2604.02028v1 Announce Type: new Abstract: Diffusion models have become a standard approach for generative modeling in continuous domains, yet their application to discrete data remains challenging. We investigate why Gaussian diffusion models with the DDPM solver struggle to sample from discrete distributions that are represented as a mixture of delta-distributions in the continuous space. Using a toy Random Hierarchy Model, we identify a critical sampling interval in which the density of noisified data becomes multimodal. In this regime, DDPM occasionally enters low-density regions between modes producing out-of-distribution inputs for the model and degrading sample quality. We show that existing heuristics, including self-conditioning and a solver we term q-sampling, help alleviate


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!