Live
Black Hat USAAI BusinessBlack Hat AsiaAI BusinessExplained: The Source Code Leak that hit AI Giant Anthropic - Cyber MagazineGoogle News: ClaudeDespite Skepticism, Survey Shows Widespread AI Use at Cal State - Inside Higher EdGoogle News: ChatGPTBig Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.Dev.to AIYour AI Agent Did Something It Wasn't Supposed To. Now What?Dev.to AITrust drives Korea’s generative AI adoption; usability and interaction sustain use - CHOSUNBIZ - ChosunbizGoogle News: Generative AIThe Model You Love Is Probably Just the One You UseO'Reilly Radar3 of Your AI Agents Crashed and You Found Out From CustomersDev.to AIYour AI Agent Is Running Wild and You Can't Stop ItDev.to AIYour AI Agent Spent $500 Overnight and Nobody NoticedDEV CommunityWhy Software Project Estimates Are Always Wrong (And How to Fix It)DEV CommunityChatGPT vs. Claude: 7 real-life benchmarks that crown the 2026 AI Madness Champion - Tom's GuideGoogle News: ChatGPTHow to Build a Responsible AI Framework for Transparent, Ethical, and Secure AppsDev.to AIBlack Hat USAAI BusinessBlack Hat AsiaAI BusinessExplained: The Source Code Leak that hit AI Giant Anthropic - Cyber MagazineGoogle News: ClaudeDespite Skepticism, Survey Shows Widespread AI Use at Cal State - Inside Higher EdGoogle News: ChatGPTBig Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.Dev.to AIYour AI Agent Did Something It Wasn't Supposed To. Now What?Dev.to AITrust drives Korea’s generative AI adoption; usability and interaction sustain use - CHOSUNBIZ - ChosunbizGoogle News: Generative AIThe Model You Love Is Probably Just the One You UseO'Reilly Radar3 of Your AI Agents Crashed and You Found Out From CustomersDev.to AIYour AI Agent Is Running Wild and You Can't Stop ItDev.to AIYour AI Agent Spent $500 Overnight and Nobody NoticedDEV CommunityWhy Software Project Estimates Are Always Wrong (And How to Fix It)DEV CommunityChatGPT vs. Claude: 7 real-life benchmarks that crown the 2026 AI Madness Champion - Tom's GuideGoogle News: ChatGPTHow to Build a Responsible AI Framework for Transparent, Ethical, and Secure AppsDev.to AI

Privacy Guard & Token Parsimony by Prompt and Context Handling and LLM Routing

arXiv cs.CRby Alessio LangiuApril 1, 20261 min read0 views
Source Quiz

arXiv:2603.28972v1 Announce Type: new Abstract: The large-scale adoption of Large Language Models (LLMs) forces a trade-off between operational cost (OpEx) and data privacy. Current routing frameworks reduce costs but ignore prompt sensitivity, exposing users and institutions to leakage risks towards third-party cloud providers. We formalise the "Inseparability Paradigm": advanced context management intrinsically coincides with privacy management. We propose a local "Privacy Guard" -- a holistic contextual observer powered by an on-premise Small Language Model (SLM) -- that performs abstractive summarisation and Automatic Prompt Optimisation (APO) to decompose prompts into focused sub-tasks, re-routing high-risk queries to Zero-Trust or NDA-covered models. This dual mechanism simultaneousl

View PDF HTML (experimental)

Abstract:The large-scale adoption of Large Language Models (LLMs) forces a trade-off between operational cost (OpEx) and data privacy. Current routing frameworks reduce costs but ignore prompt sensitivity, exposing users and institutions to leakage risks towards third-party cloud providers. We formalise the "Inseparability Paradigm": advanced context management intrinsically coincides with privacy management. We propose a local "Privacy Guard" -- a holistic contextual observer powered by an on-premise Small Language Model (SLM) -- that performs abstractive summarisation and Automatic Prompt Optimisation (APO) to decompose prompts into focused sub-tasks, re-routing high-risk queries to Zero-Trust or NDA-covered models. This dual mechanism simultaneously eliminates sensitive inference vectors (Zero Leakage) and reduces cloud token payloads (OpEx Reduction). A LIFO-based context compacting mechanism further bounds working memory, limiting the emergent leakage surface. We validate the framework through a 2x2 benchmark (Lazy vs. Expert users; Personal vs. Institutional secrets) on a 1,000-sample dataset, achieving a 45% blended OpEx reduction, 100% redaction success on personal secrets, and -- via LLM-as-a-Judge evaluation -- an 85% preference rate for APO-compressed responses over raw baselines. Our results demonstrate that Token Parsimony and Zero Leakage are mathematically dual projections of the same contextual compression operator.

Subjects:

Cryptography and Security (cs.CR); Artificial Intelligence (cs.AI)

Cite as: arXiv:2603.28972 [cs.CR]

(or arXiv:2603.28972v1 [cs.CR] for this version)

https://doi.org/10.48550/arXiv.2603.28972

arXiv-issued DOI via DataCite

Submission history

From: Alessio Langiu [view email] [v1] Mon, 30 Mar 2026 20:16:42 UTC (1,689 KB)

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by AI News Hub · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

modellanguage modelbenchmark

Knowledge Map

Knowledge Map
TopicsEntitiesSource
Privacy Gua…modellanguage mo…benchmarkannouncevaluationemergentarXiv cs.CR

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 202 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Models