Live
Black Hat USAAI BusinessBlack Hat AsiaAI BusinessWhat’s the point of an AI novel?Financial Times TechI Built an AI Content Pipeline That Publishes 4 SEO-Optimized Articles Per Day — Here's the ArchitectureDEV CommunityMy Reading Journey: Jan-Mar 2026DEV CommunityBuilding a Second Brain for Claude CodeDEV CommunityThe Perfect CLAUDE.md: How to Set Up Your Project for Agentic CodingDEV CommunityClaude Code Advanced Workflow: Subagents, Commands & Multi-SessionDEV CommunityHow to Build a Custom MCP Server for Claude Code: A Step-by-Step TutorialDEV CommunityHetzner Cloud for AI Projects — Complete GPU Server Setup & Cost Breakdown 2026DEV CommunityOllama + Open WebUI Self-Hosting Guide 2026 — Run Your Own AI for $0DEV CommunityHow to Self-Host Your Entire Dev Stack for Under $20/Month in 2026DEV CommunityTop 15 MCP Servers Every Developer Should Install in 2026DEV CommunityHow to emotionally grasp the risks of AI SafetyLessWrong AIBlack Hat USAAI BusinessBlack Hat AsiaAI BusinessWhat’s the point of an AI novel?Financial Times TechI Built an AI Content Pipeline That Publishes 4 SEO-Optimized Articles Per Day — Here's the ArchitectureDEV CommunityMy Reading Journey: Jan-Mar 2026DEV CommunityBuilding a Second Brain for Claude CodeDEV CommunityThe Perfect CLAUDE.md: How to Set Up Your Project for Agentic CodingDEV CommunityClaude Code Advanced Workflow: Subagents, Commands & Multi-SessionDEV CommunityHow to Build a Custom MCP Server for Claude Code: A Step-by-Step TutorialDEV CommunityHetzner Cloud for AI Projects — Complete GPU Server Setup & Cost Breakdown 2026DEV CommunityOllama + Open WebUI Self-Hosting Guide 2026 — Run Your Own AI for $0DEV CommunityHow to Self-Host Your Entire Dev Stack for Under $20/Month in 2026DEV CommunityTop 15 MCP Servers Every Developer Should Install in 2026DEV CommunityHow to emotionally grasp the risks of AI SafetyLessWrong AI
AI NEWS HUBbyEIGENVECTOREigenvector

Out of Sight but Not Out of Mind: Hybrid Memory for Dynamic Video World Models

HuggingFace Papersby Kaijin Chen ,March 26, 20262 min read1 views
Source Quiz

Hybrid Memory enables video world models to maintain consistent tracking of dynamic subjects during occlusion by combining archival storage for static backgrounds with active tracking for moving objects, using a specialized architecture with tokenized memory and spatiotemporal retrieval mechanisms. (68 upvotes on HuggingFace)

Abstract

Hybrid Memory enables video world models to maintain consistent tracking of dynamic subjects during occlusion by combining archival storage for static backgrounds with active tracking for moving objects, using a specialized architecture with tokenized memory and spatiotemporal retrieval mechanisms.

AI-generated summary

Video world models have shown immense potential in simulating the physical world, yet existing memory mechanisms primarily treat environments as static canvases. When dynamic subjects hide out of sight and later re-emerge, current methods often struggle, leading to frozen, distorted, or vanishing subjects. To address this, we introduce Hybrid Memory, a novel paradigm requiring models to simultaneously act as precise archivists for static backgrounds and vigilant trackers for dynamic subjects, ensuring motion continuity during out-of-view intervals. To facilitate research in this direction, we construct HM-World, the first large-scale video dataset dedicated to hybrid memory. It features 59K high-fidelity clips with decoupled camera and subject trajectories, encompassing 17 diverse scenes, 49 distinct subjects, and meticulously designed exit-entry events to rigorously evaluate hybrid coherence. Furthermore, we propose HyDRA, a specialized memory architecture that compresses memory into tokens and utilizes a spatiotemporal relevance-driven retrieval mechanism. By selectively attending to relevant motion cues, HyDRA effectively preserves the identity and motion of hidden subjects. Extensive experiments on HM-World demonstrate that our method significantly outperforms state-of-the-art approaches in both dynamic subject consistency and overall generation quality.

View arXiv page View PDF Project page GitHub 86 Add to collection

Get this paper in your agent:

hf papers read 2603.25716

Don't have the latest CLI?

curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 1

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2603.25716 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2603.25716 in a Space README.md to link it from this page.

Collections including this paper 6

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

researchpaperarxiv

Knowledge Map

Knowledge Map
TopicsEntitiesSource
Out of Sigh…researchpaperarxivvideo world…hybrid memo…dynamic sub…HuggingFace…

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 165 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Research Papers