VideoAtlas: Navigating Long-Form Video in Logarithmic Compute
Abstract
VideoAtlas enables lossless video representation through hierarchical grids, enabling efficient long-context processing via recursive language models with adaptive compute allocation.
Extending language models to video introduces two challenges: representation, where existing methods rely on lossy approximations, and long-context, where caption- or agent-based pipelines collapse video into text and lose visual fidelity. To overcome this, we introduce VideoAtlas, a task-agnostic environment to represent video as a hierarchical grid that is simultaneously lossless, navigable, scalable, caption- and preprocessing-free. An overview of the video is available at a glance, and any region can be recursively zoomed into, with the same visual representation used uniformly for the video, intermediate investigations, and the agent's memory, eliminating lossy text conversion end-to-end. This hierarchical structure ensures access depth grows only logarithmically with video length. For long-context, Recursive Language Models (RLMs) recently offered a powerful solution for long text, but extending them to visual domain requires a structured environment to recurse into, which VideoAtlas provides. VideoAtlas as a Markov Decision Process unlocks Video-RLM: a parallel Master-Worker architecture where a Master coordinates global exploration while Workers concurrently drill into assigned regions to accumulate lossless visual evidence. We demonstrate three key findings: (1)~logarithmic compute growth with video duration, further amplified by a 30-60\% multimodal cache hit rate arising from the grid's structural reuse. (2)~environment budgeting, where bounding the maximum exploration depth provides a principled compute-accuracy hyperparameter. (3)~emergent adaptive compute allocation that scales with question granularity. When scaling from 1-hour to 10-hour benchmarks, Video-RLM remains the most duration-robust method with minimal accuracy degradation, demonstrating that structured environment navigation is a viable and scalable paradigm for video understanding.
Community
What if understanding a video was more like navigating a map? 🤔
And what if that made compute scale logarithmically (not linearly) with video length?!
VideoAtlas: Navigating Long-Form Video in Logarithmic Compute
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- VideoThinker: Building Agentic VideoLLMs with LLM-Guided Tool Reasoning (2026)
- Hierarchical Long Video Understanding with Audiovisual Entity Cohesion and Agentic Search (2026)
- LongVideo-R1: Smart Navigation for Low-cost Long Video Understanding (2026)
- FocusGraph: Graph-Structured Frame Selection for Embodied Long Video Question Answering (2026)
- Event-Anchored Frame Selection for Effective Long-Video Understanding (2026)
- ReMoRa: Multimodal Large Language Model based on Refined Motion Representation for Long-Video Understanding (2026)
- QuickGrasp: Responsive Video-Language Querying Service via Accelerated Tokenization and Edge-Augmented Inference (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper