TLDR: VIDEO-RTS is a novel AI model that significantly improves video reasoning by using a highly data-efficient reinforcement learning approach (skipping costly fine-tuning) combined with an adaptive “sparse-to-dense” inference strategy. This allows it to achieve higher accuracy with far less training data and more efficient use of computational resources during analysis.
In the rapidly evolving field of artificial intelligence, particularly concerning how large language models (LLMs) understand and reason about video content, a new approach called VIDEO-RTS is making waves. This innovative system aims to tackle significant challenges faced by current methods, such as the high costs associated with data collection and fine-tuning.
Traditional methods for video reasoning often demand extensive supervised fine-tuning (SFT) using vast amounts of video data and detailed Chain-of-Thought (CoT) annotations. This process is not only expensive but also difficult to scale for more complex tasks. VIDEO-RTS, however, offers a fresh perspective by integrating data-efficient reinforcement learning (RL) with a clever video-adaptive test-time scaling (TTS) strategy.
One of the core innovations of VIDEO-RTS lies in its training methodology. Unlike its predecessors, it completely bypasses the resource-intensive SFT step. Instead, it employs a pure-RL training approach, which relies on output-based rewards. This means the model learns effectively without needing additional annotations or extensive fine-tuning, drastically improving data efficiency. For instance, VIDEO-RTS achieves comparable performance to systems that use 165,000 SFT examples plus 4,000 RL examples, but it does so with merely 6,000 video-question pairs for its RL training. This remarkable efficiency is achieved by adapting a technique called Group Relative Policy Optimization (GRPO), which simplifies the reward mechanism to focus directly on the correctness of the final answer, alongside a ‘format reward’ that encourages a structured reasoning process.
The second major advancement introduced by VIDEO-RTS is its dynamic sparse-to-dense video test-time scaling. Researchers observed that beyond a certain point (around 6,000 training samples), adding more video question-answering data yielded only marginal improvements in RL training. This insight led to the idea of reallocating computational resources from the training phase to the inference stage. The sparse-to-dense TTS strategy allows the model to adaptively select the appropriate temporal context for a video. Initially, it uses a sparse set of frames. If its multiple reasoning attempts lead to inconsistent answers, it dynamically adds more frames until a consensus is reached or a maximum frame limit is hit. This adaptive approach ensures that computational effort is applied precisely where and when it’s needed, based on the complexity of each video query.
The results of VIDEO-RTS are compelling. Across multiple video reasoning benchmarks, it surpasses existing models by an average of 2.4% in accuracy, all while utilizing only 3.6% of the training samples typically required. For example, it achieved a 4.2% improvement on Video-Holmes, a challenging new benchmark, and a 2.6% improvement on MMVU. This demonstrates that the pure RL training and the adaptive video TTS are complementary, with RL enhancing the model’s reasoning capabilities and TTS optimizing the use of visual information.
Also Read:
- Tempo-R0: Advancing Video Understanding with Enhanced Temporal Grounding
- Connecting Vision and Language: A Graph-Based Approach for Detailed Video Descriptions
In essence, VIDEO-RTS represents a significant step forward in making video reasoning with AI models more efficient and effective. By rethinking how reinforcement learning is applied and introducing an adaptive inference strategy, it sets a new standard for performance with substantially reduced data and computational overhead. For more details, you can refer to the original research paper.


