OpenGVL - Benchmarking Visual Temporal Progress for Data Curation
Abstract
OpenGVL is a benchmark for task progress prediction in robotics using vision-language models, showing open-source models underperform compared to closed-source ones and enabling automated data curation.
Data scarcity remains one of the most limiting factors in driving progress in robotics. However, the amount of available robotics data in the wild is growing exponentially, creating new opportunities for large-scale data utilization. Reliable temporal task completion prediction could help automatically annotate and curate this data at scale. The Generative Value Learning (GVL) approach was recently proposed, leveraging the knowledge embedded in vision-language models (VLMs) to predict task progress from visual observations. Building upon GVL, we propose OpenGVL, a comprehensive benchmark for estimating task progress across diverse challenging manipulation tasks involving both robotic and human embodiments. We evaluate the capabilities of publicly available open-source foundation models, showing that open-source model families significantly underperform closed-source counterparts, achieving only approximately 70% of their performance on temporal progress prediction tasks. Furthermore, we demonstrate how OpenGVL can serve as a practical tool for automated data curation and filtering, enabling efficient quality assessment of large-scale robotics datasets. We release the benchmark along with the complete codebase at github.com/budzianowski/opengvl{OpenGVL}.
Community
Check out our leaderboard on HF!
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Galaxea Open-World Dataset and G0 Dual-System VLA Model (2025)
- A Vision-Language-Action-Critic Model for Robotic Real-World Reinforcement Learning (2025)
- ROVER: Recursive Reasoning Over Videos with Vision-Language Models for Embodied Tasks (2025)
- Few-Shot Vision-Language Reasoning for Satellite Imagery via Verifiable Rewards (2025)
- MolmoAct: Action Reasoning Models that can Reason in Space (2025)
- RealMirror: A Comprehensive, Open-Source Vision-Language-Action Platform for Embodied AI (2025)
- SimpleVLA-RL: Scaling VLA Training via Reinforcement Learning (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper