Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
video
video
5.97
140
label
class label
500 classes
0bc_z-BC_Z-11233
1bc_z-BC_Z-1163
2bc_z-BC_Z-13018
3bc_z-BC_Z-13472
4bc_z-BC_Z-13955
5bc_z-BC_Z-14160
6bc_z-BC_Z-15947
7bc_z-BC_Z-16448
8bc_z-BC_Z-1686
9bc_z-BC_Z-17602
10bc_z-BC_Z-17631
11bc_z-BC_Z-1767
12bc_z-BC_Z-17787
13bc_z-BC_Z-181
14bc_z-BC_Z-18255
15bc_z-BC_Z-18893
16bc_z-BC_Z-19007
17bc_z-BC_Z-19515
18bc_z-BC_Z-21027
19bc_z-BC_Z-21350
20bc_z-BC_Z-21538
21bc_z-BC_Z-2279
22bc_z-BC_Z-23559
23bc_z-BC_Z-24544
24bc_z-BC_Z-24851
25bc_z-BC_Z-25101
26bc_z-BC_Z-25744
27bc_z-BC_Z-26655
28bc_z-BC_Z-27184
29bc_z-BC_Z-27568
30bc_z-BC_Z-27776
31bc_z-BC_Z-27803
32bc_z-BC_Z-28295
33bc_z-BC_Z-29674
34bc_z-BC_Z-31122
35bc_z-BC_Z-31301
36bc_z-BC_Z-32707
37bc_z-BC_Z-33638
38bc_z-BC_Z-34392
39bc_z-BC_Z-35454
40bc_z-BC_Z-37597
41bc_z-BC_Z-38929
42bc_z-BC_Z-39189
43bc_z-BC_Z-5379
44bc_z-BC_Z-5456
45bc_z-BC_Z-5818
46bc_z-BC_Z-7759
47bc_z-BC_Z-9301
48bc_z-BC_Z-9322
49bc_z-BC_Z-9987
50bridge-10094
51bridge-11429
52bridge-11704
53bridge-13687
54bridge-14215
55bridge-1515
56bridge-15227
57bridge-16649
58bridge-1677
59bridge-19530
60bridge-21072
61bridge-22519
62bridge-23093
63bridge-23531
64bridge-25135
65bridge-26664
66bridge-26665
67bridge-27093
68bridge-28156
69bridge-28341
70bridge-29302
71bridge-31043
72bridge-31400
73bridge-31544
74bridge-31657
75bridge-34337
76bridge-34957
77bridge-35046
78bridge-37131
79bridge-37413
80bridge-39198
81bridge-39890
82bridge-41233
83bridge-41259
84bridge-41829
85bridge-42806
86bridge-4284
87bridge-42926
88bridge-43883
89bridge-44471
90bridge-44700
91bridge-4509
92bridge-4598
93bridge-47711
94bridge-49872
95bridge-52430
96bridge-5965
97bridge-6992
98bridge-7114
99bridge-8886
End of preview. Expand in Data Studio

RoboFine-Bench

RoboFine-Bench is a benchmark for evaluating Vision-Language Models on fine-grained robotic manipulation understanding. It tests whether VLMs can comprehend execution-level details in robot videos, including contact regions, approach directions, motion trajectories, gripper states, and temporal ordering.

Dataset Structure

RoboFine-Bench/
├── videos/                    # Robot manipulation videos (mp4)
│   ├── BC-Z/
│   ├── BridgeDataV2/
│   ├── DROID-Robointer/
│   ├── Galaxea/
│   ├── RDT/
│   ├── RH20T-RoboInter/
│   ├── RT-1/
│   ├── RoboCoin/
│   ├── RoboMINDV1/
│   └── RoboMINDV2/
├── EvalSets.json              # Caption evaluation set (GT annotations)
├── QAEvalSets.json            # VQA evaluation set (questions + answers)
└── GT_AtomicFacts.jsonl       # Atomic facts for caption scoring

Evaluation Metrics

RoboFine-Bench evaluates VLMs through two complementary metrics:

1. VQA Evaluation

Multiple-choice questions targeting 13 fine-grained capability dimensions. Models receive video frames and answer questions about manipulation details.

2. Caption Evaluation (Atomic Fact Alignment)

Models generate dense captions for robot videos. Captions are decomposed into atomic facts and aligned against human-annotated ground truth across 13 dimensions:

  • action_primitive, actor_identity, object_recognition, object_disambiguation
  • contact_region, source_state_or_location, trajectory_and_orientation
  • placement_specification, interaction_with_other_objects
  • success_failure_retry, gripper_state, temporal_order_and_step_boundary, body_motion

Scoring measures: coverage (GT facts recalled), consistency (no contradictions), anti-hallucination (no invented facts).

Data Sources

Videos are sourced from 10 public robot manipulation datasets spanning diverse embodiments, environments, and task complexities.

Usage

from datasets import load_dataset

dataset = load_dataset("FineVLA/RoboFine-Bench")

For evaluation code, see: github.com/EricsXt/FineVLA

Citation

@article{hu2026finevla,
  title={FineVLA: Fine-Grained Instruction Alignment for Steerable Vision-Language-Action Policies},
  author={Hu, Xintong and Huang, Xuhong and Zhang, Jinyu and Yao, Yutong and Sun, Yuchong and Wang, Qiuyue and Li, Mingsheng and Liu, Yitao and Chen, Yixuan and Zheng, Yingming and Bai, Shuai and Yu, Tao},
  year={2026}
}
Downloads last month
1,021