File size: 4,509 Bytes
2b77c62
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5de3cd2
 
 
 
cadeeca
5de3cd2
cadeeca
 
 
5de3cd2
dc4cdb1
5de3cd2
cadeeca
dc4cdb1
5de3cd2
dc4cdb1
5de3cd2
 
 
 
cadeeca
5de3cd2
cadeeca
 
5de3cd2
 
 
 
dc4cdb1
394d9c0
 
6951efe
 
 
1b7f7e7
 
394d9c0
 
 
5de3cd2
 
 
aeebea2
 
cadeeca
dc4cdb1
5de3cd2
cadeeca
dc4cdb1
5de3cd2
cadeeca
 
 
394d9c0
1b7f7e7
394d9c0
9305e2f
 
 
 
 
 
 
1b7f7e7
 
5de3cd2
cadeeca
dfed77b
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
---
license: mit
language:
- en
tags:
- behavior
- motion
- human
- egocentric
- language
- llm
- vlm
- esk
size_categories:
- 10K<n<100K
task_categories:
- question-answering
---

# πŸ‹ EPFL-Smart-Kitchen: Lemonade benchmark

![title](media/title.svg)

## πŸ“š Introduction
 we introduce Lemonade: **L**anguage models **E**valuation of **MO**tion a**N**d **A**ction-**D**riven **E**nquiries.
Lemonade consists of <span style="color: orange;">36,521</span> closed-ended QA pairs linked to egocentric video clips, categorized in three groups and six subcategories. <span style="color: orange;">18,857</span> QAs focus on behavior understanding, leveraging the rich ground truth behavior annotations of the EPFL-Smart Kitchen to interrogate models about perceived actions <span style="color: tomato;">(Perception)</span> and reason over unseen behaviors <span style="color: tomato;">(Reasoning)</span>. <span style="color: orange;">8,210</span> QAs involve longer video clips, challenging models in summarization <span style="color: gold;">(Summarization)</span> and session-level inference <span style="color: gold;">(Session properties)</span>. The remaining <span style="color: orange;">9,463</span> QAs leverage the 3D pose estimation data to infer hand shapes, joint angles <span style="color: skyblue;">(Physical attributes)</span>, or trajectory velocities <span style="color: skyblue;">(Kinematics)</span> from visual information.

## πŸ’Ύ Content
The current repository contains all egocentric videos recorded in the EPFL-Smart-Kitchen-30 dataset and the question answer pairs of the Lemonade benchmark. Please refer to the [main GitHub repository](https://github.com/amathislab/EPFL-Smart-Kitchen#) to find the other benchmarks and links to download other modalities of the EPFL-Smart-Kitchen-30 dataset.

### πŸ—ƒοΈ Repository structure

```
Lemonade
β”œβ”€β”€ MCQs
|   └── lemonade_benchmark.csv
β”œβ”€β”€ videos
|   β”œβ”€β”€ YH2002_2023_12_04_10_15_23_hololens.mp4
|   └── ..
└── README.md
```

`lemonade_benchmark.csv` : Table with the following fields:

**Question**  : Question to be answered. </br>
**QID** : Question identifier, an integer from 0 to 30. </br>
**Answers** : A list of possible answers to the question. This can be a multiple-choice set or open-ended responses. </br>
**Correct Answer** : The answer that is deemed correct from the list of provided answers. </br>
**Clip** : A reference to the video clip related to the question. </br>
**Start** : The timestamp (in frames) in the clip where the question context begins. </br>
**End** : The timestamp (in frames) in the clip where the question context ends. </br>
**Category** : The broad topic under which the question falls (Behavior understanding, Long-term understanding or Motion and Biomechanics). </br>
**Subcategory** : A more refined classification within the category (Perception, Reasoning, Summarization, Session properties, Physical attributes, Kinematics). </br>
**Difficulty** : The complexity level of the question (e.g., Easy, Medium, Hard).

`videos` : Folder with all egocentric videos from the EPFL-Smart-Kitchen-30 benchmark. Video names are structured as `[Participant_ID]_[Session_name]_hololens.mp4`.

> We refer the reader to the associated publication for details about data processing and tasks description.

## πŸ“ˆ Evaluation results
![evaluation_results](media/evaluation_results.svg)

## 🌈 Usage
The evaluation of the benchmark can be done through the following github repository: [https://github.com/amathislab/lmms-eval-lemonade](https://github.com/amathislab/lmms-eval-lemonade)



## 🌟 Citations
Please cite our work!
```
@misc{bonnetto2025epflsmartkitchen,
      title={EPFL-Smart-Kitchen-30: Densely annotated cooking dataset with 3D kinematics to challenge video and language models}, 
      author={Andy Bonnetto and Haozhe Qi and Franklin Leong and Matea Tashkovska and Mahdi Rad and Solaiman Shokur and Friedhelm Hummel and Silvestro Micera and Marc Pollefeys and Alexander Mathis},
      year={2025},
      eprint={2506.01608},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2506.01608}, 
}
```

## ❀️ Acknowledgments

Our work was funded by EPFL, Swiss SNF grant (320030-227871), Microsoft Swiss Joint Research Center and a Boehringer Ingelheim Fonds PhD stipend (H.Q.). We are grateful to the Brain Mind Institute for providing funds for hardware and to the Neuro-X Institute for providing funds for services.