|
--- |
|
language: |
|
- fr |
|
- es |
|
- zh |
|
size_categories: |
|
- 1K<n<10K |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: test |
|
path: "test/data-00000-of-00001.arrow" |
|
--- |
|
|
|
# MultiNRC: Multilingual Native Reasoning Challenge |
|
|
|
MultiNRC is a challenging evaluation benchmark for large language models, designed to assess multilingual reasoning ability in French, Spanish, and Chinese. Unlike existing benchmarks that simply translate English-centric content, MultiNRC consists of over 1,000 native-authored reasoning questions, crafted by native speakers to capture linguistic and cultural nuances. |
|
|
|
## Features |
|
|
|
- **Languages:** French, Spanish, Chinese |
|
- **Categories:** |
|
- Language-specific Linguistic Reasoning |
|
- Wordplay & Riddles |
|
- Cultural Reasoning & Traditions |
|
- Math Reasoning with Cultural Relevance |
|
- **English Equivalents:** For Cultural/Tradition and Math, human-translated English versions are provided for direct comparison. |
|
- **Ground Truth Final Answers:** Short, objective answers accompany each prompt for automatic evaluation. |
|
|
|
|
|
## Dataset Structure |
|
|
|
Each entry includes: |
|
- A native-language prompt and answer (`i18n_prompt`, `i18n_gtfa`) |
|
- (For Math Reasoning and Cultural Reasoning category tasks) An English-equivalent prompt and answer (`english_prompt`, `english_gtfa`) |
|
- Metadata: `task_id`, `language`, `category` |
|
|
|
## Citation |
|
|
|
If you use MultiNRC in your research, please cite: |
|
|
|
```bibtex |
|
@article{fabbri2025multinrc, |
|
title = {MultiNRC: A Challenging Native Multilingual Reasoning Evaluation Benchmark for LLMs}, |
|
author = {Fabbri, Alexander R. and Mares, Diego and Flores, Jorge and Mankikar, Meher and Hernandez, Ernesto and Lee, Dean and Liu, Bing and Xing, Chen}, |
|
year = {2025}, |
|
note = {arXiv preprint, arXiv:XXXX.XXXXX} |
|
} |