Ryan McConville
Add submission page
07c76e7
|
raw
history blame
2.98 kB
# LLM Hallucination Detection Leaderboard Submission Guidelines
Thank you for your interest in contributing to the **LLM Hallucination Detection Leaderboard**! We welcome submissions from researchers and practitioners who have built or finetuned language models that can be evaluated on our hallucination benchmarks.
---
## 1. What to Send
Please email **ryan@kluster.ai** with the subject line:
```
[Verify Leaderboard Submission] <Your-Model-Name>
```
Attach **one ZIP file** that contains **all of the following**:
1. **`model_card.md`** – A short Markdown file describing your model:
β€’ Name and version
β€’ Architecture / base model
β€’ Training or finetuning procedure
β€’ License
β€’ Intended use & known limitations
β€’ Contact information
2. **`results.csv`** – A CSV file with **one row per prompt** and **one column per field** (see schema below).
3. (Optional) **`extra_notes.md`** – Anything else you would like us to know (e.g., additional analysis).
---
## 2. CSV Schema
| Column | Description |
|--------------------|---------------------------------------------------------------------------|
| `request` | The exact input prompt shown to the model. |
| `response` | The raw output produced by the model. |
| `verify_response` | The Verify judgment or explanation regarding hallucination. |
| `verify_label` | The final boolean / categorical label (e.g., `TRUE`, `FALSE`). |
| `task` | The benchmark or dataset name the sample comes from. |
**Important:** Use UTF-8 encoding and **do not** add additional columns without prior discussion; extra information should go in the `metadata` field. You must use Verify by kluster.ai to ensure fairness in the leaderboard.
---
## 3. Evaluation Datasets
Run your model on the following public datasets and include *all* examples in your CSV. You can load them directly from Hugging Face:
| Dataset | Hugging Face Link |
|---------|-------------------|
| HaluEval QA (qa_samples subet with Question and Knowledge column) | https://huggingface.co/datasets/pminervini/HaluEval |
| UltraChat | https://huggingface.co/datasets/kluster-ai/ultrachat-sampled |
---
## 5. Example Row
```csv
request,response,verify_response,verify_label,task
"What is the capital of the UK?","London is the capital of the UK.","The statement is factually correct.",CORRECT,TruthfulQA
```
---
## 6. Review Process
1. We will sanity-check the file format and reproduce a random subset.
2. If everything looks good, your scores will appear on the public leaderboard.
3. We may reach out for clarifications, please keep an eye on your inbox.
---
## 7. Contact
Questions? Email **ryan@kluster.ai**.
We look forward to your submissions and to advancing reliable language models together!