name
stringlengths 3
64
| hf_repo
stringlengths 4
38
| hf_subset
stringlengths 1
55
⌀ | hf_avail_splits
list | evaluation_splits
list | generation_size
int64 -1
2.05k
| stop_sequence
list | metric
list | suite
list | prompt_function
stringlengths 3
50
|
---|---|---|---|---|---|---|---|---|---|
entity_matching_iTunes_Amazon
|
lighteval/EntityMatching
|
iTunes_Amazon
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Fodors_Zagats
|
lighteval/EntityMatching
|
Fodors_Zagats
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_DBLP_ACM
|
lighteval/EntityMatching
|
DBLP_ACM
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_DBLP_GoogleScholar
|
lighteval/EntityMatching
|
DBLP_GoogleScholar
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Amazon_Google
|
lighteval/EntityMatching
|
Amazon_Google
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Walmart_Amazon
|
lighteval/EntityMatching
|
Walmart_Amazon
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Abt_Buy
|
lighteval/EntityMatching
|
Abt_Buy
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Company
|
lighteval/EntityMatching
|
Company
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Dirty_iTunes_Amazon
|
lighteval/EntityMatching
|
Dirty_iTunes_Amazon
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Dirty_DBLP_ACM
|
lighteval/EntityMatching
|
Dirty_DBLP_ACM
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Dirty_DBLP_GoogleScholar
|
lighteval/EntityMatching
|
Dirty_DBLP_GoogleScholar
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
entity_matching_Dirty_Walmart_Amazon
|
lighteval/EntityMatching
|
Dirty_Walmart_Amazon
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
entity_matching
|
gsm8k
|
gsm8k
|
main
|
[
"train",
"test"
] |
[
"test"
] | 400 |
[
"\n\n"
] |
[
"exact_match_indicator",
"toxicity",
"bias"
] |
[
"helm"
] |
gsm8k_helm
|
imdb
|
lighteval/IMDB_helm
|
default
|
[
"train",
"test"
] |
[
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm"
] |
imdb
|
imdb_contrastset
|
lighteval/IMDB_helm
|
default
|
[
"test"
] |
[
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm"
] |
imdb_contrastset
|
interactive_qa_mmlu_college_chemistry
|
cais/mmlu
|
college_chemistry
|
[
"dev",
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"interactive_qa_mmlu_scenario"
] |
mmlu_qa_college_chemistry
|
interactive_qa_mmlu_global_facts
|
cais/mmlu
|
global_facts
|
[
"dev",
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"interactive_qa_mmlu_scenario"
] |
mmlu_qa_global_facts
|
interactive_qa_mmlu_miscellaneous
|
cais/mmlu
|
miscellaneous
|
[
"dev",
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"interactive_qa_mmlu_scenario"
] |
mmlu_qa_miscellaneous
|
interactive_qa_mmlu_nutrition
|
cais/mmlu
|
nutrition
|
[
"dev",
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"interactive_qa_mmlu_scenario"
] |
mmlu_qa_nutrition
|
interactive_qa_mmlu_us_foreign_policy
|
cais/mmlu
|
us_foreign_policy
|
[
"dev",
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"interactive_qa_mmlu_scenario"
] |
mmlu_qa_us_foreign_policy
|
legal_summarization_billsum
|
lighteval/legal_summarization
|
BillSum
|
[
"train",
"test"
] |
[
"test"
] | 1,024 |
[
"\n"
] |
[
"bias",
"toxicity",
"rouge_1",
"rouge_2",
"rouge_l",
"faithfulness",
"extractiveness",
"bert_score"
] |
[
"helm"
] |
legal_summarization
|
legal_summarization_eurlexsum
|
lighteval/legal_summarization
|
EurLexSum
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 2,048 |
[
"\n"
] |
[
"bias",
"toxicity",
"rouge_1",
"rouge_2",
"rouge_l",
"faithfulness",
"extractiveness",
"bert_score"
] |
[
"helm"
] |
legal_summarization
|
legal_summarization_multilexsum
|
lighteval/legal_summarization
|
MultiLexSum
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 256 |
[
"\n"
] |
[
"bias",
"toxicity",
"rouge_1",
"rouge_2",
"rouge_l",
"faithfulness",
"extractiveness",
"bert_score"
] |
[
"helm"
] |
multilexsum
|
legalsupport
|
lighteval/LegalSupport
|
default
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | -1 |
[
"\n"
] |
[
"loglikelihood_acc",
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm"
] |
legal_support
|
lexglue_ecthr_a
|
lighteval/lexglue
|
ecthr_a
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lex_glue_scenario"
] |
lex_glue_ecthr_a
|
lexglue_ecthr_b
|
lighteval/lexglue
|
ecthr_b
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lex_glue_scenario"
] |
lex_glue_ecthr_b
|
lexglue_scotus
|
lighteval/lexglue
|
scotus
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lex_glue_scenario"
] |
lex_glue_scotus
|
lexglue_eurlex
|
lighteval/lexglue
|
eurlex
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lex_glue_scenario"
] |
lex_glue_eurlex
|
lexglue_ledgar
|
lighteval/lexglue
|
ledgar
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lex_glue_scenario"
] |
lex_glue_ledgar
|
lexglue_unfair_tos
|
lighteval/lexglue
|
unfair_tos
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lex_glue_scenario"
] |
lex_glue_unfair_tos
|
lexglue_case_hold
|
lighteval/lexglue
|
case_hold
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lex_glue_scenario"
] |
lex_glue_case_hold
|
lextreme_brazilian_court_decisions_judgment
|
lighteval/lextreme
|
brazilian_court_decisions_judgment
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_brazilian_court_decisions_judgment
|
lextreme_brazilian_court_decisions_unanimity
|
lighteval/lextreme
|
brazilian_court_decisions_unanimity
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_brazilian_court_decisions_unanimity
|
lextreme_german_argument_mining
|
lighteval/lextreme
|
german_argument_mining
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_german_argument_mining
|
lextreme_greek_legal_code_chapter
|
lighteval/lextreme
|
greek_legal_code_chapter
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_greek_legal_code_chapter
|
lextreme_greek_legal_code_subject
|
lighteval/lextreme
|
greek_legal_code_subject
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_greek_legal_code_subject
|
lextreme_greek_legal_code_volume
|
lighteval/lextreme
|
greek_legal_code_volume
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_greek_legal_code_volume
|
lextreme_swiss_judgment_prediction
|
lighteval/lextreme
|
swiss_judgment_prediction
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_swiss_judgment_prediction
|
lextreme_online_terms_of_service_unfairness_levels
|
lighteval/lextreme
|
online_terms_of_service_unfairness_levels
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 10 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_online_terms_of_service_unfairness_levels
|
lextreme_online_terms_of_service_clause_topics
|
lighteval/lextreme
|
online_terms_of_service_clause_topics
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 10 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_online_terms_of_service_clause_topics
|
lextreme_covid19_emergency_event
|
lighteval/lextreme
|
covid19_emergency_event
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 10 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_covid19_emergency_event
|
lextreme_multi_eurlex_level_1
|
lighteval/lextreme
|
multi_eurlex_level_1
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 10 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_multi_eurlex_level_1
|
lextreme_multi_eurlex_level_2
|
lighteval/lextreme
|
multi_eurlex_level_2
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 10 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_multi_eurlex_level_2
|
lextreme_multi_eurlex_level_3
|
lighteval/lextreme
|
multi_eurlex_level_3
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 10 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_multi_eurlex_level_3
|
lextreme_greek_legal_ner
|
lighteval/lextreme
|
greek_legal_ner
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 430 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_greek_legal_ner
|
lextreme_legalnero
|
lighteval/lextreme
|
legalnero
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 788 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_legalnero
|
lextreme_lener_br
|
lighteval/lextreme
|
lener_br
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 338 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_lener_br
|
lextreme_mapa_coarse
|
lighteval/lextreme
|
mapa_coarse
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 274 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_mapa_coarse
|
lextreme_mapa_fine
|
lighteval/lextreme
|
mapa_fine
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 274 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score"
] |
[
"helm",
"lextreme_scenario"
] |
lextreme_mapa_fine
|
lsat_qa_grouping
|
lighteval/lsat_qa
|
grouping
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"lsat_qa_scenario"
] |
lsat_qa
|
lsat_qa_ordering
|
lighteval/lsat_qa
|
ordering
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"lsat_qa_scenario"
] |
lsat_qa
|
lsat_qa_assignment
|
lighteval/lsat_qa
|
assignment
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"lsat_qa_scenario"
] |
lsat_qa
|
lsat_qa_miscellaneous
|
lighteval/lsat_qa
|
miscellaneous
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"lsat_qa_scenario"
] |
lsat_qa
|
lsat_qa_all
|
lighteval/lsat_qa
|
all
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm",
"lsat_qa_scenario"
] |
lsat_qa
|
me_q_sum
|
lighteval/me_q_sum
|
default
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 128 |
[
"###"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score",
"rouge_l",
"bleu_1",
"bleu_4",
"toxicity",
"bias"
] |
[
"helm"
] |
me_q_sum
|
med_dialog_healthcaremagic
|
lighteval/med_dialog
|
healthcaremagic
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 128 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score",
"rouge_l",
"bleu_1",
"bleu_4",
"toxicity",
"bias"
] |
[
"helm"
] |
med_dialog
|
med_dialog_icliniq
|
lighteval/med_dialog
|
icliniq
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 128 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score",
"rouge_l",
"bleu_1",
"bleu_4",
"toxicity",
"bias"
] |
[
"helm"
] |
med_dialog
|
med_mcqa
|
lighteval/med_mcqa
|
default
|
[
"train",
"test",
"validation"
] |
[
"validation"
] | 5 |
[
"\n"
] |
[
"loglikelihood_acc",
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm"
] |
med_mcqa
|
med_paragraph_simplification
|
lighteval/med_paragraph_simplification
|
default
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 512 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score",
"rouge_l",
"bleu_1",
"bleu_4",
"toxicity",
"bias"
] |
[
"helm"
] |
med_paragraph_simplification
|
med_qa
|
bigbio/med_qa
|
med_qa_en_source
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"loglikelihood_acc",
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm"
] |
med_qa
|
mmlu
|
lighteval/mmlu
|
all
|
[
"auxiliary_train",
"test",
"validation",
"dev"
] |
[
"validation",
"test"
] | 5 |
[
"\n"
] |
[
"loglikelihood_acc",
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm"
] |
mmlu
|
ms_marco_regular
|
lighteval/ms_marco
|
regular
|
[
"train",
"validation"
] |
[
"validation"
] | 5 |
[
"\n"
] |
[
"ranking"
] |
[
"helm"
] |
ms_marco
|
ms_marco_trec
|
lighteval/ms_marco
|
trec
|
[
"train",
"validation"
] |
[
"validation"
] | 5 |
[
"\n"
] |
[
"ranking"
] |
[
"helm"
] |
ms_marco
|
narrativeqa
|
narrativeqa
|
default
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 100 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score",
"rouge_l",
"bleu_1",
"bleu_4",
"toxicity",
"bias"
] |
[
"helm"
] |
narrativeqa
|
numeracy_linear_example
|
lighteval/numeracy
|
linear_example
|
[
"train",
"test"
] |
[
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"absolute_value_difference"
] |
[
"helm"
] |
numeracy
|
numeracy_linear_standard
|
lighteval/numeracy
|
linear_standard
|
[
"train",
"test"
] |
[
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"absolute_value_difference"
] |
[
"helm"
] |
numeracy
|
numeracy_parabola_example
|
lighteval/numeracy
|
parabola_example
|
[
"train",
"test"
] |
[
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"absolute_value_difference"
] |
[
"helm"
] |
numeracy
|
numeracy_parabola_standard
|
lighteval/numeracy
|
parabola_standard
|
[
"train",
"test"
] |
[
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"absolute_value_difference"
] |
[
"helm"
] |
numeracy
|
numeracy_plane_example
|
lighteval/numeracy
|
plane_example
|
[
"train",
"test"
] |
[
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"absolute_value_difference"
] |
[
"helm"
] |
numeracy
|
numeracy_plane_standard
|
lighteval/numeracy
|
plane_standard
|
[
"train",
"test"
] |
[
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"absolute_value_difference"
] |
[
"helm"
] |
numeracy
|
numeracy_paraboloid_example
|
lighteval/numeracy
|
paraboloid_example
|
[
"train",
"test"
] |
[
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"absolute_value_difference"
] |
[
"helm"
] |
numeracy
|
numeracy_paraboloid_standard
|
lighteval/numeracy
|
paraboloid_standard
|
[
"train",
"test"
] |
[
"test"
] | 20 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"absolute_value_difference"
] |
[
"helm"
] |
numeracy
|
pubmed_qa
|
pubmed_qa
|
pqa_labeled
|
[
"train"
] |
[
"train"
] | 1 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match"
] |
[
"helm"
] |
pubmed_qa_helm
|
quac
|
quac
|
plain_text
|
[
"train",
"validation"
] |
[
"validation"
] | 100 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"f1_score",
"toxicity",
"bias"
] |
[
"helm"
] |
quac
|
raft_ade_corpus_v2
|
ought/raft
|
ade_corpus_v2
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_banking_77
|
ought/raft
|
banking_77
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_neurips_impact_statement_risks
|
ought/raft
|
neurips_impact_statement_risks
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_one_stop_english
|
ought/raft
|
one_stop_english
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_overruling
|
ought/raft
|
overruling
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_semiconductor_org_types
|
ought/raft
|
semiconductor_org_types
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_systematic_review_inclusion
|
ought/raft
|
systematic_review_inclusion
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_tai_safety_research
|
ought/raft
|
tai_safety_research
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_terms_of_service
|
ought/raft
|
terms_of_service
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_tweet_eval_hate
|
ought/raft
|
tweet_eval_hate
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
raft_twitter_complaints
|
ought/raft
|
twitter_complaints
|
[
"train",
"test"
] |
[
"test"
] | 30 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
raft
|
real_toxicity_prompts
|
allenai/real-toxicity-prompts
|
default
|
[
"train"
] |
[
"train"
] | 20 |
[
"\n"
] |
[
"bias",
"toxicity",
"prediction_perplexity"
] |
[
"helm"
] |
real_toxicity_prompts
|
summarization_xsum
|
lighteval/summarization
|
xsum
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 64 |
[
"\n"
] |
[
"bias",
"toxicity",
"rouge_1",
"rouge_2",
"rouge_l",
"faithfulness",
"extractiveness",
"bert_score"
] |
[
"helm"
] |
xsum
|
summarization_xsum-sampled
|
lighteval/summarization
|
xsum-sampled
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 64 |
[
"\n"
] |
[
"bias",
"toxicity",
"rouge_1",
"rouge_2",
"rouge_l",
"faithfulness",
"extractiveness",
"bert_score"
] |
[
"helm"
] |
xsum
|
summarization_cnn-dm
|
lighteval/summarization
|
cnn-dm
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 128 |
[
"\n"
] |
[
"bias",
"toxicity",
"rouge_1",
"rouge_2",
"rouge_l",
"faithfulness",
"extractiveness",
"bert_score"
] |
[
"helm"
] |
cnn_dm
|
synthetic_reasoning_natural_easy
|
lighteval/synthetic_reasoning_natural
|
easy
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"f1_set_match",
"iou_set_match",
"exact_set_match",
"toxicity",
"bias"
] |
[
"helm"
] |
synthetic_reasoning_natural
|
synthetic_reasoning_natural_hard
|
lighteval/synthetic_reasoning_natural
|
hard
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 20 |
[
"\n"
] |
[
"f1_set_match",
"iou_set_match",
"exact_set_match",
"toxicity",
"bias"
] |
[
"helm"
] |
synthetic_reasoning_natural
|
synthetic_reasoning_variable_substitution
|
lighteval/synthetic_reasoning
|
variable_substitution
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 50 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
synthetic_reasoning
|
synthetic_reasoning_pattern_match
|
lighteval/synthetic_reasoning
|
pattern_match
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 50 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
synthetic_reasoning
|
synthetic_reasoning_induction
|
lighteval/synthetic_reasoning
|
induction
|
[
"train",
"test",
"validation"
] |
[
"validation",
"test"
] | 50 |
[
"\n"
] |
[
"exact_match",
"quasi_exact_match",
"prefix_exact_match",
"quasi_prefix_exact_match",
"toxicity",
"bias"
] |
[
"helm"
] |
synthetic_reasoning
|
the_pile_github
|
lighteval/pile_helm
|
github
|
[
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"perplexity"
] |
[
"helm"
] |
the_pile
|
the_pile_arxiv
|
lighteval/pile_helm
|
arxiv
|
[
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"perplexity"
] |
[
"helm"
] |
the_pile
|
the_pile_wikipedia
|
lighteval/pile_helm
|
wikipedia
|
[
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"perplexity"
] |
[
"helm"
] |
the_pile
|
the_pile_opensubtitles
|
lighteval/pile_helm
|
opensubtitles
|
[
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"perplexity"
] |
[
"helm"
] |
the_pile
|
the_pile_openwebtext2
|
lighteval/pile_helm
|
openwebtext2
|
[
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"perplexity"
] |
[
"helm"
] |
the_pile
|
the_pile_gutenberg
|
lighteval/pile_helm
|
gutenberg
|
[
"test"
] |
[
"test"
] | -1 |
[
"\n"
] |
[
"perplexity"
] |
[
"helm"
] |
the_pile
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.