Update leaderboard_data.jsonl
Browse files- leaderboard_data.jsonl +27 -27
leaderboard_data.jsonl
CHANGED
@@ -1,27 +1,27 @@
|
|
1 |
-
{"Model": "meta-llama/Llama-3.2-1B-Instruct", "Precision": "bfloat16", "#Params (B)": 1.24, "Part Multiple Choice": 28.28, "ARC Easy": 47.1, "ARC Challenge": 39.0, "MMLU Pro": 12.17, "AUT Multiple Choice Persian": 36.88, "Hub License": "llama3.2", "Model sha": "9213176726f574b556790deb65791e0c5aa438b6", "model_name_for_query": "meta-llama/Llama-3.2-1B-Instruct"}
|
2 |
-
{"Model": "PartAI/Dorna2-Llama3.1-8B-Instruct", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 34.48, "ARC Easy": 79.59, "ARC Challenge": 64.42, "MMLU Pro": 21.47, "AUT Multiple Choice Persian": 53.64, "Hub License": "llama3.1", "Model sha": "b78e4bd261100c96e511ed5090ca0ce0e1f4b340", "model_name_for_query": "PartAI/Dorna2-Llama3.1-8B-Instruct"}
|
3 |
-
{"Model": "Qwen/Qwen2.5-7B-Instruct", "Precision": "bfloat16", "#Params (B)": 7.62, "Part Multiple Choice": 36.72, "ARC Easy": 79.02, "ARC Challenge": 69.13, "MMLU Pro": 21.96, "AUT Multiple Choice Persian": 52.66, "Hub License": "apache-2.0", "Model sha": "a09a35458c702b33eeacc393d103063234e8bc28", "model_name_for_query": "Qwen/Qwen2.5-7B-Instruct"}
|
4 |
-
{"Model": "CohereForAI/c4ai-command-r7b-12-2024", "Precision": "float16", "#Params (B)": 8.03, "Part Multiple Choice": 34.03, "ARC Easy": 77.01, "ARC Challenge": 66.44, "MMLU Pro": 23.62, "AUT Multiple Choice Persian": 50.49, "Hub License": "cc-by-nc-4.0", "Model sha": "ff3e3c9c990d8d7576a4f8fa839281e11ebabc09", "model_name_for_query": "CohereForAI/c4ai-command-r7b-12-2024"}
|
5 |
-
{"Model": "google/gemma-3-4b-it", "Precision": "bfloat16", "#Params (B)": 4.3, "Part Multiple Choice": 34.4, "ARC Easy": 77.01, "ARC Challenge": 63.76, "MMLU Pro": 19.81, "AUT Multiple Choice Persian": 50.3, "Hub License": "gemma", "Model sha": "093f9f388b31de276ce2de164bdc2081324b9767", "model_name_for_query": "google/gemma-3-4b-it"}
|
6 |
-
{"Model": "Qwen/Qwen2-7B-Instruct", "Precision": "bfloat16", "#Params (B)": 7.62, "Part Multiple Choice": 35.9, "ARC Easy": 77.3, "ARC Challenge": 68.46, "MMLU Pro": 23.87, "AUT Multiple Choice Persian": 51.68, "Hub License": "apache-2.0", "Model sha": "f2826a00ceef68f0f2b946d945ecc0477ce4450c", "model_name_for_query": "Qwen/Qwen2-7B-Instruct"}
|
7 |
-
{"Model": "meta-llama/Meta-Llama-3-8B-Instruct", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 34.99, "ARC Easy": 72.9, "ARC Challenge": 57.7, "MMLU Pro": 25.54, "AUT Multiple Choice Persian": 53.85, "Hub License": "llama3", "Model sha": "5f0b02c75b57c5855da9ae460ce51323ea669d8a", "model_name_for_query": "meta-llama/Meta-Llama-3-8B-Instruct"}
|
8 |
-
{"Model": "NousResearch/Hermes-3-Llama-3.1-8B", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 35.01, "ARC Easy": 77.01, "ARC Challenge": 58.39, "MMLU Pro": 21.0, "AUT Multiple Choice Persian": 52.46, "Hub License": "llama3", "Model sha": "896ea440e5a9e6070e3d8a2774daf2b481ab425b", "model_name_for_query": "NousResearch/Hermes-3-Llama-3.1-8B"}
|
9 |
-
{"Model": "CohereForAI/aya-expanse-32b", "Precision": "float16", "#Params (B)": 32.3, "Part Multiple Choice": 43.36, "ARC Easy": 93.1, "ARC Challenge": 79.87, "MMLU Pro": 31.03, "AUT Multiple Choice Persian": 62.33, "Hub License": "cc-by-nc-4.0", "Model sha": "94bda1dcb97d260f732d230b832c7c685ae91e23", "model_name_for_query": "CohereForAI/aya-expanse-32b"}
|
10 |
-
{"Model": "CohereForAI/aya-23-8B", "Precision": "float16", "#Params (B)": 8.03, "Part Multiple Choice": 32.82, "ARC Easy": 80.46, "ARC Challenge": 64.43, "MMLU Pro": 18.62, "AUT Multiple Choice Persian": 52.86, "Hub License": "cc-by-nc-4.0", "Model sha": "2a1a63b24af8f591616fdf58936ee576d63ca835", "model_name_for_query": "CohereForAI/aya-23-8B"}
|
11 |
-
{"Model": "meta-llama/Llama-3.1-8B-Instruct", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 36.68, "ARC Easy": 78.4, "ARC Challenge": 60.4, "MMLU Pro": 21.0, "AUT Multiple Choice Persian": 54.24, "Hub License": "llama3.1", "Model sha": "0e9e39f249a16976918f6564b8830bc894c89659", "model_name_for_query": "meta-llama/Llama-3.1-8B-Instruct"}
|
12 |
-
{"Model": "google/gemma-2-2b-it", "Precision": "bfloat16", "#Params (B)": 2.61, "Part Multiple Choice": 31.12, "ARC Easy": 71.26, "ARC Challenge": 57.72, "MMLU Pro": 16.23, "AUT Multiple Choice Persian": 49.9, "Hub License": "gemma", "Model sha": "299a8560bedf22ed1c72a8a11e7dce4a7f9f51f8", "model_name_for_query": "google/gemma-2-2b-it"}
|
13 |
-
{"Model": "meta-llama/Meta-Llama-3-8B", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 34.32, "ARC Easy": 72.12, "ARC Challenge": 56.37, "MMLU Pro": 19.33, "AUT Multiple Choice Persian": 48.32, "Hub License": "llama3", "Model sha": "8cde5ca8380496c9a6cc7ef3a8b46a0372a1d920", "model_name_for_query": "meta-llama/Meta-Llama-3-8B"}
|
14 |
-
{"Model": "google/gemma-2-9b-it", "Precision": "bfloat16", "#Params (B)": 9.24, "Part Multiple Choice": 42.7, "ARC Easy": 93.1, "ARC Challenge": 84.56, "MMLU Pro": 31.74, "AUT Multiple Choice Persian": 62.33, "Hub License": "gemma", "Model sha": "11c9b309abf73637e4b6f9a3fa1e92e615547819", "model_name_for_query": "google/gemma-2-9b-it"}
|
15 |
-
{"Model": "PartAI/Dorna-Llama3-8B-Instruct", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 33.88, "ARC Easy": 70.4, "ARC Challenge": 61.07, "MMLU Pro": 23.39, "AUT Multiple Choice Persian": 52.86, "Hub License": "llama3", "Model sha": "fb268bb51b950b4db5b7c82c1b73d9e803020eed", "model_name_for_query": "PartAI/Dorna-Llama3-8B-Instruct"}
|
16 |
-
{"Model": "MaralGPT/Maral-7B-alpha-1", "Precision": "bfloat16", "#Params (B)": 7.24, "Part Multiple Choice": 26.67, "ARC Easy": 44.54, "ARC Challenge": 30.87, "MMLU Pro": 15.99, "AUT Multiple Choice Persian": 36.09, "Hub License": "mit", "Model sha": "2ab5ca2a0d1a4454a78b4ca911e595bb9da2fe2f", "model_name_for_query": "MaralGPT/Maral-7B-alpha-1"}
|
17 |
-
{"Model": "Qwen/QwQ-32B-Preview", "Precision": "bfloat16", "#Params (B)": 32.8, "Part Multiple Choice": 46.64, "ARC Easy": 91.95, "ARC Challenge": 87.24, "MMLU Pro": 37.94, "AUT Multiple Choice Persian": 60.15, "Hub License": "apache-2.0", "Model sha": "91906fe41a48b6a89ce2970abfd1269eefee170e", "model_name_for_query": "Qwen/QwQ-32B-Preview"}
|
18 |
-
{"Model": "Qwen/QwQ-32B", "Precision": "bfloat16", "#Params (B)": 32.8, "Part Multiple Choice": 46.48, "ARC Easy": 89.37, "ARC Challenge": 82.55, "MMLU Pro": 35.32, "AUT Multiple Choice Persian": 57.4, "Hub License": "apache-2.0", "Model sha": "976055f8c83f394f35dbd3ab09a285a984907bd0", "model_name_for_query": "Qwen/QwQ-32B"}
|
19 |
-
{"Model": "CohereForAI/aya-23-35B", "Precision": "float16", "#Params (B)": 35, "Part Multiple Choice": 36.79, "ARC Easy": 87.93, "ARC Challenge": 72.48, "MMLU Pro": 25.54, "AUT Multiple Choice Persian": 61.14, "Hub License": "cc-by-nc-4.0", "Model sha": "5e72bd5ad83e5e1612ee7f56a0c1a439a7cfb887", "model_name_for_query": "CohereForAI/aya-23-35B"}
|
20 |
-
{"Model": "google/gemma-3-1b-it", "Precision": "bfloat16", "#Params (B)": 1, "Part Multiple Choice": 27.15, "ARC Easy": 41.67, "ARC Challenge": 31.54, "MMLU Pro": 16.23, "AUT Multiple Choice Persian": 36.88, "Hub License": "gemma", "Model sha": "dcc83ea841ab6100d6b47a070329e1ba4cf78752", "model_name_for_query": "google/gemma-3-1b-it"}
|
21 |
-
{"Model": "CohereForAI/aya-expanse-8b", "Precision": "float16", "#Params (B)": 8.03, "Part Multiple Choice": 34.91, "ARC Easy": 79.6, "ARC Challenge": 70.47, "MMLU Pro": 25.06, "AUT Multiple Choice Persian": 58.38, "Hub License": "cc-by-nc-4.0", "Model sha": "0ad43ec1e309e1351faa4b1d22713c065e37359a", "model_name_for_query": "CohereForAI/aya-expanse-8b"}
|
22 |
-
{"Model": "google/gemma-2-27b-it", "Precision": "bfloat16", "#Params (B)": 27.2, "Part Multiple Choice": 46.03, "ARC Easy": 95.98, "ARC Challenge": 85.91, "MMLU Pro": 36.28, "AUT Multiple Choice Persian": 63.12, "Hub License": "gemma", "Model sha": "aaf20e6b9f4c0fcf043f6fb2a2068419086d77b0", "model_name_for_query": "google/gemma-2-27b-it"}
|
23 |
-
{"Model": "Qwen/Qwen2.5-32B-Instruct", "Precision": "bfloat16", "#Params (B)": 32.8, "Part Multiple Choice": 46.06, "ARC Easy": 90.8, "ARC Challenge": 85.91, "MMLU Pro": 38.19, "AUT Multiple Choice Persian": 61.34, "Hub License": "apache-2.0", "Model sha": "5ede1c97bbab6ce5cda5812749b4c0bdf79b18dd", "model_name_for_query": "Qwen/Qwen2.5-32B-Instruct"}
|
24 |
-
{"Model": "universitytehran/PersianMind-v1.0", "Precision": "bfloat16", "#Params (B)": 6.82, "Part Multiple Choice": 29.27, "ARC Easy": 58.91, "ARC Challenge": 48.32, "MMLU Pro": 15.51, "AUT Multiple Choice Persian": 45.36, "Hub License": "cc-by-nc-sa-4.0", "Model sha": "af603eeb074138e2a613fbc95d89f018afbd3041", "model_name_for_query": "universitytehran/PersianMind-v1.0"}
|
25 |
-
{"Model": "google/gemma-3-27b-it", "Precision": "bfloat16", "#Params (B)": 27.4, "Part Multiple Choice": 48.56, "ARC Easy": 95.69, "ARC Challenge": 90.6, "MMLU Pro": 40.1, "AUT Multiple Choice Persian": 64.3, "Hub License": "gemma", "Model sha": "005ad3404e59d6023443cb575daa05336842228a", "model_name_for_query": "google/gemma-3-27b-it"}
|
26 |
-
{"Model": "meta-llama/Llama-3.1-8B", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 35.52, "ARC Easy": 75.28, "ARC Challenge": 53.02, "MMLU Pro": 24.1, "AUT Multiple Choice Persian": 53.45, "Hub License": "llama3.1", "Model sha": "d04e592bb4f6aa9cfee91e2e20afa771667e1d4b", "model_name_for_query": "meta-llama/Llama-3.1-8B"}
|
27 |
-
{"Model": "google/gemma-3-12b-it", "Precision": "bfloat16", "#Params (B)": 12.2, "Part Multiple Choice": 43.52, "ARC Easy": 93.39, "ARC Challenge": 81.21, "MMLU Pro": 29.36, "AUT Multiple Choice Persian": 57.0, "Hub License": "gemma", "Model sha": "96b6f1eccf38110c56df3a15bffe176da04bfd80", "model_name_for_query": "google/gemma-3-12b-it"}
|
|
|
1 |
+
{"Model": "meta-llama/Llama-3.2-1B-Instruct", "Precision": "bfloat16", "#Params (B)": 1.24, "Part Multiple Choice": 28.28, "ARC Easy": 47.1, "ARC Challenge": 39.0, "MMLU Pro": 12.17, "AUT Multiple Choice Persian": 36.88, "Hub License": "llama3.2", "Model sha": "9213176726f574b556790deb65791e0c5aa438b6", "model_name_for_query": "meta-llama/Llama-3.2-1B-Instruct", "General Knowledge": "29.59%", "GSM8K": "4.10%", "Homegraph Easy": "50.93%", "Homograph Hard": "52.53%", "PIQA": "53.79%", "Proverb": "28.65%", "Verb Tense": "26.11%", "Winogrande": "49.84%"}
|
2 |
+
{"Model": "PartAI/Dorna2-Llama3.1-8B-Instruct", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 34.48, "ARC Easy": 79.59, "ARC Challenge": 64.42, "MMLU Pro": 21.47, "AUT Multiple Choice Persian": 53.64, "Hub License": "llama3.1", "Model sha": "b78e4bd261100c96e511ed5090ca0ce0e1f4b340", "model_name_for_query": "PartAI/Dorna2-Llama3.1-8B-Instruct", "General Knowledge": "48.72%", "GSM8K": "11.90%", "Homegraph Easy": "44.44%", "Homograph Hard": "72.81%", "PIQA": "69.07%", "Proverb": "42.97%", "Verb Tense": "42.06%", "Winogrande": "54.19%"}
|
3 |
+
{"Model": "Qwen/Qwen2.5-7B-Instruct", "Precision": "bfloat16", "#Params (B)": 7.62, "Part Multiple Choice": 36.72, "ARC Easy": 79.02, "ARC Challenge": 69.13, "MMLU Pro": 21.96, "AUT Multiple Choice Persian": 52.66, "Hub License": "apache-2.0", "Model sha": "a09a35458c702b33eeacc393d103063234e8bc28", "model_name_for_query": "Qwen/Qwen2.5-7B-Instruct", "General Knowledge": "51.02%", "GSM8K": "18.00%", "Homegraph Easy": "52.78%", "Homograph Hard": "79.26%", "PIQA": "69.74%", "Proverb": "47.84%", "Verb Tense": "44.44%", "Winogrande": "58.79%"}
|
4 |
+
{"Model": "CohereForAI/c4ai-command-r7b-12-2024", "Precision": "float16", "#Params (B)": 8.03, "Part Multiple Choice": 34.03, "ARC Easy": 77.01, "ARC Challenge": 66.44, "MMLU Pro": 23.62, "AUT Multiple Choice Persian": 50.49, "Hub License": "cc-by-nc-4.0", "Model sha": "ff3e3c9c990d8d7576a4f8fa839281e11ebabc09", "model_name_for_query": "CohereForAI/c4ai-command-r7b-12-2024", "General Knowledge": "52.55%", "GSM8K": "10.10%", "Homegraph Easy": "47.22%", "Homograph Hard": "71.43%", "PIQA": "73.68%", "Proverb": "56.76%", "Verb Tense": "42.45%", "Winogrande": "55.56%"}
|
5 |
+
{"Model": "google/gemma-3-4b-it", "Precision": "bfloat16", "#Params (B)": 4.3, "Part Multiple Choice": 34.4, "ARC Easy": 77.01, "ARC Challenge": 63.76, "MMLU Pro": 19.81, "AUT Multiple Choice Persian": 50.3, "Hub License": "gemma", "Model sha": "093f9f388b31de276ce2de164bdc2081324b9767", "model_name_for_query": "google/gemma-3-4b-it", "General Knowledge": "45.92%", "GSM8K": "9.60%", "Homegraph Easy": "42.59%", "Homograph Hard": "72.58%", "PIQA": "70.80%", "Proverb": "53.78%", "Verb Tense": "45.30%", "Winogrande": "53.63%"}
|
6 |
+
{"Model": "Qwen/Qwen2-7B-Instruct", "Precision": "bfloat16", "#Params (B)": 7.62, "Part Multiple Choice": 35.9, "ARC Easy": 77.3, "ARC Challenge": 68.46, "MMLU Pro": 23.87, "AUT Multiple Choice Persian": 51.68, "Hub License": "apache-2.0", "Model sha": "f2826a00ceef68f0f2b946d945ecc0477ce4450c", "model_name_for_query": "Qwen/Qwen2-7B-Instruct", "General Knowledge": "52.04%", "GSM8K": "14.50%", "Homegraph Easy": "54.63%", "Homograph Hard": "72.81%", "PIQA": "68.88%", "Proverb": "50.54%", "Verb Tense": "40.62%", "Winogrande": "57.74%"}
|
7 |
+
{"Model": "meta-llama/Meta-Llama-3-8B-Instruct", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 34.99, "ARC Easy": 72.9, "ARC Challenge": 57.7, "MMLU Pro": 25.54, "AUT Multiple Choice Persian": 53.85, "Hub License": "llama3", "Model sha": "5f0b02c75b57c5855da9ae460ce51323ea669d8a", "model_name_for_query": "meta-llama/Meta-Llama-3-8B-Instruct", "General Knowledge": "52.04%", "GSM8K": "10.40%", "Homegraph Easy": "41.67%", "Homograph Hard": "81.11%", "PIQA": "69.26%", "Proverb": "42.97%", "Verb Tense": "38.93%", "Winogrande": "55.24%"}
|
8 |
+
{"Model": "NousResearch/Hermes-3-Llama-3.1-8B", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 35.01, "ARC Easy": 77.01, "ARC Challenge": 58.39, "MMLU Pro": 21.0, "AUT Multiple Choice Persian": 52.46, "Hub License": "llama3", "Model sha": "896ea440e5a9e6070e3d8a2774daf2b481ab425b", "model_name_for_query": "NousResearch/Hermes-3-Llama-3.1-8B", "General Knowledge": "49.49%", "GSM8K": "10.20%", "Homegraph Easy": "44.44%", "Homograph Hard": "79.72%", "PIQA": "69.26%", "Proverb": "47.84%", "Verb Tense": "48.94%", "Winogrande": "54.03%"}
|
9 |
+
{"Model": "CohereForAI/aya-expanse-32b", "Precision": "float16", "#Params (B)": 32.3, "Part Multiple Choice": 43.36, "ARC Easy": 93.1, "ARC Challenge": 79.87, "MMLU Pro": 31.03, "AUT Multiple Choice Persian": 62.33, "Hub License": "cc-by-nc-4.0", "Model sha": "94bda1dcb97d260f732d230b832c7c685ae91e23", "model_name_for_query": "CohereForAI/aya-expanse-32b", "General Knowledge": "73.72%", "GSM8K": "17.50%", "Homegraph Easy": "62.96%", "Homograph Hard": "87.56%", "PIQA": "88.57%", "Proverb": "77.03%", "Verb Tense": "61.95%", "Winogrande": "67.18%"}
|
10 |
+
{"Model": "CohereForAI/aya-23-8B", "Precision": "float16", "#Params (B)": 8.03, "Part Multiple Choice": 32.82, "ARC Easy": 80.46, "ARC Challenge": 64.43, "MMLU Pro": 18.62, "AUT Multiple Choice Persian": 52.86, "Hub License": "cc-by-nc-4.0", "Model sha": "2a1a63b24af8f591616fdf58936ee576d63ca835", "model_name_for_query": "CohereForAI/aya-23-8B", "General Knowledge": "52.30%", "GSM8K": "6.10%", "Homegraph Easy": "52.78%", "Homograph Hard": "76.27%", "PIQA": "78.77%", "Proverb": "44.32%", "Verb Tense": "39.30%", "Winogrande": "54.27%"}
|
11 |
+
{"Model": "meta-llama/Llama-3.1-8B-Instruct", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 36.68, "ARC Easy": 78.4, "ARC Challenge": 60.4, "MMLU Pro": 21.0, "AUT Multiple Choice Persian": 54.24, "Hub License": "llama3.1", "Model sha": "0e9e39f249a16976918f6564b8830bc894c89659", "model_name_for_query": "meta-llama/Llama-3.1-8B-Instruct", "General Knowledge": "52.55%", "GSM8K": "12.00%", "Homegraph Easy": "43.52%", "Homograph Hard": "79.03%", "PIQA": "69.26%", "Proverb": "47.57%", "Verb Tense": "42.91%", "Winogrande": "53.79%"}
|
12 |
+
{"Model": "google/gemma-2-2b-it", "Precision": "bfloat16", "#Params (B)": 2.61, "Part Multiple Choice": 31.12, "ARC Easy": 71.26, "ARC Challenge": 57.72, "MMLU Pro": 16.23, "AUT Multiple Choice Persian": 49.9, "Hub License": "gemma", "Model sha": "299a8560bedf22ed1c72a8a11e7dce4a7f9f51f8", "model_name_for_query": "google/gemma-2-2b-it", "General Knowledge": "32.91%", "GSM8K": "6.40%", "Homegraph Easy": "47.22%", "Homograph Hard": "74.65%", "PIQA": "65.80%", "Proverb": "45.68%", "Verb Tense": "36.18%", "Winogrande": "53.23%"}
|
13 |
+
{"Model": "meta-llama/Meta-Llama-3-8B", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 34.32, "ARC Easy": 72.12, "ARC Challenge": 56.37, "MMLU Pro": 19.33, "AUT Multiple Choice Persian": 48.32, "Hub License": "llama3", "Model sha": "8cde5ca8380496c9a6cc7ef3a8b46a0372a1d920", "model_name_for_query": "meta-llama/Meta-Llama-3-8B", "General Knowledge": "47.70%", "GSM8K": "10.30%", "Homegraph Easy": "41.67%", "Homograph Hard": "74.42%", "PIQA": "63.59%", "Proverb": "37.30%", "Verb Tense": "39.46%", "Winogrande": "54.60%"}
|
14 |
+
{"Model": "google/gemma-2-9b-it", "Precision": "bfloat16", "#Params (B)": 9.24, "Part Multiple Choice": 42.7, "ARC Easy": 93.1, "ARC Challenge": 84.56, "MMLU Pro": 31.74, "AUT Multiple Choice Persian": 62.33, "Hub License": "gemma", "Model sha": "11c9b309abf73637e4b6f9a3fa1e92e615547819", "model_name_for_query": "google/gemma-2-9b-it", "General Knowledge": "64.03%", "GSM8K": "17.40%", "Homegraph Easy": "59.26%", "Homograph Hard": "90.55%", "PIQA": "84.63%", "Proverb": "69.19%", "Verb Tense": "58.25%", "Winogrande": "67.26%"}
|
15 |
+
{"Model": "PartAI/Dorna-Llama3-8B-Instruct", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 33.88, "ARC Easy": 70.4, "ARC Challenge": 61.07, "MMLU Pro": 23.39, "AUT Multiple Choice Persian": 52.86, "Hub License": "llama3", "Model sha": "fb268bb51b950b4db5b7c82c1b73d9e803020eed", "model_name_for_query": "PartAI/Dorna-Llama3-8B-Instruct", "General Knowledge": "41.33%", "GSM8K": "10.30%", "Homegraph Easy": "40.74%", "Homograph Hard": "74.65%", "PIQA": "64.84%", "Proverb": "35.41%", "Verb Tense": "34.74%", "Winogrande": "54.27%"}
|
16 |
+
{"Model": "MaralGPT/Maral-7B-alpha-1", "Precision": "bfloat16", "#Params (B)": 7.24, "Part Multiple Choice": 26.67, "ARC Easy": 44.54, "ARC Challenge": 30.87, "MMLU Pro": 15.99, "AUT Multiple Choice Persian": 36.09, "Hub License": "mit", "Model sha": "2ab5ca2a0d1a4454a78b4ca911e595bb9da2fe2f", "model_name_for_query": "MaralGPT/Maral-7B-alpha-1", "General Knowledge": "31.63%", "GSM8K": "6.10%", "Homegraph Easy": "43.52%", "Homograph Hard": "47.47%", "PIQA": "52.45%", "Proverb": "22.16%", "Verb Tense": "28.96%", "Winogrande": "50.40%"}
|
17 |
+
{"Model": "Qwen/QwQ-32B-Preview", "Precision": "bfloat16", "#Params (B)": 32.8, "Part Multiple Choice": 46.64, "ARC Easy": 91.95, "ARC Challenge": 87.24, "MMLU Pro": 37.94, "AUT Multiple Choice Persian": 60.15, "Hub License": "apache-2.0", "Model sha": "91906fe41a48b6a89ce2970abfd1269eefee170e", "model_name_for_query": "Qwen/QwQ-32B-Preview", "General Knowledge": "63.27%", "GSM8K": "34.70%", "Homegraph Easy": "61.11%", "Homograph Hard": "88.25%", "PIQA": "78.58%", "Proverb": "58.11%", "Verb Tense": "51.97%", "Winogrande": "71.45%"}
|
18 |
+
{"Model": "Qwen/QwQ-32B", "Precision": "bfloat16", "#Params (B)": 32.8, "Part Multiple Choice": 46.48, "ARC Easy": 89.37, "ARC Challenge": 82.55, "MMLU Pro": 35.32, "AUT Multiple Choice Persian": 57.4, "Hub License": "apache-2.0", "Model sha": "976055f8c83f394f35dbd3ab09a285a984907bd0", "model_name_for_query": "Qwen/QwQ-32B", "General Knowledge": "60.71%", "GSM8K": "29.30%", "Homegraph Easy": "58.33%", "Homograph Hard": "88.25%", "PIQA": "79.06%", "Proverb": "59.19%", "Verb Tense": "52.31%", "Winogrande": "68.79%"}
|
19 |
+
{"Model": "CohereForAI/aya-23-35B", "Precision": "float16", "#Params (B)": 35, "Part Multiple Choice": 36.79, "ARC Easy": 87.93, "ARC Challenge": 72.48, "MMLU Pro": 25.54, "AUT Multiple Choice Persian": 61.14, "Hub License": "cc-by-nc-4.0", "Model sha": "5e72bd5ad83e5e1612ee7f56a0c1a439a7cfb887", "model_name_for_query": "CohereForAI/aya-23-35B", "General Knowledge": "63.27%", "GSM8K": "10.00%", "Homegraph Easy": "55.56%", "Homograph Hard": "83.64%", "PIQA": "87.51%", "Proverb": "67.03%", "Verb Tense": "47.32%", "Winogrande": "62.66%"}
|
20 |
+
{"Model": "google/gemma-3-1b-it", "Precision": "bfloat16", "#Params (B)": 1, "Part Multiple Choice": 27.15, "ARC Easy": 41.67, "ARC Challenge": 31.54, "MMLU Pro": 16.23, "AUT Multiple Choice Persian": 36.88, "Hub License": "gemma", "Model sha": "dcc83ea841ab6100d6b47a070329e1ba4cf78752", "model_name_for_query": "google/gemma-3-1b-it", "General Knowledge": "26.02%", "GSM8K": "4.30%", "Homegraph Easy": "49.07%", "Homograph Hard": "51.15%", "PIQA": "56.96%", "Proverb": "28.92%", "Verb Tense": "27.67%", "Winogrande": "50.65%"}
|
21 |
+
{"Model": "CohereForAI/aya-expanse-8b", "Precision": "float16", "#Params (B)": 8.03, "Part Multiple Choice": 34.91, "ARC Easy": 79.6, "ARC Challenge": 70.47, "MMLU Pro": 25.06, "AUT Multiple Choice Persian": 58.38, "Hub License": "cc-by-nc-4.0", "Model sha": "0ad43ec1e309e1351faa4b1d22713c065e37359a", "model_name_for_query": "CohereForAI/aya-expanse-8b", "General Knowledge": "58.67%", "GSM8K": "9.80%", "Homegraph Easy": "51.85%", "Homograph Hard": "80.65%", "PIQA": "78.00%", "Proverb": "60.00%", "Verb Tense": "48.06%", "Winogrande": "59.84%"}
|
22 |
+
{"Model": "google/gemma-2-27b-it", "Precision": "bfloat16", "#Params (B)": 27.2, "Part Multiple Choice": 46.03, "ARC Easy": 95.98, "ARC Challenge": 85.91, "MMLU Pro": 36.28, "AUT Multiple Choice Persian": 63.12, "Hub License": "gemma", "Model sha": "aaf20e6b9f4c0fcf043f6fb2a2068419086d77b0", "model_name_for_query": "google/gemma-2-27b-it", "General Knowledge": "68.11%", "GSM8K": "26.70%", "Homegraph Easy": "60.19%", "Homograph Hard": "91.24%", "PIQA": "87.61%", "Proverb": "73.51%", "Verb Tense": "61.16%", "Winogrande": "71.85%"}
|
23 |
+
{"Model": "Qwen/Qwen2.5-32B-Instruct", "Precision": "bfloat16", "#Params (B)": 32.8, "Part Multiple Choice": 46.06, "ARC Easy": 90.8, "ARC Challenge": 85.91, "MMLU Pro": 38.19, "AUT Multiple Choice Persian": 61.34, "Hub License": "apache-2.0", "Model sha": "5ede1c97bbab6ce5cda5812749b4c0bdf79b18dd", "model_name_for_query": "Qwen/Qwen2.5-32B-Instruct", "General Knowledge": "61.73%", "GSM8K": "50.10%", "Homegraph Easy": "67.59%", "Homograph Hard": "91.47%", "PIQA": "81.65%", "Proverb": "63.24%", "Verb Tense": "54.58%", "Winogrande": "75.40%"}
|
24 |
+
{"Model": "universitytehran/PersianMind-v1.0", "Precision": "bfloat16", "#Params (B)": 6.82, "Part Multiple Choice": 29.27, "ARC Easy": 58.91, "ARC Challenge": 48.32, "MMLU Pro": 15.51, "AUT Multiple Choice Persian": 45.36, "Hub License": "cc-by-nc-sa-4.0", "Model sha": "af603eeb074138e2a613fbc95d89f018afbd3041", "model_name_for_query": "universitytehran/PersianMind-v1.0", "General Knowledge": "30.61%", "GSM8K": "2.30%", "Homegraph Easy": "41.67%", "Homograph Hard": "65.90%", "PIQA": "59.94%", "Proverb": "34.32%", "Verb Tense": "26.26%", "Winogrande": "52.26%"}
|
25 |
+
{"Model": "google/gemma-3-27b-it", "Precision": "bfloat16", "#Params (B)": 27.4, "Part Multiple Choice": 48.56, "ARC Easy": 95.69, "ARC Challenge": 90.6, "MMLU Pro": 40.1, "AUT Multiple Choice Persian": 64.3, "Hub License": "gemma", "Model sha": "005ad3404e59d6023443cb575daa05336842228a", "model_name_for_query": "google/gemma-3-27b-it", "General Knowledge": "73.72%", "GSM8K": "28.30%", "Homegraph Easy": "63.89%", "Homograph Hard": "92.40%", "PIQA": "85.30%", "Proverb": "78.92%", "Verb Tense": "66.02%", "Winogrande": "73.47%"}
|
26 |
+
{"Model": "meta-llama/Llama-3.1-8B", "Precision": "bfloat16", "#Params (B)": 8.03, "Part Multiple Choice": 35.52, "ARC Easy": 75.28, "ARC Challenge": 53.02, "MMLU Pro": 24.1, "AUT Multiple Choice Persian": 53.45, "Hub License": "llama3.1", "Model sha": "d04e592bb4f6aa9cfee91e2e20afa771667e1d4b", "model_name_for_query": "meta-llama/Llama-3.1-8B", "General Knowledge": "49.23%", "GSM8K": "10.80%", "Homegraph Easy": "46.30%", "Homograph Hard": "72.12%", "PIQA": "65.90%", "Proverb": "35.95%", "Verb Tense": "39.91%", "Winogrande": "53.39%"}
|
27 |
+
{"Model": "google/gemma-3-12b-it", "Precision": "bfloat16", "#Params (B)": 12.2, "Part Multiple Choice": 43.52, "ARC Easy": 93.39, "ARC Challenge": 81.21, "MMLU Pro": 29.36, "AUT Multiple Choice Persian": 57.0, "Hub License": "gemma", "Model sha": "96b6f1eccf38110c56df3a15bffe176da04bfd80", "model_name_for_query": "google/gemma-3-12b-it", "General Knowledge": "68.37%", "GSM8K": "20.20%", "Homegraph Easy": "67.59%", "Homograph Hard": "91.24%", "PIQA": "85.11%", "Proverb": "72.97%", "Verb Tense": "63.39%", "Winogrande": "69.84%"}
|