type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
listlengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model
|
Easy-Systems/easy-ko-Llama3-8b-Instruct-v1
| 1 |
[
"main"
] | 16,069,723,822 |
model
|
Litzy619/PHI30512HMAB25H
| 1 |
[
"main"
] | 8,194,835,164 |
model
|
presencesw/xlm-roberta-large-snli_neutral-triplet
| 1 |
[
"main"
] | 2,239,630,384 |
model
|
kowlick/c4ai-command-r-v01-Q8_0-GGUF
| 1 |
[
"main"
] | 37,179,016,744 |
model
|
AlignmentResearch/robust_llm_pythia-1b_niki-052_imdb_gcg_seed-0
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
sally9805/bert-base-uncased-finetuned-news-2008
| 1 |
[
"main"
] | 439,054,284 |
model
|
Lifeisegg199/llama-3-cat-8b-instruct-v1-Q5_K_S-GGUF
| 1 |
[
"main"
] | 5,599,296,664 |
model
|
kkotkar1/llama3-reft
| 1 |
[
"main"
] | 304,531 |
model
|
MoMonir/Yi-1.5-9B-Chat-GGUF
| 1 |
[
"main"
] | 18,832,860,799 |
model
|
AdhamEhab/StoryGen
| 1 |
[
"main"
] | 501,176,492 |
model
|
shkna1368/mt5-small-finetuned-mt5-small-poemV1
| 1 |
[
"main"
] | 1,519 |
model
|
kaengpil/iland-kangjiwon
| 1 |
[
"main"
] | 74,761,391 |
model
|
twodigit/Meta-Llama-3-8B-Instruct-koconv2_4327k-sft-full-500000-keyword_23k
| 1 |
[
"main"
] | 16,069,725,414 |
model
|
ranggaaldosas/llama-8b
| 1 |
[
"main"
] | 680,289,508 |
model
|
kapilrk04/indicbart_based_enhi_hi_mt_model
| 1 |
[
"main"
] | 983,355,869 |
model
|
isenbek/OrpoLlama-3-8B
| 1 |
[
"main"
] | 16,069,758,159 |
model
|
bhsai2709/T7_Llama_readmission_prediction
| 1 |
[
"main"
] | 13,479,240,438 |
model
|
compressa-ai/openchat-3.5-0106-GPTQ-medchat-LoRA
| 1 |
[
"main"
] | 56,864,876 |
model
|
nrishabh/llama3-8b-instruct-qlora-minute
| 1 |
[
"main"
] | 227,283,220 |
model
|
mg0515/my_awesome_eli5_mlm_model
| 1 |
[
"main"
] | 328,722,553 |
model
|
NousResearch/Hermes-2-Theta-Llama-3-8B-GGUF
| 1 |
[
"main"
] | 41,859,407,184 |
model
|
MoMonir/Yi-1.5-6B-Chat-GGUF
| 1 |
[
"main"
] | 12,952,682,912 |
model
|
Crysiss/llama-3-8B-sql-dpo-v0.2
| 1 |
[
"main"
] | 671,151,993 |
model
|
ranggaaldosas/llama-8b-16bit
| 1 |
[
"main"
] | 16,069,788,954 |
model
|
rzhen0615/my_man_David_LoRA
| 1 |
[
"main"
] | 23,523,797 |
model
|
Rajpatel013/BARTModel_for_Ecommerce
| 1 |
[
"main"
] | 561,303,374 |
model
|
compressa-ai/openchat-3.5-0106-medchat-LoRA
| 1 |
[
"main"
] | 84,462,557 |
model
|
microsoft/kosmos-2.5
| 1 |
[
"main"
] | 11,673,963,089 |
model
|
llama-duo/gemma7b-summarize
| 1 |
[
"main"
] | 42,560,731 |
model
|
rzou/Fine-tuned
| 1 |
[
"main"
] | 176,971,839 |
model
|
lpaovien8957/bert-finetuned-squad
| 1 |
[
"main"
] | 1,519 |
model
|
diwanshus/codequalbert
| 1 |
[
"main"
] | 498,623,516 |
model
|
vilarin/Yi-1.5-6B-Chat-CN-Law
| 1 |
[
"main"
] | 12,126,729,383 |
model
|
bethcherie/butterflies
| 1 |
[
"main"
] | 1,518,418 |
model
|
compressa-ai/Meta-Llama-3-8B-Instruct-medchat-LoRA
| 1 |
[
"main"
] | 91,350,898 |
model
|
mssma/ko-solar-10.7b-v0.2b
| 1 |
[
"main"
] | 21,704,888,242 |
model
|
YuChern/desired-model-name
| 1 |
[
"main"
] | 268,784,208 |
model
|
RayanNan/Llama3-8b-4bits
| 1 |
[
"main"
] | 1,519 |
model
|
Ankita802/test_part
| 1 |
[
"main"
] | 1,396,786 |
model
|
Karlsen/Class_university_exercise
| 1 |
[
"main"
] | 1,878,029 |
model
|
Litzy619/PHI30512HMAB19H
| 1 |
[
"main"
] | 8,194,835,164 |
model
|
shkna1368/mt5-small-finetuned-mt5-small-poem-hemen
| 1 |
[
"main"
] | 1,205,049,818 |
model
|
Crysiss/llama-3-8B-sql-dpo-v0.3
| 1 |
[
"main"
] | 671,151,993 |
model
|
lainshower/Llama3-8b-alpaca-v2
| 1 |
[
"main"
] | 16,069,723,618 |
model
|
BahaaEldin0/Mistral-7b-Mixtral-7b
| 1 |
[
"main"
] | 14,485,814,324 |
model
|
tctsung/TinyLlama-1.1B-chat-v1.0-awq
| 1 |
[
"main"
] | 767,599,581 |
model
|
Ichigo2899/01-ai-Yi-1.5-9B-AWQ
| 1 |
[
"main"
] | 5,368,656,864 |
model
|
blockblockblock/Yi-1.5-34B-Chat-bpw5-exl2
| 1 |
[
"main"
] | 42,833,433,504 |
model
|
Crowlley/DatasetToolsUpscalerModels
| 1 |
[
"main"
] | 926,654,115 |
model
|
Mantis-VL/mantis-8b-idefics2_8192_qlora
| 1 |
[
"main"
] | 49,026,891 |
model
|
PJMixers/MV02-PB-Mixture-v1-run_15-SFT-7B-Latest-QLoRA
| 1 |
[
"main"
] | 1,197,766,249 |
model
|
mostafiz26/imdbreviews_classification_codet5
| 1 |
[
"main"
] | 1,519 |
model
|
Wilber87vn/whisper-tiny-hass-vn
| 1 |
[
"main"
] | 453,245,790 |
model
|
gogo8232/pythia-1.3b-before-healing
| 1 |
[
"main"
] | 2,627,901,852 |
model
|
tkempto1/hybrid-qa1
| 1 |
[
"main"
] | 1,158,074,265 |
model
|
RichardErkhov/google_-_recurrentgemma-2b-4bits
| 1 |
[
"main"
] | 2,508,685,786 |
model
|
Ichigo2899/01-ai-Yi-1.5-9B-Chat-AWQ
| 1 |
[
"main"
] | 5,368,658,713 |
model
|
cpajitha/t5-small-finetuned-new-gettitle
| 1 |
[
"main"
] | 245,302,284 |
model
|
Wilber87vn/faster-whisper-tiny-hass-vn
| 1 |
[
"main"
] | 79,142,216 |
model
|
arutema47/mistral-7b-v0.2-trtllm
| 1 |
[
"main"
] | 14,503,179,777 |
model
|
RichardErkhov/google_-_recurrentgemma-2b-8bits
| 1 |
[
"main"
] | 3,387,453,666 |
model
|
spsither/tibetan_RoBERTa_S_e2
| 1 |
[
"main"
] | 340,721,009 |
model
|
Sancheon/AiDok
| 1 |
[
"main"
] | 1,543 |
model
|
hsgwktb/model_xl
| 1 |
[
"main"
] | 37,145,080,374 |
model
|
Rebecca19990101/westbreeze-7b-q4_k_m.gguf
| 1 |
[
"main"
] | 4,469,709,293 |
model
|
epiverseai/llama3-8b-r-data-science
| 1 |
[
"main"
] | 680,298,717 |
model
|
hskhyl/05-13_4
| 1 |
[
"main"
] | 43,105,509 |
model
|
sally9805/bert-base-uncased-finetuned-news-2003-2007
| 1 |
[
"main"
] | 439,068,934 |
model
|
Thatsnazzyartist22/JaxV2
| 1 |
[
"main"
] | 84,831,770 |
model
|
yuyu328/drive-lm-test
| 1 |
[
"main"
] | 13,431,368 |
model
|
HuanYangChang/ppo-LunarLander-v2
| 1 |
[
"main"
] | 487,455 |
model
|
gogo8232/pythia-1.3b-after-healing
| 1 |
[
"main"
] | 2,630,021,169 |
model
|
senthilkumar18/testing-model
| 1 |
[
"main"
] | 1,536 |
model
|
mssma/ko-solar-10.7b-v0.1b
| 1 |
[
"main"
] | 21,704,888,242 |
model
|
RichardErkhov/pfnet_-_plamo-13b-gguf
| 1 |
[
"main"
] | 172,166,305,475 |
model
|
SimplCup/MoistcriticalDereverbed
| 1 |
[
"main"
] | 87,368,313 |
model
|
byh711/Phi-3-mini-4k-dementia
| 1 |
[
"main"
] | 103,101,943 |
model
|
Zawak8/manicure1
| 1 |
[
"main"
] | 1,519 |
model
|
Sinha07/fine_tuned_model
| 1 |
[
"main"
] | 497,781,941 |
model
|
kat0t0y/syriltest
| 1 |
[
"main"
] | 1,550 |
model
|
epiverseai/llama3-8b-sivirep
| 1 |
[
"main"
] | 680,306,473 |
model
|
XueyingJia/llama3_gsm8k_8_shots_answer_version_full
| 1 |
[
"main"
] | 93,084,783 |
model
|
nluai/question-generation-vietnamese-v2
| 1 |
[
"main"
] | 2,350,356,960 |
model
|
shkna1368/mt5-base-finetuned-mt5-base-poem-hemen
| 1 |
[
"main"
] | 2,333,957,464 |
model
|
cjsanjay/llama-3-8B-gorilla-meraki_v2
| 1 |
[
"main"
] | 16,069,724,261 |
model
|
Shanzey/distilbert-base-uncased-finetuned-emotion
| 1 |
[
"main"
] | 268,808,166 |
model
|
mssma/ko-solar-10.7b-v0.3b
| 1 |
[
"main"
] | 21,704,888,242 |
model
|
Moon-Ahn/komistral-7b-v1-q4f16_1-MLC
| 1 |
[
"main"
] | 4,074,885,053 |
model
|
gogo8232/pythia-1.4b-healed-expaned-raw
| 1 |
[
"main"
] | 2,831,457,065 |
model
|
RomBor/PPO-LunarLander-v2
| 1 |
[
"main"
] | 475,358 |
model
|
Litzy619/PHI30512HMAB26H
| 1 |
[
"main"
] | 8,194,835,164 |
model
|
arvnoodle/llama3-coder-8b-nativeformat-xml-json-051324
| 1 |
[
"main"
] | 167,835,094 |
model
|
MJerome/V58_LoRA_V31_Leon-Chess-350k-Plus_13k_low_elo_20E_r64
| 1 |
[
"main"
] | 9,453,210 |
model
|
krishnakalyan3/emo_xfw5q8hr_125
| 1 |
[
"main"
] | 382,292,136 |
model
|
djsull/setfit_classifier
| 1 |
[
"main"
] | 445,446,831 |
model
|
ImagineIt/alpha-r-s-v-12
| 1 |
[
"main"
] | 16,069,720,125 |
model
|
Holarissun/dpo_tldrtldr_contrast_subset10000_modelgemma2b_maxsteps5000_bz8_lr1e-05
| 1 |
[
"main"
] | 32,322,805 |
model
|
LA1512/led-pubmed-20K
| 1 |
[
"main"
] | 650,989,324 |
model
|
ikhsanlaode/lora_model
| 1 |
[
"main"
] | 1,679 |
model
|
songhyundong/my-new-shiny-tokenizer
| 1 |
[
"main"
] | 1,680,529 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.