type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
sequencelengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model | Easy-Systems/easy-ko-Llama3-8b-Instruct-v1 | 1 | [
"main"
] | 16,069,723,822 |
model | Litzy619/PHI30512HMAB25H | 1 | [
"main"
] | 8,194,835,164 |
model | presencesw/xlm-roberta-large-snli_neutral-triplet | 1 | [
"main"
] | 2,239,630,384 |
model | kowlick/c4ai-command-r-v01-Q8_0-GGUF | 1 | [
"main"
] | 37,179,016,744 |
model | AlignmentResearch/robust_llm_pythia-1b_niki-052_imdb_gcg_seed-0 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | sally9805/bert-base-uncased-finetuned-news-2008 | 1 | [
"main"
] | 439,054,284 |
model | Lifeisegg199/llama-3-cat-8b-instruct-v1-Q5_K_S-GGUF | 1 | [
"main"
] | 5,599,296,664 |
model | kkotkar1/llama3-reft | 1 | [
"main"
] | 304,531 |
model | MoMonir/Yi-1.5-9B-Chat-GGUF | 1 | [
"main"
] | 18,832,860,799 |
model | AdhamEhab/StoryGen | 1 | [
"main"
] | 501,176,492 |
model | shkna1368/mt5-small-finetuned-mt5-small-poemV1 | 1 | [
"main"
] | 1,519 |
model | kaengpil/iland-kangjiwon | 1 | [
"main"
] | 74,761,391 |
model | twodigit/Meta-Llama-3-8B-Instruct-koconv2_4327k-sft-full-500000-keyword_23k | 1 | [
"main"
] | 16,069,725,414 |
model | ranggaaldosas/llama-8b | 1 | [
"main"
] | 680,289,508 |
model | kapilrk04/indicbart_based_enhi_hi_mt_model | 1 | [
"main"
] | 983,355,869 |
model | isenbek/OrpoLlama-3-8B | 1 | [
"main"
] | 16,069,758,159 |
model | bhsai2709/T7_Llama_readmission_prediction | 1 | [
"main"
] | 13,479,240,438 |
model | compressa-ai/openchat-3.5-0106-GPTQ-medchat-LoRA | 1 | [
"main"
] | 56,864,876 |
model | nrishabh/llama3-8b-instruct-qlora-minute | 1 | [
"main"
] | 227,283,220 |
model | mg0515/my_awesome_eli5_mlm_model | 1 | [
"main"
] | 328,722,553 |
model | NousResearch/Hermes-2-Theta-Llama-3-8B-GGUF | 1 | [
"main"
] | 41,859,407,184 |
model | MoMonir/Yi-1.5-6B-Chat-GGUF | 1 | [
"main"
] | 12,952,682,912 |
model | Crysiss/llama-3-8B-sql-dpo-v0.2 | 1 | [
"main"
] | 671,151,993 |
model | ranggaaldosas/llama-8b-16bit | 1 | [
"main"
] | 16,069,788,954 |
model | rzhen0615/my_man_David_LoRA | 1 | [
"main"
] | 23,523,797 |
model | Rajpatel013/BARTModel_for_Ecommerce | 1 | [
"main"
] | 561,303,374 |
model | compressa-ai/openchat-3.5-0106-medchat-LoRA | 1 | [
"main"
] | 84,462,557 |
model | microsoft/kosmos-2.5 | 1 | [
"main"
] | 11,673,963,089 |
model | llama-duo/gemma7b-summarize | 1 | [
"main"
] | 42,560,731 |
model | rzou/Fine-tuned | 1 | [
"main"
] | 176,971,839 |
model | lpaovien8957/bert-finetuned-squad | 1 | [
"main"
] | 1,519 |
model | diwanshus/codequalbert | 1 | [
"main"
] | 498,623,516 |
model | vilarin/Yi-1.5-6B-Chat-CN-Law | 1 | [
"main"
] | 12,126,729,383 |
model | bethcherie/butterflies | 1 | [
"main"
] | 1,518,418 |
model | compressa-ai/Meta-Llama-3-8B-Instruct-medchat-LoRA | 1 | [
"main"
] | 91,350,898 |
model | mssma/ko-solar-10.7b-v0.2b | 1 | [
"main"
] | 21,704,888,242 |
model | YuChern/desired-model-name | 1 | [
"main"
] | 268,784,208 |
model | RayanNan/Llama3-8b-4bits | 1 | [
"main"
] | 1,519 |
model | Ankita802/test_part | 1 | [
"main"
] | 1,396,786 |
model | Karlsen/Class_university_exercise | 1 | [
"main"
] | 1,878,029 |
model | Litzy619/PHI30512HMAB19H | 1 | [
"main"
] | 8,194,835,164 |
model | shkna1368/mt5-small-finetuned-mt5-small-poem-hemen | 1 | [
"main"
] | 1,205,049,818 |
model | Crysiss/llama-3-8B-sql-dpo-v0.3 | 1 | [
"main"
] | 671,151,993 |
model | lainshower/Llama3-8b-alpaca-v2 | 1 | [
"main"
] | 16,069,723,618 |
model | BahaaEldin0/Mistral-7b-Mixtral-7b | 1 | [
"main"
] | 14,485,814,324 |
model | tctsung/TinyLlama-1.1B-chat-v1.0-awq | 1 | [
"main"
] | 767,599,581 |
model | Ichigo2899/01-ai-Yi-1.5-9B-AWQ | 1 | [
"main"
] | 5,368,656,864 |
model | blockblockblock/Yi-1.5-34B-Chat-bpw5-exl2 | 1 | [
"main"
] | 42,833,433,504 |
model | Crowlley/DatasetToolsUpscalerModels | 1 | [
"main"
] | 926,654,115 |
model | Mantis-VL/mantis-8b-idefics2_8192_qlora | 1 | [
"main"
] | 49,026,891 |
model | PJMixers/MV02-PB-Mixture-v1-run_15-SFT-7B-Latest-QLoRA | 1 | [
"main"
] | 1,197,766,249 |
model | mostafiz26/imdbreviews_classification_codet5 | 1 | [
"main"
] | 1,519 |
model | Wilber87vn/whisper-tiny-hass-vn | 1 | [
"main"
] | 453,245,790 |
model | gogo8232/pythia-1.3b-before-healing | 1 | [
"main"
] | 2,627,901,852 |
model | tkempto1/hybrid-qa1 | 1 | [
"main"
] | 1,158,074,265 |
model | RichardErkhov/google_-_recurrentgemma-2b-4bits | 1 | [
"main"
] | 2,508,685,786 |
model | Ichigo2899/01-ai-Yi-1.5-9B-Chat-AWQ | 1 | [
"main"
] | 5,368,658,713 |
model | cpajitha/t5-small-finetuned-new-gettitle | 1 | [
"main"
] | 245,302,284 |
model | Wilber87vn/faster-whisper-tiny-hass-vn | 1 | [
"main"
] | 79,142,216 |
model | arutema47/mistral-7b-v0.2-trtllm | 1 | [
"main"
] | 14,503,179,777 |
model | RichardErkhov/google_-_recurrentgemma-2b-8bits | 1 | [
"main"
] | 3,387,453,666 |
model | spsither/tibetan_RoBERTa_S_e2 | 1 | [
"main"
] | 340,721,009 |
model | Sancheon/AiDok | 1 | [
"main"
] | 1,543 |
model | hsgwktb/model_xl | 1 | [
"main"
] | 37,145,080,374 |
model | Rebecca19990101/westbreeze-7b-q4_k_m.gguf | 1 | [
"main"
] | 4,469,709,293 |
model | epiverseai/llama3-8b-r-data-science | 1 | [
"main"
] | 680,298,717 |
model | hskhyl/05-13_4 | 1 | [
"main"
] | 43,105,509 |
model | sally9805/bert-base-uncased-finetuned-news-2003-2007 | 1 | [
"main"
] | 439,068,934 |
model | Thatsnazzyartist22/JaxV2 | 1 | [
"main"
] | 84,831,770 |
model | yuyu328/drive-lm-test | 1 | [
"main"
] | 13,431,368 |
model | HuanYangChang/ppo-LunarLander-v2 | 1 | [
"main"
] | 487,455 |
model | gogo8232/pythia-1.3b-after-healing | 1 | [
"main"
] | 2,630,021,169 |
model | senthilkumar18/testing-model | 1 | [
"main"
] | 1,536 |
model | mssma/ko-solar-10.7b-v0.1b | 1 | [
"main"
] | 21,704,888,242 |
model | RichardErkhov/pfnet_-_plamo-13b-gguf | 1 | [
"main"
] | 172,166,305,475 |
model | SimplCup/MoistcriticalDereverbed | 1 | [
"main"
] | 87,368,313 |
model | byh711/Phi-3-mini-4k-dementia | 1 | [
"main"
] | 103,101,943 |
model | Zawak8/manicure1 | 1 | [
"main"
] | 1,519 |
model | Sinha07/fine_tuned_model | 1 | [
"main"
] | 497,781,941 |
model | kat0t0y/syriltest | 1 | [
"main"
] | 1,550 |
model | epiverseai/llama3-8b-sivirep | 1 | [
"main"
] | 680,306,473 |
model | XueyingJia/llama3_gsm8k_8_shots_answer_version_full | 1 | [
"main"
] | 93,084,783 |
model | nluai/question-generation-vietnamese-v2 | 1 | [
"main"
] | 2,350,356,960 |
model | shkna1368/mt5-base-finetuned-mt5-base-poem-hemen | 1 | [
"main"
] | 2,333,957,464 |
model | cjsanjay/llama-3-8B-gorilla-meraki_v2 | 1 | [
"main"
] | 16,069,724,261 |
model | Shanzey/distilbert-base-uncased-finetuned-emotion | 1 | [
"main"
] | 268,808,166 |
model | mssma/ko-solar-10.7b-v0.3b | 1 | [
"main"
] | 21,704,888,242 |
model | Moon-Ahn/komistral-7b-v1-q4f16_1-MLC | 1 | [
"main"
] | 4,074,885,053 |
model | gogo8232/pythia-1.4b-healed-expaned-raw | 1 | [
"main"
] | 2,831,457,065 |
model | RomBor/PPO-LunarLander-v2 | 1 | [
"main"
] | 475,358 |
model | Litzy619/PHI30512HMAB26H | 1 | [
"main"
] | 8,194,835,164 |
model | arvnoodle/llama3-coder-8b-nativeformat-xml-json-051324 | 1 | [
"main"
] | 167,835,094 |
model | MJerome/V58_LoRA_V31_Leon-Chess-350k-Plus_13k_low_elo_20E_r64 | 1 | [
"main"
] | 9,453,210 |
model | krishnakalyan3/emo_xfw5q8hr_125 | 1 | [
"main"
] | 382,292,136 |
model | djsull/setfit_classifier | 1 | [
"main"
] | 445,446,831 |
model | ImagineIt/alpha-r-s-v-12 | 1 | [
"main"
] | 16,069,720,125 |
model | Holarissun/dpo_tldrtldr_contrast_subset10000_modelgemma2b_maxsteps5000_bz8_lr1e-05 | 1 | [
"main"
] | 32,322,805 |
model | LA1512/led-pubmed-20K | 1 | [
"main"
] | 650,989,324 |
model | ikhsanlaode/lora_model | 1 | [
"main"
] | 1,679 |
model | songhyundong/my-new-shiny-tokenizer | 1 | [
"main"
] | 1,680,529 |
Subsets and Splits