type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
sequencelengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model | ashnadua01/model_muril_small_noise | 1 | [
"main"
] | 957,531,409 |
model | gbueno86/Meta-Llama-3-120b-LumiLumimaid | 1 | [
"main"
] | 243,795,340,360 |
model | liilliliiiliil/VALORANT_SOVA | 1 | [
"main"
] | 115,055,119 |
model | mradermacher/Rhea-125-v0.5-i1-GGUF | 1 | [
"main"
] | 1,207,383,908,874 |
model | sel639/llava-1.5-7b-hf-ft-bill-epoch-50-continue | 1 | [
"main"
] | 1,519 |
model | NicolasAdrian07/Lockjaw_TRTF | 1 | [
"main"
] | 54,609,143 |
model | fth66/2nd_sub | 1 | [
"main"
] | 328,031 |
model | GitBag/rebel_nectar_1715368613 | 12 | [
"step2201",
"step1201",
"step1401",
"step1801",
"step1001",
"step2001",
"step801",
"step1601",
"step201",
"step601",
"step401",
"main"
] | 1,519 |
model | AwsafAlam/flan-t5-large-code-review-lora-may-12-2024 | 1 | [
"main"
] | 75,550,890 |
model | GoldoBasic/Plory_RVC | 1 | [
"main"
] | 62,227,240 |
model | alexandra-udrescu/qa_toolformer | 1 | [
"main"
] | 134,266,255 |
model | Felladrin/gguf-sharded-Phi-3-mini-4k-instruct-iMat | 1 | [
"main"
] | 2,815,287,511 |
model | saeidebbei/Azkoon | 1 | [
"main"
] | 1,519 |
model | azsxscdvfb/VetMedGPT-1B-V0.2 | 1 | [
"main"
] | 8,038,858,144 |
model | enriquesaou/phi2_mrqa_v2 | 1 | [
"main"
] | 104,894,177 |
model | med-alex/uzn-roberta-base-ft-qa-tr-mt-to-uzn | 1 | [
"main"
] | 352,186,225 |
model | paulo037/stable-code-instruct-3b-spider-13000-steps | 1 | [
"main"
] | 11,183,918,197 |
model | austindavis/gpt2-lichess-uci-201302b | 1 | [
"main"
] | 102,099,077 |
model | NBA55/Final_DPO_All_3_issue_with_Cleaned_Dataset_epoch_2 | 1 | [
"main"
] | 201,360,658 |
model | RichardErkhov/42dot_-_42dot_LLM-SFT-1.3B-4bits | 1 | [
"main"
] | 1,109,709,848 |
model | Meggido/Pantheon-RP-1.0-8b-Llama-3-6.5bpw-h8-exl2 | 1 | [
"main"
] | 7,251,514,066 |
model | CyberHarem/sruler_fgo | 1 | [
"main"
] | 2,277,439,745 |
model | JoshuaAAX/bert-finetuned-ner | 1 | [
"main"
] | 431,897,920 |
model | RichardErkhov/42dot_-_42dot_LLM-SFT-1.3B-8bits | 1 | [
"main"
] | 1,651,222,088 |
model | Zannatul/google-vit-base-patch16-384-in21k-batch_16_epoch_4_classes_24_final_withAug_12th_May | 1 | [
"main"
] | 344,687,276 |
model | abc88767/22c5 | 1 | [
"main"
] | 3,295,853,121 |
model | stafdif/Milf | 1 | [
"main"
] | 19,027,847 |
model | techie4life/llama-2-7b-boolq | 1 | [
"main"
] | 26,956,116,765 |
model | soba1911/xlrs-wav2vec2-lr1e-4 | 1 | [
"main"
] | 3,764,638,114 |
model | Ransss/kukuspice-7B-Q8_0-GGUF | 1 | [
"main"
] | 7,695,860,369 |
model | mberghouse/llama3-8b-python-codes-25k | 1 | [
"main"
] | 16,069,719,273 |
model | BlackRose104/EpiCRealism | 1 | [
"main"
] | 2,134,468,753 |
model | RichardErkhov/NousResearch_-_Hermes-2-Pro-Llama-3-8B-gguf | 1 | [
"main"
] | 99,039,359,917 |
model | sajithrpc/gemma-2b-mt-German-to-English | 1 | [
"main"
] | 5,034,185,800 |
model | akbargherbal/ar_poem_v_01_lora | 1 | [
"main"
] | 221,872,167 |
model | abc88767/3sc4 | 1 | [
"main"
] | 3,295,853,121 |
model | 5w4n/poneyate-xl-v1 | 1 | [
"main"
] | 13,878,866,235 |
model | alexandra-udrescu/qa_toolformer_2 | 1 | [
"main"
] | 67,133,370 |
model | RichardErkhov/BAAI_-_Bunny-v1_0-3B-4bits | 1 | [
"main"
] | 1,941,407,307 |
model | abc88767/5c4 | 1 | [
"main"
] | 3,295,853,120 |
model | cpellens/wavecoder-ultra-6.7b-Q4_K_M-GGUF | 1 | [
"main"
] | 4,081,565,367 |
model | RichardErkhov/BAAI_-_Bunny-v1_0-3B-8bits | 1 | [
"main"
] | 3,049,487,475 |
model | eitan3/infini_llama_3_v2 | 1 | [
"main"
] | 16,069,733,072 |
model | cisnes68/Carl | 1 | [
"main"
] | 1,519 |
model | nnheui/gpt2-sft-index-op-500 | 1 | [
"main"
] | 253,206,785 |
model | Grayx/sad_papa_7 | 1 | [
"main"
] | 16,069,724,863 |
model | rahulmak92/mistral-finetuned-alpaca | 1 | [
"main"
] | 1,519 |
model | parthkher123/Gujarati_Llama_3_Model | 1 | [
"main"
] | 176,971,858 |
model | AI-001/Jellyfishd | 1 | [
"main"
] | 11,098,027 |
model | Bertinho24/Sumin | 1 | [
"main"
] | 77,519,514 |
model | ivykopal/wikiann_sk_adapter_100k | 1 | [
"main"
] | 775,538,706 |
model | Bertinho24/Sieun | 1 | [
"main"
] | 70,840,174 |
model | Zannatul/google-vit-base-patch32-224-in21k-batch_16_epoch_4_classes_24 | 1 | [
"main"
] | 1,519 |
model | Bertinho24/Isa | 1 | [
"main"
] | 80,404,044 |
model | Bertinho24/Seeun | 1 | [
"main"
] | 74,751,625 |
model | cpellens/granite-8b-code-instruct-Q4_K_M-GGUF | 1 | [
"main"
] | 4,882,861,592 |
model | Bertinho24/Yoon | 1 | [
"main"
] | 78,775,739 |
model | santoshtyss/lex-32k-100 | 1 | [
"main"
] | 28,444,991,914 |
model | rohan-2810/image-captioning-blip | 1 | [
"main"
] | 5,638,817,966 |
model | gauravsirola/my_awesome_eli5_clm-model | 1 | [
"main"
] | 327,665,516 |
model | Kikiraw/MOONFILM | 1 | [
"main"
] | 11,076,427,155 |
model | llmsec/Llama-3-8B-sft-lora-ultrachat | 1 | [
"main"
] | 227,284,323 |
model | Bertinho24/J | 1 | [
"main"
] | 80,149,181 |
model | oliverkm/translation | 1 | [
"main"
] | 1,563 |
model | rohan-2810/imagecap_blipbase | 1 | [
"main"
] | 1,981,045,889 |
model | PlayerBPlaytime/MJ | 1 | [
"main"
] | 1,295,311,081 |
model | nbeerbower/llama-3-stinky-v2-8B | 1 | [
"main"
] | 16,069,731,749 |
model | tomaszki/stablelm-57 | 1 | [
"main"
] | 3,295,880,749 |
model | bogya/segformer-b0-scene-parse-150 | 1 | [
"main"
] | 1,519 |
model | omar-sala7/falcon7binstruct_qac_updated_context | 1 | [
"main"
] | 263,894,631 |
model | meherajj/Bangla-Llama-2-7b-base-v0.1 | 1 | [
"main"
] | 13,479,236,859 |
model | Fralet/personality | 1 | [
"main"
] | 6,932,305,260 |
model | Miguelsbdh/llama-3-finetuned-bases-de-datos | 1 | [
"main"
] | 11,477,845,288 |
model | AlignmentResearch/robust_llm_pythia-160m_niki-046_enronspam_random-token-1280_seed-2 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | NBA55/Final_Experiment_with_trained_model_Final_DPO_for_all_3_epoch_2_with_cleaned_dataset | 1 | [
"main"
] | 25,198,968 |
model | Norphel/dzoQA_f2f | 1 | [
"main"
] | 431,832,088 |
model | ernestsborro/erfdse | 1 | [
"main"
] | 1,552 |
model | solidrust/Llama-3-monika-ddlc-8b-v1-AWQ | 1 | [
"main"
] | 5,737,143,613 |
model | mradermacher/Exodius-70B-i1-GGUF | 1 | [
"main"
] | 704,070,265,469 |
model | nev/phi-3-4k-saex-test | 1 | [
"main"
] | 137,798,166,759 |
model | lalreddy/HW_7 | 1 | [
"main"
] | 1,519 |
model | iujinasena/llama-2-7b-conv | 1 | [
"main"
] | 13,476,896,283 |
model | AlignmentResearch/robust_llm_pythia-160m_niki-046_enronspam_random-token-1280_seed-1 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | thrunlab/cats_exp | 1 | [
"main"
] | 28,967,008,061 |
model | xizak/bert-base-uncased-MR | 1 | [
"main"
] | 1,543 |
model | axiomailabs/lora_model_fine_tuned_llama3_8b | 1 | [
"main"
] | 167,835,069 |
model | axiomailabs/lora_model_fine_tuned_llama3_8b_tokenizer | 1 | [
"main"
] | 9,143,476 |
model | MaestraAgah/whisper-large-v3-en | 1 | [
"main"
] | 1,519 |
model | sahlebrahim/codeparrot-ds | 1 | [
"main"
] | 500,348,413 |
model | Felladrin/gguf-sharded-vicuna-160m | 1 | [
"main"
] | 498,963,465 |
model | memoire/llama3_cicis2017_01 | 1 | [
"main"
] | 4,920,735,799 |
model | DewiBrynJones/wav2vec2-xls-r-2b-ft-btb-cy | 1 | [
"main"
] | 8,637,694,226 |
model | zainalikhokhar/llama2_VLLM_LORA | 1 | [
"main"
] | 162,319,417 |
model | fundahe/fundi | 1 | [
"main"
] | 1,519 |
model | zainalikhokhar/llama2_VLLM_4bit | 1 | [
"main"
] | 3,868,394,285 |
model | zainalikhokhar/llama2_VLLM_16bit | 1 | [
"main"
] | 26,431,251,386 |
model | RichardErkhov/SanjiWatsuki_-_Kunoichi-DPO-v2-7B-gguf | 1 | [
"main"
] | 87,511,270,912 |
model | DUAL-GPO/phi-2-gpo-v2-i1 | 1 | [
"main"
] | 171,307,061 |
model | mnoukhov/EleutherAI_pythia-1b-deduped__sft__tldr_55513 | 1 | [
"main"
] | 4,049,276,940 |
model | RichardErkhov/PygmalionAI_-_pygmalion-2-13b-4bits | 1 | [
"main"
] | 7,795,699,829 |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.