type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
sequencelengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model | Holarissun/RM-TLDR_gpt3_loraR64_-1_gemma2b_lr1e-06_bs2_g4 | 1 | [
"main"
] | 51,309,747 |
model | karoldobiczek/roberta-large-fomc_long | 1 | [
"main"
] | 1,424,878,398 |
model | RichardErkhov/google_-_recurrentgemma-2b-it-8bits | 1 | [
"main"
] | 3,387,455,898 |
model | santoshbt/mistral-finetuned-samsum | 1 | [
"main"
] | 1,519 |
model | Saibo-creator/bpe_compressor | 1 | [
"main"
] | 1,044,709 |
model | Holarissun/RM-TLDR_gpt3_loraR64_-1_gemma2b_lr5e-05_bs2_g4 | 1 | [
"main"
] | 51,309,747 |
model | Saibo-creator/bpe_compressor_ar_train | 1 | [
"main"
] | 1,044,709 |
model | Saibo-creator/bpe_compressor_zh_train | 1 | [
"main"
] | 1,038,127 |
model | Saibo-creator/bpe_compressor_de_train | 1 | [
"main"
] | 1,039,197 |
model | domenicrosati/freeze_layers_ten_twenty_meta-llama_Llama-2-7b-chat-hf_minimality-mmd_defence_steps_10000 | 1 | [
"main"
] | 13,476,896,564 |
model | Saibo-creator/bpe_compressor_es_train | 1 | [
"main"
] | 1,038,781 |
model | YorkieOH10/Yi-1.5-34B-Q8_0-GGUF | 1 | [
"main"
] | 36,542,284,598 |
model | Saibo-creator/bpe_compressor_fr_train | 1 | [
"main"
] | 1,039,675 |
model | Saibo-creator/bpe_compressor_ru_train | 1 | [
"main"
] | 1,046,521 |
model | Saibo-creator/bpe_compressor_tr_train | 1 | [
"main"
] | 1,044,709 |
model | emilykang/Phi_medprob-biochemistry_lora | 1 | [
"main"
] | 50,630,613 |
model | maneln/gpttest | 1 | [
"main"
] | 3,099,537,192 |
model | amc9qx/ADRv2024 | 1 | [
"main"
] | 1,628,281,029 |
model | domenicrosati/security_vectors_meta-llama_Llama-2-7b-chat-hf_3e-5_1k | 1 | [
"main"
] | 8,412,876 |
model | emilykang/Gemma_medprob-anatomy_lora | 1 | [
"main"
] | 96,062,130 |
model | LarryAIDraw/Ren_Yamashiro_anylora30r42r-000006 | 1 | [
"main"
] | 37,876,145 |
model | nathantablang/question-answering-qa-may-tablang-LOCAL | 1 | [
"main"
] | 1,337,378,625 |
model | LarryAIDraw/mki-genshin108-1_5-v3 | 1 | [
"main"
] | 110,731,836 |
model | LarryAIDraw/Achreon_SDXL_V1-000014 | 1 | [
"main"
] | 170,553,361 |
model | LarryAIDraw/hoshimi_miyabi | 1 | [
"main"
] | 37,886,552 |
model | mradermacher/Llama-3-8B-Instruct-abliterated-v2-i1-GGUF | 1 | [
"main"
] | 81,008,928,315 |
model | LarryAIDraw/cherno_maid | 1 | [
"main"
] | 456,492,892 |
model | RichardErkhov/beomi_-_KoAlpaca-Polyglot-5.8B-4bits | 1 | [
"main"
] | 3,668,477,585 |
model | rhysjones/Phi-3-mini-4k-mango-2 | 1 | [
"main"
] | 7,644,571,115 |
model | emilykang/Phi_medprob-biochemistry | 1 | [
"main"
] | 5,562,839,047 |
model | Jamessjunk/AnzuFutaba | 1 | [
"main"
] | 83,555,311 |
model | Litzy619/PHI30512HMAB4H | 1 | [
"main"
] | 8,194,835,162 |
model | mradermacher/Padma-SLM-7b-v1.0-GGUF | 1 | [
"main"
] | 73,415,913,811 |
model | ehristoforu/pm-v0.1 | 1 | [
"main"
] | 14,485,322,692 |
model | AlignmentResearch/robust_llm_pythia-1b_niki-052_imdb_gcg_seed-1 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | yaelmd/Llama-3-8b-chat-translation-finetune-10k-1epoch | 1 | [
"main"
] | 16,069,813,160 |
model | flammenai/flammen24X-mistral-7B | 1 | [
"main"
] | 14,485,815,043 |
model | RichardErkhov/beomi_-_KoAlpaca-Polyglot-5.8B-8bits | 1 | [
"main"
] | 6,138,875,998 |
model | maneln/gpt2-finetuned | 1 | [
"main"
] | 1,519 |
model | omparghale/biogpt_ft | 1 | [
"main"
] | 694,959,927 |
model | maneln/gpt-finetuned | 1 | [
"main"
] | 3,099,547,482 |
model | Frixi/Fuwawa_HoloEN | 1 | [
"main"
] | 150,609,468 |
model | Shalie/PenitentOnePonyXL | 1 | [
"main"
] | 111,085,912 |
model | thorirhrafn/llama_DPO_model_e2 | 1 | [
"main"
] | 269,573,685 |
model | kapliff89/roberta-base-finetuned-emotion-aug | 1 | [
"main"
] | 443,518,017 |
model | Su1ka/polki | 1 | [
"main"
] | 228,477,403 |
model | pmolodo/Taxi-v3-qtable | 1 | [
"main"
] | 150,426 |
model | emilykang/Phi_medprob-dental_lora | 1 | [
"main"
] | 50,632,651 |
model | santhoshml/e2erag-arctic-m | 1 | [
"main"
] | 97,882,887 |
model | RichardErkhov/shenzhi-wang_-_Llama3-8B-Chinese-Chat-4bits | 1 | [
"main"
] | 6,037,160,874 |
model | Padlex/results | 1 | [
"main"
] | 355,072,740 |
model | YorkieOH10/Yi-1.5-34B-Chat-Q8_0-GGUF | 1 | [
"main"
] | 36,542,285,096 |
model | bryan-tchakote/fine-tuned-llm | 1 | [
"main"
] | 83,952,725 |
model | nathantablang/question-answering-qa-may-12-tablang-LOCAL | 1 | [
"main"
] | 1,337,376,492 |
model | domenicrosati/security_vectors_meta-llama_Llama-2-7b-chat-hf_8e-5_10k | 1 | [
"main"
] | 8,412,876 |
model | emilykang/Phi_medprob-dental | 1 | [
"main"
] | 5,562,839,047 |
model | 2nzi/videomae-surf-analytics | 1 | [
"main"
] | 4,301,493,771 |
model | ddidiadiandiana/idefics2-8b-docvqa-finetuned-tutorial | 1 | [
"main"
] | 93,387,892 |
model | RichardErkhov/shenzhi-wang_-_Llama3-8B-Chinese-Chat-8bits | 1 | [
"main"
] | 9,096,030,656 |
model | vicgalle/Configurable-Yi-1.5-9B-Chat | 1 | [
"main"
] | 17,663,508,969 |
model | TheTsar1209/llama3-carp-v0.4.1 | 1 | [
"main"
] | 16,069,720,937 |
model | Litzy619/PHI30512HMAB1H | 1 | [
"main"
] | 8,194,835,226 |
model | JuanRetamales/lora_model | 1 | [
"main"
] | 120,104,108 |
model | crusoeai/dolphin-2.9.1-qwen-110b-GGUF | 1 | [
"main"
] | 843,225,871,845 |
model | bartowski/Yi-1.5-9B-Chat-GGUF | 1 | [
"main"
] | 131,323,826,881 |
model | collectionkpro/TBAI | 1 | [
"main"
] | 1,556 |
model | Tpratap04/gpt2_finetuned_tej | 1 | [
"main"
] | 501,149,921 |
model | emilykang/Phi_medprob-medicine_lora | 1 | [
"main"
] | 50,632,494 |
model | terry69/mistral-10p-adv-full | 1 | [
"main"
] | 14,485,819,285 |
model | Rhma/llamaF | 1 | [
"main"
] | 4,829,133,225 |
model | Jonregi/MISTR_INT8 | 1 | [
"main"
] | 1,519 |
model | emilykang/Gemma_medprob-anatomy | 1 | [
"main"
] | 5,029,943,795 |
model | Yumejichi/dqn-IceHockeyNoFrameskip-v4 | 1 | [
"main"
] | 54,784,063 |
model | terry69/llama3-10p-adv-full | 1 | [
"main"
] | 16,069,724,598 |
model | Jonregi/MISTR_INT82 | 1 | [
"main"
] | 1,519 |
model | elijahriley/text_classification_model | 1 | [
"main"
] | 268,805,036 |
model | mradermacher/Llama-3-8B-Instruct-abliterated-i1-GGUF | 1 | [
"main"
] | 81,008,927,460 |
model | emilykang/Phi_medprob-medicine | 1 | [
"main"
] | 5,562,839,047 |
model | sorg20/autotrain-sd-pic | 1 | [
"main"
] | 16,964,633 |
model | abc88767/2c31 | 1 | [
"main"
] | 3,295,853,121 |
model | Scherzando/bert-fashion-mnist | 1 | [
"main"
] | 116,554,479 |
model | RichardErkhov/Qwen_-_Qwen-14B-gguf | 1 | [
"main"
] | 197,187,921,635 |
model | ehristoforu/pm-v0.2 | 1 | [
"main"
] | 14,485,322,401 |
model | RichardErkhov/rinna_-_llama-3-youko-8b-4bits | 1 | [
"main"
] | 6,037,121,013 |
model | Manuel-O/swin-base-patch4-window7-224-in22k-finetuned-CT | 1 | [
"main"
] | 347,515,473 |
model | pavanvicky1998/Group5_NewsChatBot | 1 | [
"main"
] | 1,519 |
model | emilykang/Gemma_medprob-biochemistry_lora | 1 | [
"main"
] | 56,831,292 |
model | Jonregi/MISTR_INT8_2 | 1 | [
"main"
] | 28,969,283,700 |
model | abc88767/3sc31 | 1 | [
"main"
] | 3,295,853,122 |
model | GTsuya/not_enough_milk_pony | 1 | [
"main"
] | 464,615,914 |
model | dtorber/BioNLP-tech-intro-disc-PLOS | 1 | [
"main"
] | 651,033,097 |
model | Dampfinchen/Llama-3-8B-Ultra-Instruct-SaltSprinkle | 1 | [
"main"
] | 16,069,723,588 |
model | nroggendorff/birthday-detector | 1 | [
"main"
] | 268,840,348 |
model | dalishi188/aoliaoli666 | 1 | [
"main"
] | 1,548 |
model | abc88767/4sc31 | 1 | [
"main"
] | 3,295,853,122 |
model | blockblockblock/Yi-1.5-34B-Chat-bpw3.5-exl2 | 1 | [
"main"
] | 36,583,725,232 |
model | emilykang/Phi_medprob-microbiology_lora | 1 | [
"main"
] | 50,629,570 |
model | RichardErkhov/rinna_-_llama-3-youko-8b-8bits | 1 | [
"main"
] | 9,095,990,795 |
model | emilykang/Gemma_medprob-biochemistry | 1 | [
"main"
] | 5,029,943,795 |
model | SD449/bert-finetuned-ner | 1 | [
"main"
] | 1,519 |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.