type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
sequencelengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model | EGORsGOR/vit-spam | 1 | [
"main"
] | 343,242,409 |
model | kaya-kedi/Bubble-TADC | 1 | [
"main"
] | 54,960,071 |
model | Minbyul/llama2-7b-wo-kqa_golden-iter-dpo-step2 | 1 | [
"main"
] | 13,479,247,039 |
model | ccihvun/zziclo | 1 | [
"main"
] | 1,519 |
model | blockblockblock/Dark-Miqu-70B-bpw6-exl2 | 1 | [
"main"
] | 52,013,010,661 |
model | jidlso/RVC-models | 1 | [
"main"
] | 215,983,495 |
model | copper029/videomae-base-finetuned-ucf101-subset | 1 | [
"main"
] | 1,519 |
model | AlignmentResearch/robust_llm_pythia-160m_niki-046_enronspam_random-token-1280_seed-0 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | chujiezheng/tulu-2-dpo-13b | 1 | [
"main"
] | 26,034,153,661 |
model | rhye/man | 1 | [
"main"
] | 5,482,649,875 |
model | wikiray/llama | 1 | [
"main"
] | 1,546 |
model | theskrtnerd/corgi-dog | 1 | [
"main"
] | 5,482,649,438 |
model | AlignmentResearch/robust_llm_pythia-70m_niki-047_wl_random-token-1280_seed-1 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | AlignmentResearch/robust_llm_pythia-70m_niki-047_wl_random-token-1280_seed-0 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | nickrwu/deberta-v3-large-finetuned-mathqa | 1 | [
"main"
] | 1,519 |
model | ke-lly/46401345_1 | 1 | [
"main"
] | 1,519 |
model | CarlRD/bloom-7b1-lora-tagger | 1 | [
"main"
] | 1,519 |
model | DUAL-GPO/zephyr-7b-gpo-v4-i2 | 1 | [
"main"
] | 673,548,959 |
model | Thangaraj/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,688 |
model | ABF54/testingdulu | 1 | [
"main"
] | 1,548 |
model | tayyibsupercool/llama-2-7b-miniguanaco-qlora | 1 | [
"main"
] | 16,826,198 |
model | AdityaXPV/Llama-3-8b-Instruct-law-sage-v0.1-GGUF | 1 | [
"main"
] | 37,653,351,430 |
model | Mohyuddin1973/Puppy | 1 | [
"main"
] | 1,561 |
model | Caiyun-AI/DCPythia-6.9B | 1 | [
"main"
] | 14,809,074,522 |
model | SubhasishSaha/Resume-Classifier | 1 | [
"main"
] | 438,995,324 |
model | RichardErkhov/YeungNLP_-_firefly-mixtral-8x7b-gguf | 1 | [
"main"
] | 246,760,208,286 |
model | snshrivas10/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,688 |
model | bhassi01/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,688 |
model | sunjupskilling/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,943 |
model | praveenku32k/zephyr-LLM-chatbot-General | 1 | [
"main"
] | 29,585,772 |
model | Yosingh1/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,943 |
model | SachinGenAIMaster/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,943 |
model | PraveenCMR/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,943 |
model | ittailup/whisper-small-hi | 1 | [
"main"
] | 1,519 |
model | Tristan/pythia-70m_default | 1 | [
"main"
] | 283,812,633 |
model | ydavidpan/bert-base-uncased-finetuned-kaggle-arena-0510 | 1 | [
"main"
] | 1,519 |
model | SaravanaPriyan/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,688 |
model | ivilson/llama-3-8b-chinese-function-calling-AWQ-4bit | 1 | [
"main"
] | 5,737,141,885 |
model | med-alex/kaz-roberta-base-ft-qa-en-mt-to-kaz | 1 | [
"main"
] | 368,017,203 |
model | theskrtnerd/cocacola-drink | 1 | [
"main"
] | 5,482,649,460 |
model | Ayush-1722/Llama-3-8B-Instruct-v0.1-32K-QLoRA-Merged | 1 | [
"main"
] | 16,069,789,399 |
model | Kearm/file.pt | 1 | [
"main"
] | 1,519 |
model | smorodinovmv/models1 | 1 | [
"main"
] | 1,519 |
model | aariz120/zephyr-support-chatbot | 1 | [
"main"
] | 29,592,464 |
model | arctic126/lab_h2o_35 | 1 | [
"main"
] | 3,664,812,902 |
model | leimu/44 | 1 | [
"main"
] | 228,455,987 |
model | npedrazzini/BERTislav | 1 | [
"main"
] | 1,432,810,330 |
model | jujusosmart/bert-base-chinese-Guilt | 1 | [
"main"
] | 409,107,863 |
model | Luna-Skywalker/phi-3-finetuned-lora | 1 | [
"main"
] | 120,104,109 |
model | blockblockblock/neo_7b-bpw2.25-exl2 | 1 | [
"main"
] | 2,634,083,599 |
model | desi3/ppo-LunarLander-v2 | 1 | [
"main"
] | 428,244 |
model | omark03/llama2-qlora-finetunined-med-convo-v1 | 1 | [
"main"
] | 1,519 |
model | vinitlondhe21/zephyr-support-chatbot | 1 | [
"main"
] | 29,585,688 |
model | leimu/45 | 1 | [
"main"
] | 228,458,107 |
model | Jackie999/mistral-10p-adv | 1 | [
"main"
] | 32,101,344 |
model | Aspik101/llama1111 | 1 | [
"main"
] | 32,130,281,223 |
model | AlignmentResearch/robust_llm_pythia-410m_niki-047_wl_random-token-1280_seed-2 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | SimplCup/OlivisRodrigo | 1 | [
"main"
] | 215,944,159 |
model | AlignmentResearch/robust_llm_pythia-70m_niki-047_wl_random-token-1280_seed-2 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | AlignmentResearch/robust_llm_pythia-160m_niki-047_wl_random-token-1280_seed-0 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | rhye/cat_ppl_less | 1 | [
"main"
] | 15,797,585,024 |
model | mmnga/Meta-Llama-3-8B-Instruct-gguf | 1 | [
"main"
] | 99,819,606,841 |
model | sachi020991/my_awesome_model | 1 | [
"main"
] | 268,801,869 |
model | LordNoah/schedule-iter0-b | 1 | [
"main"
] | 28,968,791,095 |
model | thorirhrafn/GPT1B_domar_RLHF_e3 | 1 | [
"main"
] | 50,345,320 |
model | fine-tuned/jina-embeddings-v2-base-en-5122024-3toh-webapp | 1 | [
"main"
] | 550,554,550 |
model | Mr-Bhaskar/fbt-gemma-7b-inst | 1 | [
"main"
] | 17,093,038,793 |
model | acmench/testcopy | 1 | [
"main"
] | 311,838,132 |
model | blockblockblock/neo_7b-bpw2.5-exl2 | 1 | [
"main"
] | 2,862,572,846 |
model | MrOvkill/Llama-3-8B-sft-lora-ultrachat | 1 | [
"main"
] | 1,519 |
model | AlignmentResearch/robust_llm_pythia-410m_niki-047_wl_random-token-1280_seed-0 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | sally9805/bert-base-uncased-finetuned-news-1908-1909 | 1 | [
"main"
] | 439,048,284 |
model | AlignmentResearch/robust_llm_pythia-1b_niki-047_wl_random-token-1280_seed-1 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | techsivam/zephyr-support-chatbot | 1 | [
"main"
] | 1,519 |
model | yaojingguo/codeparrot-ds | 1 | [
"main"
] | 1,519 |
model | med-alex/kaz-roberta-base-ft-qa-tr-mt-to-kaz | 1 | [
"main"
] | 368,348,009 |
model | krishnakalyan3/emo_nmlaozqe_67 | 1 | [
"main"
] | 382,272,147 |
model | DerekTrayn15/theolasad | 1 | [
"main"
] | 97,432,764 |
model | Rohandebnath243536287/Llama-2-7b-chat-finetune | 1 | [
"main"
] | 13,479,328,407 |
model | Demon1221/Temp | 1 | [
"main"
] | 1,547 |
model | win10/phi3-128k-6b | 1 | [
"main"
] | 11,721,743,675 |
model | voxmenthe/Meta-Llama-3-70B-Instruct-8bit | 1 | [
"main"
] | 159,603,261,670 |
model | clxudiajazmin/summarization-TFM-Claudia | 1 | [
"main"
] | 245,315,332 |
model | PB7-DUT-2023/finetuned_Bloomz_1b1_v6 | 1 | [
"main"
] | 2,145,231,559 |
model | nguyenx/vietnamese-correction-v2 | 1 | [
"main"
] | 1,588,918,422 |
model | TrgTuan10/Famelink_hand_negative | 1 | [
"main"
] | 28,077 |
model | jiangcongtao/llama3-8b-unsloth-merged | 1 | [
"main"
] | 16,069,789,678 |
model | SuiGio/roberta_pubmesh | 1 | [
"main"
] | 331,930,175 |
model | blockblockblock/neo_7b-bpw3-exl2 | 1 | [
"main"
] | 3,323,677,998 |
model | Mag0g/Ezekiel25_15 | 1 | [
"main"
] | 3,295,853,092 |
model | Hyx66/LandscaperPriv | 1 | [
"main"
] | 2,607,440,397 |
model | cwei13/bert-base-japanese-ghost_rate | 1 | [
"main"
] | 442,892,896 |
model | OwOpeepeepoopoo/test12 | 1 | [
"main"
] | 3,295,853,400 |
model | S4nto/lora-dpo-finetuned-stage2-iter40000-sft | 1 | [
"main"
] | 23,066,479,892 |
model | jiangcongtao/llama3-8b-unsloth | 1 | [
"main"
] | 167,835,070 |
model | enchan1/q-FrozenLake-v1-4x4-noSlippery | 1 | [
"main"
] | 34,548 |
model | Mudiit4/gemma-Code-aws-Finetune-test | 1 | [
"main"
] | 5,034,185,180 |
model | jin-cheon/bert-finetuned-ner | 1 | [
"main"
] | 862,797,421 |
model | Mr-Bhaskar/fbt-mistral7b-instruct | 1 | [
"main"
] | 14,485,390,390 |
model | netcat420/MFANNv0.9 | 1 | [
"main"
] | 32,130,241,100 |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.