type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
sequencelengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model | unrahul/dolly-v2-3b-sym_int5 | 1 | [
"main"
] | 2,255,918,027 |
model | andricValdez/bert-base-cased-finetuned-pan24-augmented | 1 | [
"main"
] | 434,163,704 |
model | unrahul/dolly-v2-3b-asym_int5 | 1 | [
"main"
] | 2,338,583,628 |
model | unrahul/dolly-v2-3b-sym_int8 | 1 | [
"main"
] | 3,330,570,827 |
model | unrahul/dolly-v2-3b-nf3 | 1 | [
"main"
] | 1,594,593,222 |
model | unrahul/dolly-v2-3b-nf4 | 1 | [
"main"
] | 1,925,255,622 |
model | unrahul/dolly-v2-3b-fp4 | 1 | [
"main"
] | 1,925,255,622 |
model | unrahul/dolly-v2-3b-fp8 | 1 | [
"main"
] | 3,330,570,822 |
model | unrahul/dolly-v2-3b-fp8_e4m3 | 1 | [
"main"
] | 3,330,570,827 |
model | unrahul/dolly-v2-3b-fp8_e5m2 | 1 | [
"main"
] | 3,330,570,827 |
model | unrahul/dolly-v2-3b-fp16 | 1 | [
"main"
] | 5,552,481,139 |
model | unrahul/dolly-v2-3b-bf16 | 1 | [
"main"
] | 5,552,481,140 |
model | unrahul/dolly-v2-3b-mixed_fp4 | 1 | [
"main"
] | 1,925,255,628 |
model | unrahul/dolly-v2-3b-mixed_fp8 | 1 | [
"main"
] | 3,330,570,828 |
model | unrahul/open_llama_3b_v2-sym_int4 | 1 | [
"main"
] | 2,178,577,275 |
model | unrahul/open_llama_3b_v2-asym_int4 | 1 | [
"main"
] | 2,282,449,276 |
model | unrahul/open_llama_3b_v2-sym_int5 | 1 | [
"main"
] | 2,594,065,275 |
model | unrahul/open_llama_3b_v2-asym_int5 | 1 | [
"main"
] | 2,697,937,276 |
model | unrahul/open_llama_3b_v2-sym_int8 | 1 | [
"main"
] | 3,944,401,275 |
model | unrahul/open_llama_3b_v2-nf3 | 1 | [
"main"
] | 1,763,089,270 |
model | unrahul/open_llama_3b_v2-nf4 | 1 | [
"main"
] | 2,178,577,270 |
model | unrahul/open_llama_3b_v2-fp4 | 1 | [
"main"
] | 2,178,577,270 |
model | unrahul/open_llama_3b_v2-fp8 | 1 | [
"main"
] | 3,944,401,270 |
model | unrahul/open_llama_3b_v2-fp8_e4m3 | 1 | [
"main"
] | 3,944,401,275 |
model | unrahul/open_llama_3b_v2-fp8_e5m2 | 1 | [
"main"
] | 3,944,401,275 |
model | unrahul/open_llama_3b_v2-fp16 | 1 | [
"main"
] | 6,855,446,526 |
model | ikura31/test_merge_and_unload_model | 1 | [
"main"
] | 4,125,695,550 |
model | unrahul/open_llama_3b_v2-bf16 | 1 | [
"main"
] | 6,855,446,591 |
model | unrahul/open_llama_3b_v2-mixed_fp4 | 1 | [
"main"
] | 2,178,577,276 |
model | unrahul/open_llama_3b_v2-mixed_fp8 | 1 | [
"main"
] | 3,944,401,276 |
model | unrahul/gpt-neo-1.3B-sym_int4 | 1 | [
"main"
] | 1,130,352,245 |
model | unrahul/gpt-neo-1.3B-asym_int4 | 1 | [
"main"
] | 1,171,317,430 |
model | unrahul/gpt-neo-1.3B-sym_int5 | 1 | [
"main"
] | 1,294,212,981 |
model | unrahul/gpt-neo-1.3B-asym_int5 | 1 | [
"main"
] | 1,335,178,166 |
model | unrahul/gpt-neo-1.3B-sym_int8 | 1 | [
"main"
] | 1,826,760,373 |
model | AlignmentResearch/robust_llm_pythia-14m_niki-046_enronspam_random-token-1280_seed-2 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | unrahul/gpt-neo-1.3B-nf3 | 1 | [
"main"
] | 966,491,504 |
model | unrahul/gpt-neo-1.3B-nf4 | 1 | [
"main"
] | 1,130,352,240 |
model | unrahul/gpt-neo-1.3B-fp4 | 1 | [
"main"
] | 1,130,352,240 |
model | unrahul/gpt-neo-1.3B-fp8 | 1 | [
"main"
] | 1,826,760,368 |
model | unrahul/gpt-neo-1.3B-fp8_e4m3 | 1 | [
"main"
] | 1,826,760,373 |
model | unrahul/gpt-neo-1.3B-fp8_e5m2 | 1 | [
"main"
] | 1,826,760,373 |
model | AlignmentResearch/robust_llm_pythia-14m_niki-046_enronspam_random-token-1280_seed-1 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | AlignmentResearch/robust_llm_pythia-31m_niki-046_enronspam_random-token-1280_seed-1 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | unrahul/gpt-neo-1.3B-fp16 | 1 | [
"main"
] | 2,840,483,889 |
model | unrahul/gpt-neo-1.3B-bf16 | 1 | [
"main"
] | 2,634,631,022 |
model | unrahul/gpt-neo-1.3B-mixed_fp4 | 1 | [
"main"
] | 1,130,352,246 |
model | AlignmentResearch/robust_llm_pythia-31m_niki-046_enronspam_random-token-1280_seed-0 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | unrahul/gpt-neo-1.3B-mixed_fp8 | 1 | [
"main"
] | 1,826,760,374 |
model | unrahul/deepseek-coder-1.3b-instruct-sym_int4 | 1 | [
"main"
] | 946,264,508 |
model | unrahul/deepseek-coder-1.3b-instruct-asym_int4 | 1 | [
"main"
] | 986,274,237 |
model | AlignmentResearch/robust_llm_pythia-14m_niki-046_enronspam_random-token-1280_seed-0 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | unrahul/deepseek-coder-1.3b-instruct-sym_int5 | 1 | [
"main"
] | 1,106,303,420 |
model | ymoslem/whisper-small-ga2en-v5.2.2 | 1 | [
"main"
] | 968,933,669 |
model | unrahul/deepseek-coder-1.3b-instruct-asym_int5 | 1 | [
"main"
] | 1,146,313,149 |
model | unrahul/deepseek-coder-1.3b-instruct-sym_int8 | 1 | [
"main"
] | 1,626,429,884 |
model | ximcer/Llama3 | 1 | [
"main"
] | 1,546 |
model | unrahul/deepseek-coder-1.3b-instruct-nf3 | 1 | [
"main"
] | 786,225,591 |
model | unrahul/deepseek-coder-1.3b-instruct-nf4 | 1 | [
"main"
] | 946,264,503 |
model | unrahul/deepseek-coder-1.3b-instruct-fp4 | 1 | [
"main"
] | 946,264,503 |
model | trung0209/rumi | 1 | [
"main"
] | 1,519 |
model | unrahul/deepseek-coder-1.3b-instruct-fp8 | 1 | [
"main"
] | 1,626,429,879 |
model | unrahul/deepseek-coder-1.3b-instruct-fp8_e4m3 | 1 | [
"main"
] | 1,626,429,884 |
model | unrahul/deepseek-coder-1.3b-instruct-fp8_e5m2 | 1 | [
"main"
] | 1,626,429,884 |
model | unrahul/deepseek-coder-1.3b-instruct-fp16 | 1 | [
"main"
] | 2,694,401,080 |
model | unrahul/deepseek-coder-1.3b-instruct-bf16 | 1 | [
"main"
] | 2,694,401,081 |
model | unrahul/deepseek-coder-1.3b-instruct-mixed_fp4 | 1 | [
"main"
] | 946,264,509 |
model | unrahul/deepseek-coder-1.3b-instruct-mixed_fp8 | 1 | [
"main"
] | 1,626,429,885 |
model | unrahul/neural-chat-7b-v3-2-sym_int4 | 1 | [
"main"
] | 4,305,148,025 |
model | unrahul/neural-chat-7b-v3-2-asym_int4 | 1 | [
"main"
] | 4,527,347,990 |
model | zhe0/lama3_taide_sloth | 1 | [
"main"
] | 16,237,745,724 |
model | unrahul/neural-chat-7b-v3-2-sym_int5 | 1 | [
"main"
] | 5,193,977,185 |
model | unrahul/neural-chat-7b-v3-2-asym_int5 | 1 | [
"main"
] | 5,416,176,515 |
model | unrahul/neural-chat-7b-v3-2-sym_int8 | 1 | [
"main"
] | 8,082,573,709 |
model | unrahul/neural-chat-7b-v3-2-nf3 | 1 | [
"main"
] | 3,416,348,732 |
model | AlignmentResearch/robust_llm_pythia-31m_niki-046_enronspam_random-token-1280_seed-2 | 31 | [
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model | unrahul/neural-chat-7b-v3-2-nf4 | 1 | [
"main"
] | 4,305,148,020 |
model | amirpoudel/llama2-fine-tuning | 1 | [
"main"
] | 1,519 |
model | unrahul/neural-chat-7b-v3-2-fp4 | 1 | [
"main"
] | 4,305,148,020 |
model | unrahul/neural-chat-7b-v3-2-fp8 | 1 | [
"main"
] | 8,082,573,704 |
model | unrahul/neural-chat-7b-v3-2-fp8_e4m3 | 1 | [
"main"
] | 8,082,573,709 |
model | unrahul/neural-chat-7b-v3-2-fp8_e5m2 | 1 | [
"main"
] | 8,082,573,709 |
model | unrahul/neural-chat-7b-v3-2-fp16 | 1 | [
"main"
] | 14,485,892,451 |
model | unrahul/neural-chat-7b-v3-2-bf16 | 1 | [
"main"
] | 14,485,892,452 |
model | unrahul/neural-chat-7b-v3-2-mixed_fp4 | 1 | [
"main"
] | 4,305,148,026 |
model | parthkher123/Gujarati_LLMA-3 | 1 | [
"main"
] | 1,546 |
model | el-filatova/rag | 1 | [
"main"
] | 4,127,487,585 |
model | unrahul/neural-chat-7b-v3-2-mixed_fp8 | 1 | [
"main"
] | 8,082,573,710 |
model | gokaygokay/llava-llama3-docci | 1 | [
"main"
] | 16,760,765,583 |
model | unrahul/Hermes-2-Pro-Mistral-7B-sym_int4 | 1 | [
"main"
] | 4,305,754,406 |
model | unrahul/Hermes-2-Pro-Mistral-7B-asym_int4 | 1 | [
"main"
] | 4,527,958,467 |
model | unrahul/Hermes-2-Pro-Mistral-7B-sym_int5 | 1 | [
"main"
] | 5,194,599,950 |
model | Loren85/Pina-V2 | 1 | [
"main"
] | 63,523,641 |
model | unrahul/Hermes-2-Pro-Mistral-7B-asym_int5 | 1 | [
"main"
] | 5,416,803,376 |
model | unrahul/Hermes-2-Pro-Mistral-7B-sym_int8 | 1 | [
"main"
] | 8,083,249,722 |
model | unrahul/Hermes-2-Pro-Mistral-7B-nf3 | 1 | [
"main"
] | 3,416,938,729 |
model | unrahul/Hermes-2-Pro-Mistral-7B-nf4 | 1 | [
"main"
] | 4,305,754,401 |
model | unrahul/Hermes-2-Pro-Mistral-7B-fp4 | 1 | [
"main"
] | 4,305,754,401 |
model | unrahul/Hermes-2-Pro-Mistral-7B-fp8 | 1 | [
"main"
] | 8,083,249,717 |
model | unrahul/Hermes-2-Pro-Mistral-7B-fp8_e4m3 | 1 | [
"main"
] | 8,083,249,722 |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.