id
stringlengths 7
117
| author
stringclasses 6
values | sha
null | created_at
unknown | last_modified
null | disabled
null | downloads
int64 0
18.6M
| downloads_all_time
null | gated
bool 1
class | gguf
null | inference
null | likes
int64 0
4.77k
| library_name
stringclasses 36
values | tags
sequencelengths 1
430
| pipeline_tag
stringclasses 32
values | mask_token
null | model_index
null | trending_score
int64 0
132
| architectures
sequencelengths 1
5
⌀ | bos_token_id
int64 -1
256k
⌀ | eos_token_id
int64 -1
256k
⌀ | hidden_act
stringclasses 15
values | hidden_size
int64 1
20.5k
⌀ | initializer_range
float64 0
1
⌀ | intermediate_size
int64 1
98.3k
⌀ | max_position_embeddings
int64 8
1.05M
⌀ | model_type
stringclasses 530
values | num_attention_heads
int64 1
5k
⌀ | num_hidden_layers
int64 -1
8.93k
⌀ | num_key_value_heads
int64 1
160
⌀ | rms_norm_eps
float64 0
7
⌀ | rope_theta
float64 1k
1,000B
⌀ | sliding_window
int64 0
262k
⌀ | tie_word_embeddings
bool 2
classes | torch_dtype
stringclasses 8
values | transformers_version
stringclasses 207
values | use_cache
bool 2
classes | vocab_size
int64 -1
5.03M
⌀ | attention_bias
bool 2
classes | attention_dropout
float64 0
0.5
⌀ | head_dim
int64 2
256
⌀ | mlp_bias
bool 2
classes | pretraining_tp
int64 0
8
⌀ | rope_scaling
dict |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
NexaAIDev/Octopus-v2 | null | null | "2024-03-16T17:54:22Z" | null | null | 968 | null | null | null | null | 857 | transformers | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"function calling",
"on-device language model",
"android",
"conversational",
"en",
"arxiv:2404.19296",
"arxiv:2404.11459",
"arxiv:2404.01744",
"base_model:google/gemma-2b",
"base_model:finetune:google/gemma-2b",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | null | null | 3 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu | 2,048 | 0.02 | 16,384 | 8,192 | gemma | 8 | 18 | 1 | 0.000001 | 10,000 | null | null | bfloat16 | 4.39.0.dev0 | true | 256,022 | false | 0 | 256 | null | null | null |
Qwen/CodeQwen1.5-7B-Chat | null | null | "2024-04-15T07:17:06Z" | null | null | 47,065 | null | null | null | null | 314 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.16609",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | 2 | 2 | silu | 4,096 | 0.02 | 13,440 | 65,536 | qwen2 | 32 | 32 | 4 | 0.00001 | 1,000,000 | 65,536 | false | bfloat16 | 4.39.3 | true | 92,416 | null | 0 | null | null | null | null |
aaditya/Llama3-OpenBioLLM-8B | null | null | "2024-04-20T00:53:57Z" | null | null | 18,253 | null | null | null | null | 156 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"llama-3",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"distillation",
"en",
"arxiv:2305.18290",
"arxiv:2303.13375",
"arxiv:2212.13138",
"arxiv:2305.09617",
"arxiv:2402.07023",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0.dev0 | false | 128,256 | false | 0 | null | null | 1 | null |
shenzhi-wang/Llama3-8B-Chinese-Chat | null | null | "2024-04-21T15:11:45Z" | null | null | 7,593 | null | null | null | null | 645 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-factory",
"orpo",
"conversational",
"en",
"zh",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"doi:10.57967/hf/2316",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0 | true | 128,256 | false | 0 | null | null | 1 | null |
microsoft/Phi-3-mini-4k-instruct | null | null | "2024-04-22T16:18:17Z" | null | null | 1,347,345 | null | null | null | null | 1,070 | transformers | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"nlp",
"code",
"conversational",
"custom_code",
"en",
"fr",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Phi3ForCausalLM"
] | 1 | 32,000 | silu | 3,072 | 0.02 | 8,192 | 4,096 | phi3 | 32 | 32 | 32 | 0.00001 | 10,000 | 2,047 | false | bfloat16 | 4.40.2 | true | 32,064 | false | 0 | null | null | null | null |
MLP-KTLim/llama-3-Korean-Bllossom-8B | null | null | "2024-04-25T10:16:43Z" | null | null | 89,597 | null | null | null | null | 273 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"ko",
"arxiv:2403.10882",
"arxiv:2403.11399",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.43.1 | false | 128,256 | false | 0 | null | false | 1 | null |
01-ai/Yi-1.5-34B-Chat | null | null | "2024-05-10T06:47:21Z" | null | null | 10,909 | null | null | null | null | 248 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:2403.04652",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 7,168 | 0.02 | 20,480 | 4,096 | llama | 56 | 60 | 8 | 0.000001 | 5,000,000 | null | false | bfloat16 | 4.40.0 | false | 64,000 | false | 0 | null | null | 1 | null |
mlabonne/NeuralDaredevil-8B-abliterated | null | null | "2024-05-27T19:33:23Z" | null | null | 19,314 | null | null | null | null | 148 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"dpo",
"conversational",
"dataset:mlabonne/orpo-dpo-mix-40k",
"license:llama3",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | float16 | 4.40.2 | true | 128,256 | false | 0 | null | false | 1 | null |
microsoft/Florence-2-large-ft | null | null | "2024-06-15T00:57:45Z" | null | null | 66,725 | null | null | null | null | 298 | transformers | [
"transformers",
"pytorch",
"florence2",
"text-generation",
"vision",
"image-text-to-text",
"custom_code",
"arxiv:2311.06242",
"license:mit",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 3 | [
"Florence2ForConditionalGeneration"
] | 0 | 2 | null | null | null | null | null | florence2 | null | null | null | null | null | null | null | float16 | 4.41.0.dev0 | null | 51,289 | null | null | null | null | null | null |
AI-MO/NuminaMath-7B-TIR | null | null | "2024-07-04T09:28:53Z" | null | null | 6,769 | null | null | null | null | 318 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"alignment-handbook",
"generated_from_trainer",
"conversational",
"arxiv:2309.17452",
"base_model:deepseek-ai/deepseek-math-7b-base",
"base_model:finetune:deepseek-ai/deepseek-math-7b-base",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 100,000 | 100,001 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 30 | 32 | 0.000001 | 10,000 | null | false | bfloat16 | 4.42.3 | true | 102,400 | false | 0 | null | false | 1 | null |
MaziyarPanahi/calme-2.4-rys-78b | null | null | "2024-08-07T14:38:21Z" | null | null | 484 | null | null | null | null | 42 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"qwen",
"finetune",
"chatml",
"conversational",
"en",
"dataset:MaziyarPanahi/truthy-dpo-v0.1-axolotl",
"dataset:Intel/orca_dpo_pairs",
"base_model:MaziyarPanahi/calme-2.1-rys-78b",
"base_model:finetune:MaziyarPanahi/calme-2.1-rys-78b",
"license:mit",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | null | 151,645 | silu | 8,192 | 0.02 | 29,568 | 32,768 | qwen2 | 64 | 86 | 8 | 0.000001 | 1,000,000 | null | false | float32 | 4.44.0.dev0 | false | 151,646 | null | 0 | null | null | null | null |
ystemsrx/Qwen2-Boundless | null | null | "2024-08-19T06:49:29Z" | null | null | 378 | null | null | null | null | 37 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"Qwen",
"causal-lm",
"fine-tuned",
"ethics",
"Chinese",
"text2text-generation",
"zh",
"dataset:ystemsrx/Bad_Data_Alpaca",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text2text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 1,536 | 0.02 | 8,960 | 32,768 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.42.3 | true | 151,936 | null | 0 | null | null | null | null |
MarinaraSpaghetti/NemoMix-Unleashed-12B | null | null | "2024-08-21T16:42:43Z" | null | null | 10,241 | null | null | null | null | 141 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.0 | true | 131,072 | null | 0 | 128 | null | null | null |
nbeerbower/Lyra-Gutenberg-mistral-nemo-12B | null | null | "2024-08-23T11:13:56Z" | null | null | 176 | null | null | null | null | 18 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"dataset:jondurbin/gutenberg-dpo-v0.1",
"base_model:Sao10K/MN-12B-Lyra-v1",
"base_model:finetune:Sao10K/MN-12B-Lyra-v1",
"license:cc-by-nc-4.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.2 | true | 131,072 | null | 0 | 128 | null | null | null |
speakleash/Bielik-11B-v2.3-Instruct | null | null | "2024-08-30T12:45:27Z" | null | null | 6,926 | null | null | null | null | 22 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"conversational",
"pl",
"arxiv:2005.01643",
"arxiv:2309.11235",
"arxiv:2006.09092",
"arxiv:2402.13228",
"arxiv:2410.18565",
"base_model:speakleash/Bielik-11B-v2",
"base_model:merge:speakleash/Bielik-11B-v2",
"base_model:speakleash/Bielik-11B-v2.0-Instruct",
"base_model:merge:speakleash/Bielik-11B-v2.0-Instruct",
"base_model:speakleash/Bielik-11B-v2.1-Instruct",
"base_model:merge:speakleash/Bielik-11B-v2.1-Instruct",
"base_model:speakleash/Bielik-11B-v2.2-Instruct",
"base_model:merge:speakleash/Bielik-11B-v2.2-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 32,001 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 50 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.44.2 | true | 32,128 | null | 0 | null | null | null | null |
HF1BitLLM/Llama3-8B-1.58-100B-tokens | null | null | "2024-09-10T10:35:45Z" | null | null | 26,971 | null | null | null | null | 145 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:2402.17764",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"8-bit",
"bitnet",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.44.0.dev0 | true | 128,256 | false | 0 | null | false | 1 | null |
ifable/gemma-2-Ifable-9B | null | null | "2024-09-10T11:48:51Z" | null | null | 305 | null | null | null | null | 33 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"dataset:jondurbin/gutenberg-dpo-v0.1",
"license:gemma",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,584 | 0.02 | 14,336 | 8,192 | gemma2 | 16 | 42 | 8 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.43.4 | false | 256,000 | false | 0 | 256 | null | null | null |
Qwen/Qwen2.5-3B | null | null | "2024-09-15T12:17:03Z" | null | null | 97,632 | null | null | null | null | 33 | null | [
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"arxiv:2407.10671",
"license:other",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.40.1 | true | 151,936 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-32B | null | null | "2024-09-15T12:18:33Z" | null | null | 23,072 | null | null | null | null | 28 | null | [
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"arxiv:2407.10671",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 5,120 | 0.02 | 27,648 | 131,072 | qwen2 | 40 | 64 | 8 | 0.00001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
inflatebot/MN-12B-Mag-Mell-R1 | null | null | "2024-09-16T00:38:49Z" | null | null | 154 | null | null | null | null | 18 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"arxiv:2311.03099",
"arxiv:2306.01708",
"base_model:Fizzarolli/MN-12b-Sunrose",
"base_model:merge:Fizzarolli/MN-12b-Sunrose",
"base_model:IntervitensInc/Mistral-Nemo-Base-2407-chatml",
"base_model:merge:IntervitensInc/Mistral-Nemo-Base-2407-chatml",
"base_model:anthracite-org/magnum-v2.5-12b-kto",
"base_model:merge:anthracite-org/magnum-v2.5-12b-kto",
"base_model:elinas/Chronos-Gold-12B-1.0",
"base_model:merge:elinas/Chronos-Gold-12B-1.0",
"base_model:nbeerbower/mistral-nemo-bophades-12B",
"base_model:merge:nbeerbower/mistral-nemo-bophades-12B",
"base_model:nbeerbower/mistral-nemo-gutenberg-12B-v4",
"base_model:merge:nbeerbower/mistral-nemo-gutenberg-12B-v4",
"base_model:nbeerbower/mistral-nemo-wissenschaft-12B",
"base_model:merge:nbeerbower/mistral-nemo-wissenschaft-12B",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 15 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.1 | true | 131,072 | null | 0 | 128 | null | null | null |
Qwen/Qwen2.5-0.5B-Instruct | null | null | "2024-09-16T11:52:46Z" | null | null | 419,043 | null | null | null | null | 94 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-0.5B",
"base_model:finetune:Qwen/Qwen2.5-0.5B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.43.1 | true | 151,936 | null | 0 | null | null | null | null |
thrishala/mental_health_chatbot | null | null | "2024-09-17T14:39:34Z" | null | null | 689 | null | null | null | null | 7 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.38.2 | true | 32,000 | false | 0 | null | null | 1 | null |
unsloth/Mistral-Small-Instruct-2409 | null | null | "2024-09-17T17:48:49Z" | null | null | 1,741 | null | null | null | null | 3 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 6,144 | 0.02 | 16,384 | 131,072 | mistral | 48 | 56 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.2 | true | 32,768 | null | 0 | 128 | null | null | null |
THUDM/cogvlm2-llama3-caption | null | null | "2024-09-18T11:21:11Z" | null | null | 5,095 | null | null | null | null | 55 | transformers | [
"transformers",
"safetensors",
"text-generation",
"video-text-to-text",
"custom_code",
"en",
"arxiv:2408.06072",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"license:other",
"autotrain_compatible",
"region:us"
] | video-text-to-text | null | null | 3 | [
"CogVLMVideoForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 2,048 | null | 32 | 32 | null | 0.00001 | null | null | false | bfloat16 | 4.43.1 | true | 128,256 | null | null | null | null | null | null |
BAAI/Emu3-Chat | null | null | "2024-09-26T04:43:47Z" | null | null | 4,600 | null | null | null | null | 71 | transformers | [
"transformers",
"safetensors",
"Emu3",
"text-generation",
"custom_code",
"arxiv:2409.18869",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Emu3ForCausalLM"
] | 151,849 | 151,850 | silu | 4,096 | 0.02 | 14,336 | 131,072 | Emu3 | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float32 | 4.44.0 | true | 184,622 | null | 0.1 | null | null | 1 | null |
ibm-granite/granite-3.0-2b-instruct | null | null | "2024-10-02T21:07:46Z" | null | null | 15,000 | null | null | null | null | 34 | transformers | [
"transformers",
"safetensors",
"granite",
"text-generation",
"language",
"granite-3.0",
"conversational",
"arxiv:0000.00000",
"base_model:ibm-granite/granite-3.0-2b-base",
"base_model:finetune:ibm-granite/granite-3.0-2b-base",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"GraniteForCausalLM"
] | 0 | 0 | silu | 2,048 | 0.02 | 8,192 | 4,096 | granite | 32 | 40 | 8 | 0.00001 | 10,000 | null | true | bfloat16 | 4.46.0.dev0 | true | 49,155 | false | 0.1 | null | false | null | null |
Steelskull/MSM-MS-Cydrion-22B | null | null | "2024-10-03T12:22:41Z" | null | null | 113 | null | null | null | null | 10 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"conversational",
"base_model:ArliAI/Mistral-Small-22B-ArliAI-RPMax-v1.1",
"base_model:merge:ArliAI/Mistral-Small-22B-ArliAI-RPMax-v1.1",
"base_model:TheDrummer/Cydonia-22B-v1.1",
"base_model:merge:TheDrummer/Cydonia-22B-v1.1",
"base_model:nbeerbower/Mistral-Small-Gutenberg-Doppel-22B",
"base_model:merge:nbeerbower/Mistral-Small-Gutenberg-Doppel-22B",
"base_model:rAIfle/Acolyte-22B",
"base_model:merge:rAIfle/Acolyte-22B",
"base_model:unsloth/Mistral-Small-Instruct-2409",
"base_model:merge:unsloth/Mistral-Small-Instruct-2409",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 6,144 | 0.02 | 16,384 | 131,072 | mistral | 48 | 56 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.43.2 | true | 32,768 | null | 0 | 128 | null | null | null |
AIDC-AI/Ovis1.6-Gemma2-9B-GPTQ-Int4 | null | null | "2024-10-15T12:10:13Z" | null | null | 187 | null | null | null | null | 5 | transformers | [
"transformers",
"ovis",
"text-generation",
"MLLM",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"dataset:AIDC-AI/Ovis-dataset",
"arxiv:2405.20797",
"base_model:AIDC-AI/Ovis1.6-Gemma2-9B",
"base_model:quantized:AIDC-AI/Ovis1.6-Gemma2-9B",
"license:apache-2.0",
"autotrain_compatible",
"4-bit",
"gptq",
"region:us"
] | image-text-to-text | null | null | 3 | [
"Ovis"
] | null | null | null | 3,584 | null | null | null | ovis | null | null | null | null | null | null | null | bfloat16 | 4.44.2 | true | null | null | null | null | null | null | null |
ibm-granite/granite-guardian-3.0-8b | null | null | "2024-10-15T17:49:13Z" | null | null | 1,076 | null | null | null | null | 27 | transformers | [
"transformers",
"safetensors",
"granite",
"text-generation",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"GraniteForCausalLM"
] | 0 | 0 | silu | 4,096 | 0.02 | 12,800 | 8,192 | granite | 32 | 40 | 8 | 0.00001 | 10,000 | null | true | bfloat16 | 4.45.0.dev0 | false | 49,155 | false | 0.1 | null | false | null | null |
AIDC-AI/Ovis1.6-Llama3.2-3B | null | null | "2024-10-16T13:01:34Z" | null | null | 2,360 | null | null | null | null | 35 | transformers | [
"transformers",
"safetensors",
"ovis",
"text-generation",
"MLLM",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"dataset:AIDC-AI/Ovis-dataset",
"arxiv:2405.20797",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 3 | [
"Ovis"
] | null | null | null | 3,072 | null | null | null | ovis | null | null | null | null | null | null | null | bfloat16 | 4.44.2 | false | null | null | null | null | null | null | null |
RefalMachine/ruadapt_qwen2.5_3B_ext_u48_instruct_v4 | null | null | "2024-10-18T10:23:19Z" | null | null | 1,597 | null | null | null | null | 21 | null | [
"safetensors",
"qwen2",
"text-generation",
"conversational",
"ru",
"dataset:IlyaGusev/saiga_scored",
"dataset:IlyaGusev/saiga_preferences",
"dataset:dichspace/darulm",
"base_model:RefalMachine/ruadapt_qwen2.5_3B_ext_u48_full_lr5e4_peft_mlp_32_32_bs256",
"base_model:finetune:RefalMachine/ruadapt_qwen2.5_3B_ext_u48_full_lr5e4_peft_mlp_32_32_bs256",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | null | 147,077 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | bfloat16 | 4.45.2 | true | 147,097 | null | 0 | null | null | null | null |
anthracite-org/magnum-v4-12b | null | null | "2024-10-20T02:06:55Z" | null | null | 806 | null | null | null | null | 21 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"chat",
"conversational",
"en",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.45.0.dev0 | false | 131,072 | null | 0 | 128 | null | null | null |
sarvamai/sarvam-1 | null | null | "2024-10-23T13:18:53Z" | null | null | 1,604 | null | null | null | null | 23 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"bn",
"en",
"gu",
"hi",
"kn",
"ml",
"mr",
"or",
"pa",
"ta",
"te",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 11,008 | 8,192 | llama | 16 | 28 | 8 | 0.000001 | 10,000 | null | false | bfloat16 | 4.45.2 | true | 68,096 | false | 0 | 128 | false | 1 | null |
prithivMLmods/Llama-3.2-1B-GGUF | null | null | "2024-10-26T13:00:49Z" | null | null | 1,339 | null | null | null | null | 5 | transformers | [
"transformers",
"gguf",
"llama",
"text-generation-inference",
"ollama",
"fast-llama",
"1B",
"meta",
"meta-llama",
"token",
"text-generation",
"en",
"base_model:meta-llama/Llama-3.2-1B",
"base_model:quantized:meta-llama/Llama-3.2-1B",
"license:creativeml-openrail-m",
"endpoints_compatible",
"region:us",
"conversational"
] | text-generation | null | null | 3 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
fblgit/cybertron-v4-qw7B-MGS | null | null | "2024-10-29T12:44:49Z" | null | null | 731 | null | null | null | null | 9 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"conversational",
"en",
"dataset:Magpie-Align/Magpie-Qwen2.5-Pro-1M-v0.1",
"arxiv:2410.21228",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-7B-Instruct",
"license:other",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 131,072 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.45.2 | true | 152,064 | null | 0 | null | null | null | null |
facebook/MobileLLM-350M | null | null | "2024-10-30T22:55:48Z" | null | null | 929 | null | null | null | null | 24 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mobilellm",
"text-generation",
"custom_code",
"arxiv:2402.14905",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MobileLLMForCausalLM"
] | 1 | 2 | silu | 960 | 0.02 | 2,560 | 2,048 | mobilellm | 15 | 32 | 5 | 0.00001 | 10,000 | null | false | float16 | 4.41.2 | true | 32,000 | false | 0 | 64 | false | 1 | null |
HuggingFaceTB/SmolLM2-360M | null | null | "2024-10-31T00:43:11Z" | null | null | 3,275 | null | null | null | null | 21 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 0 | 0 | silu | 960 | 0.02 | 2,560 | 8,192 | llama | 15 | 32 | 5 | 0.00001 | 100,000 | null | true | bfloat16 | 4.40.1 | true | 49,152 | false | 0 | null | null | 1 | null |
EVA-UNIT-01/EVA-Qwen2.5-32B-v0.1 | null | null | "2024-10-31T05:13:25Z" | null | null | 259 | null | null | null | null | 14 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"conversational",
"dataset:anthracite-org/kalo-opus-instruct-22k-no-refusal",
"dataset:Nopm/Opus_WritingStruct",
"dataset:Gryphe/Sonnet3.5-SlimOrcaDedupCleaned",
"dataset:Gryphe/Sonnet3.5-Charcard-Roleplay",
"dataset:Gryphe/ChatGPT-4o-Writing-Prompts",
"dataset:Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned",
"dataset:Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned",
"dataset:nothingiisreal/Reddit-Dirty-And-WritingPrompts",
"dataset:allura-org/Celeste-1.x-data-mixture",
"dataset:cognitivecomputations/dolphin-2.9.3",
"base_model:Qwen/Qwen2.5-32B",
"base_model:finetune:Qwen/Qwen2.5-32B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | null | 151,643 | silu | 5,120 | 0.02 | 27,648 | 131,072 | qwen2 | 40 | 64 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.45.2 | false | 152,064 | null | 0 | null | null | null | null |
amd/AMD-OLMo-1B-SFT-DPO | null | null | "2024-10-31T20:29:17Z" | null | null | 883 | null | null | null | null | 15 | transformers | [
"transformers",
"safetensors",
"olmo",
"text-generation",
"conversational",
"dataset:allenai/dolma",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"OlmoForCausalLM"
] | null | 50,279 | silu | 2,048 | 0.02 | 8,192 | 2,048 | olmo | 16 | 16 | 16 | null | 10,000 | null | true | float32 | 4.40.2 | true | 50,304 | false | 0 | null | null | null | null |
THU-KEG/Mistral-Crab-SFT | null | null | "2024-11-01T06:35:13Z" | null | null | 28 | null | null | null | null | 4 | null | [
"safetensors",
"mistral",
"alignment-handbook",
"generated_from_trainer",
"text-generation",
"conversational",
"en",
"arxiv:2410.24175",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.0 | true | 32,768 | null | 0 | 128 | null | null | null |
fblgit/miniclaus-qw1.5B-UNAMGS | null | null | "2024-11-01T07:07:02Z" | null | null | 4,865 | null | null | null | null | 4 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"conversational",
"en",
"dataset:Magpie-Align/Magpie-Pro-MT-300K-v0.1",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-1.5B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-1.5B-Instruct",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 1,536 | 0.02 | 8,960 | 131,072 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | null | true | bfloat16 | 4.45.2 | true | 151,936 | null | 0 | null | null | null | null |
MaziyarPanahi/SmolLM2-1.7B-Instruct-GGUF | null | null | "2024-11-01T16:40:28Z" | null | null | 635,330 | null | null | null | null | 3 | null | [
"gguf",
"mistral",
"quantized",
"2-bit",
"3-bit",
"4-bit",
"5-bit",
"6-bit",
"8-bit",
"GGUF",
"text-generation",
"base_model:HuggingFaceTB/SmolLM2-1.7B-Instruct",
"base_model:quantized:HuggingFaceTB/SmolLM2-1.7B-Instruct",
"region:us",
"imatrix",
"conversational"
] | text-generation | null | null | 3 | null | null | null | null | null | null | null | null | mistral | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
allura-org/TQ2.5-14B-Sugarquill-v1 | null | null | "2024-11-06T10:39:16Z" | null | null | 51 | null | null | null | null | 3 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"dataset:Mielikki/Erebus-87k",
"dataset:allura-org/r_shortstories_24k",
"base_model:arcee-ai/SuperNova-Medius",
"base_model:finetune:arcee-ai/SuperNova-Medius",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | null | 151,645 | silu | 5,120 | 0.02 | 13,824 | 131,072 | qwen2 | 40 | 48 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.45.1 | false | 152,064 | null | 0 | null | null | null | null |
FallenMerick/MN-Chunky-Lotus-12B | null | null | "2024-11-07T23:42:28Z" | null | null | 33 | null | null | null | null | 3 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"storywriting",
"text adventure",
"creative",
"story",
"writing",
"fiction",
"roleplaying",
"rp",
"mergekit",
"merge",
"en",
"arxiv:2306.01708",
"base_model:Epiculous/Violet_Twilight-v0.2",
"base_model:merge:Epiculous/Violet_Twilight-v0.2",
"base_model:TheDrummer/Rocinante-12B-v1.1",
"base_model:merge:TheDrummer/Rocinante-12B-v1.1",
"base_model:flammenai/Mahou-1.5-mistral-nemo-12B",
"base_model:merge:flammenai/Mahou-1.5-mistral-nemo-12B",
"base_model:nbeerbower/mistral-nemo-gutenberg-12B-v4",
"base_model:merge:nbeerbower/mistral-nemo-gutenberg-12B-v4",
"license:cc-by-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 131,072 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.0 | true | 131,072 | null | 0 | 128 | null | null | null |
theo77186/Qwen2.5-Coder-7B-Instruct-20241106 | null | null | "2024-11-08T10:00:12Z" | null | null | 6 | null | null | null | null | 3 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"conversational",
"en",
"arxiv:2409.12186",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-Coder-7B",
"base_model:finetune:Qwen/Qwen2.5-Coder-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
EleutherAI/gpt-neo-2.7B | null | null | "2022-03-02T23:29:04Z" | null | null | 191,834 | null | null | null | null | 438 | transformers | [
"transformers",
"pytorch",
"jax",
"rust",
"safetensors",
"gpt_neo",
"text-generation",
"text generation",
"causal-lm",
"en",
"dataset:EleutherAI/pile",
"arxiv:2101.00027",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"GPTNeoForCausalLM"
] | 50,256 | 50,256 | null | 2,560 | 0.02 | null | 2,048 | gpt_neo | null | null | null | null | null | null | null | null | 4.9.0.dev0 | true | 50,257 | null | 0 | null | null | null | null |
Gustavosta/MagicPrompt-Stable-Diffusion | null | null | "2022-09-17T22:34:07Z" | null | null | 321,164 | null | null | null | null | 704 | transformers | [
"transformers",
"pytorch",
"coreml",
"safetensors",
"gpt2",
"text-generation",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.23.0.dev0 | true | 50,257 | null | null | null | null | null | null |
bigscience/bloomz-560m | null | null | "2022-10-08T16:14:42Z" | null | null | 15,791,704 | null | null | null | null | 107 | transformers | [
"transformers",
"pytorch",
"tensorboard",
"safetensors",
"bloom",
"text-generation",
"ak",
"ar",
"as",
"bm",
"bn",
"ca",
"code",
"en",
"es",
"eu",
"fon",
"fr",
"gu",
"hi",
"id",
"ig",
"ki",
"kn",
"lg",
"ln",
"ml",
"mr",
"ne",
"nso",
"ny",
"or",
"pa",
"pt",
"rn",
"rw",
"sn",
"st",
"sw",
"ta",
"te",
"tn",
"ts",
"tum",
"tw",
"ur",
"vi",
"wo",
"xh",
"yo",
"zh",
"zu",
"dataset:bigscience/xP3",
"arxiv:2211.01786",
"license:bigscience-bloom-rail-1.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"BloomForCausalLM"
] | 1 | 2 | null | null | 0.02 | null | null | bloom | 16 | null | null | null | null | null | null | null | 4.20.0 | true | 250,880 | null | 0 | null | null | 1 | null |
PygmalionAI/pygmalion-6b | null | null | "2023-01-07T18:43:33Z" | null | null | 3,219 | null | null | null | null | 730 | transformers | [
"transformers",
"pytorch",
"tensorboard",
"gptj",
"text-generation",
"text generation",
"conversational",
"en",
"license:creativeml-openrail-m",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"GPTJForCausalLM"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gptj | null | null | null | null | null | null | false | float16 | 4.26.0.dev0 | true | 50,400 | null | null | null | null | null | null |
medalpaca/medalpaca-7b | null | null | "2023-03-29T17:54:49Z" | null | null | 4,771 | null | null | null | null | 71 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"medical",
"en",
"arxiv:2303.14070",
"license:cc",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 0 | 1 | silu | 4,096 | 0.02 | 11,008 | null | llama | 32 | 32 | null | 0.000001 | null | null | false | float32 | 4.28.0.dev0 | true | 32,001 | null | null | null | null | null | null |
tiiuae/falcon-7b-instruct | null | null | "2023-04-25T06:21:01Z" | null | null | 169,796 | null | null | null | null | 919 | transformers | [
"transformers",
"pytorch",
"coreml",
"safetensors",
"falcon",
"text-generation",
"conversational",
"custom_code",
"en",
"dataset:tiiuae/falcon-refinedweb",
"arxiv:2205.14135",
"arxiv:1911.02150",
"arxiv:2005.14165",
"arxiv:2104.09864",
"arxiv:2306.01116",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"FalconForCausalLM"
] | 11 | 11 | null | 4,544 | 0.02 | null | null | falcon | 71 | 32 | null | null | null | null | null | bfloat16 | 4.27.4 | true | 65,024 | null | 0 | null | null | null | null |
replit/replit-code-v1-3b | null | null | "2023-04-28T19:00:45Z" | null | null | 379 | null | null | null | null | 722 | transformers | [
"transformers",
"pytorch",
"mpt",
"text-generation",
"code",
"custom_code",
"dataset:bigcode/the-stack-dedup",
"arxiv:2211.15533",
"arxiv:2205.14135",
"arxiv:2108.12409",
"arxiv:2302.06675",
"license:cc-by-sa-4.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MPTForCausalLM"
] | null | 1 | null | null | null | null | null | mpt | null | null | null | null | null | null | null | float32 | 4.28.1 | true | 32,768 | null | null | null | null | null | null |
alibidaran/medical_transcription_generator | null | null | "2023-06-03T10:16:09Z" | null | null | 961 | null | null | null | null | 4 | transformers | [
"transformers",
"pytorch",
"safetensors",
"gpt2",
"text-generation",
"medical",
"en",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"doi:10.57967/hf/1204",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.29.2 | true | 50,257 | null | null | null | null | null | null |
TheBloke/Llama-2-7B-Chat-GPTQ | null | null | "2023-07-18T17:38:15Z" | null | null | 136,936 | null | null | null | null | 261 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-2",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"base_model:quantized:meta-llama/Llama-2-7b-chat-hf",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | null | 0.000001 | null | null | false | float16 | 4.30.0.dev0 | true | 32,000 | null | null | null | null | null | null |
lmsys/vicuna-7b-v1.5 | null | null | "2023-07-29T04:42:33Z" | null | null | 389,147 | null | null | null | null | 304 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"arxiv:2307.09288",
"arxiv:2306.05685",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | null | null | false | float16 | 4.31.0 | true | 32,000 | null | null | null | null | 1 | null |
TheBloke/Llama-2-7B-Chat-GGUF | null | null | "2023-09-04T16:38:41Z" | null | null | 90,148 | null | null | null | null | 437 | transformers | [
"transformers",
"gguf",
"llama",
"facebook",
"meta",
"pytorch",
"llama-2",
"text-generation",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"base_model:quantized:meta-llama/Llama-2-7b-chat-hf",
"license:llama2",
"region:us"
] | text-generation | null | null | 2 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
EleutherAI/llemma_7b | null | null | "2023-09-12T22:09:33Z" | null | null | 7,883 | null | null | null | null | 98 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"math",
"reasoning",
"en",
"dataset:EleutherAI/proof-pile-2",
"dataset:open-web-math/open-web-math",
"arxiv:2310.10631",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | null | null | false | bfloat16 | 4.31.0 | true | 32,016 | null | null | null | null | 1 | null |
thesephist/contra-bottleneck-t5-xl-wikipedia | null | null | "2023-09-30T21:48:13Z" | null | null | 179 | null | null | null | null | 16 | transformers | [
"transformers",
"pytorch",
"t5",
"text-generation",
"custom_code",
"en",
"dataset:wikipedia",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"BottleneckT5LMWithPerturb"
] | null | 1 | null | null | null | null | null | t5 | null | null | null | null | null | null | false | float32 | 4.33.3 | true | 32,128 | null | null | null | null | null | null |
liuhaotian/llava-v1.5-13b | null | null | "2023-10-05T18:27:40Z" | null | null | 464,536 | null | null | null | null | 481 | transformers | [
"transformers",
"pytorch",
"llava",
"text-generation",
"image-text-to-text",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 2 | [
"LlavaLlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llava | 40 | 40 | 40 | 0.00001 | null | null | false | float16 | 4.31.0 | true | 32,000 | null | null | null | null | 1 | null |
HuggingFaceH4/zephyr-7b-alpha | null | null | "2023-10-09T08:45:10Z" | null | null | 34,183 | null | null | null | null | 1,099 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"conversational",
"en",
"dataset:stingning/ultrachat",
"dataset:openbmb/UltraFeedback",
"arxiv:2305.18290",
"arxiv:2310.16944",
"arxiv:2305.14233",
"arxiv:2310.01377",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.34.0 | true | 32,000 | null | null | null | null | null | null |
TheBloke/CausalLM-14B-GGUF | null | null | "2023-10-22T16:28:29Z" | null | null | 2,983 | null | null | null | null | 151 | transformers | [
"transformers",
"gguf",
"llama",
"llama2",
"qwen",
"text-generation",
"en",
"zh",
"dataset:JosephusCheung/GuanacoDataset",
"dataset:Open-Orca/OpenOrca",
"dataset:stingning/ultrachat",
"dataset:meta-math/MetaMathQA",
"dataset:liuhaotian/LLaVA-Instruct-150K",
"dataset:jondurbin/airoboros-3.1",
"dataset:WizardLM/WizardLM_evol_instruct_V2_196k",
"dataset:RyokoAI/ShareGPT52K",
"dataset:RyokoAI/Fandom23K",
"dataset:milashkaarshif/MoeGirlPedia_wikitext_raw_archive",
"dataset:wikipedia",
"dataset:wiki_lingua",
"dataset:fnlp/moss-003-sft-data",
"dataset:garage-bAInd/Open-Platypus",
"dataset:LDJnr/Puffin",
"dataset:openbmb/llava_zh",
"dataset:BAAI/COIG",
"dataset:TigerResearch/tigerbot-zhihu-zh-10k",
"dataset:liwu/MNBVC",
"dataset:teknium/openhermes",
"base_model:CausalLM/14B",
"base_model:quantized:CausalLM/14B",
"license:wtfpl",
"region:us"
] | text-generation | null | null | 2 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
01-ai/Yi-34B | null | null | "2023-11-01T07:03:50Z" | null | null | 5,505 | null | null | null | null | 1,285 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"arxiv:2403.04652",
"arxiv:2311.16502",
"arxiv:2401.11944",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 7,168 | 0.02 | 20,480 | 4,096 | llama | 56 | 60 | 8 | 0.00001 | 5,000,000 | null | false | bfloat16 | 4.34.0 | true | 64,000 | null | null | null | null | 1 | null |
maritaca-ai/sabia-7b | null | null | "2023-11-08T22:13:36Z" | null | null | 1,104 | null | null | null | null | 104 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"pt",
"arxiv:2304.07880",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 2,048 | llama | 32 | 32 | 32 | 0.000001 | 10,000 | null | false | bfloat16 | 4.37.2 | true | 32,000 | false | 0 | null | null | 1 | null |
reglab-rrc/mistral-rrc | null | null | "2023-11-23T06:42:21Z" | null | null | 89 | null | null | null | null | 4 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"legal",
"housing",
"covenants",
"property",
"deed",
"racial-covenant",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float16 | 4.34.0 | true | 32,000 | null | null | null | null | null | null |
bigcode/starcoder2-3b | null | null | "2023-11-29T15:22:51Z" | null | null | 189,981 | null | null | null | null | 146 | transformers | [
"transformers",
"safetensors",
"starcoder2",
"text-generation",
"code",
"dataset:bigcode/the-stack-v2-train",
"arxiv:2305.13245",
"arxiv:2205.14135",
"arxiv:2004.05150",
"arxiv:2207.14255",
"arxiv:2402.19173",
"license:bigcode-openrail-m",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"Starcoder2ForCausalLM"
] | 0 | 0 | gelu_pytorch_tanh | 3,072 | 0.018042 | 12,288 | 16,384 | starcoder2 | 24 | 30 | 2 | null | 999,999.442036 | 4,096 | null | null | 4.37.0.dev0 | true | 49,152 | null | 0.1 | null | null | null | null |
TheBloke/Mistral-7B-Instruct-v0.2-GGUF | null | null | "2023-12-11T22:18:46Z" | null | null | 131,340 | null | null | null | null | 398 | transformers | [
"transformers",
"gguf",
"mistral",
"finetuned",
"text-generation",
"arxiv:2310.06825",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"region:us",
"conversational"
] | text-generation | null | null | 2 | null | null | null | null | null | null | null | null | mistral | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
upstage/SOLAR-10.7B-Instruct-v1.0 | null | null | "2023-12-12T12:39:22Z" | null | null | 130,262 | null | null | null | null | 614 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"dataset:c-s-ale/alpaca-gpt4-data",
"dataset:Open-Orca/OpenOrca",
"dataset:Intel/orca_dpo_pairs",
"dataset:allenai/ultrafeedback_binarized_cleaned",
"arxiv:2312.15166",
"arxiv:2403.19270",
"base_model:upstage/SOLAR-10.7B-v1.0",
"base_model:finetune:upstage/SOLAR-10.7B-v1.0",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 4,096 | llama | 32 | 48 | 8 | 0.00001 | 10,000 | null | false | float16 | 4.35.0 | true | 32,000 | false | null | null | null | 1 | null |
cognitivecomputations/dolphin-2.5-mixtral-8x7b | null | null | "2023-12-14T00:45:47Z" | null | null | 23,529 | null | null | null | null | 1,208 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"en",
"dataset:ehartford/dolphin",
"dataset:jondurbin/airoboros-2.2.1",
"dataset:ehartford/dolphin-coder",
"dataset:migtissera/Synthia-v1.3",
"dataset:teknium/openhermes",
"dataset:ise-uiuc/Magicoder-OSS-Instruct-75K",
"dataset:ise-uiuc/Magicoder-Evol-Instruct-110K",
"dataset:LDJnr/Pure-Dove",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MixtralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.0.dev0 | false | 32,002 | null | 0 | null | null | null | null |
cognitivecomputations/dolphin-2.6-mixtral-8x7b | null | null | "2023-12-21T22:34:09Z" | null | null | 4,920 | null | null | null | null | 195 | transformers | [
"transformers",
"pytorch",
"mixtral",
"text-generation",
"conversational",
"en",
"dataset:ehartford/dolphin",
"dataset:jondurbin/airoboros-2.2.1",
"dataset:ehartford/dolphin-coder",
"dataset:teknium/openhermes",
"dataset:ise-uiuc/Magicoder-OSS-Instruct-75K",
"dataset:ise-uiuc/Magicoder-Evol-Instruct-110K",
"dataset:LDJnr/Capybara",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MixtralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.37.0.dev0 | false | 32,002 | null | 0 | null | null | null | null |
MaralGPT/Maral-7B-alpha-1 | null | null | "2023-12-24T17:10:55Z" | null | null | 616 | null | null | null | null | 50 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"en",
"fa",
"dataset:sinarashidi/alpaca-persian",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.36.2 | true | 32,000 | null | 0 | null | null | null | null |
Vikhrmodels/Vikhr-7b-0.1 | null | null | "2024-01-02T11:14:26Z" | null | null | 280 | null | null | null | null | 53 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"ru",
"en",
"dataset:IlyaGusev/habr",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float16 | 4.37.2 | true | 40,000 | false | 0 | null | null | 1 | null |
NousResearch/Genstruct-7B | null | null | "2024-01-05T21:56:01Z" | null | null | 231 | null | null | null | null | 367 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"Mistral",
"instruct",
"finetune",
"synthetic",
"conversational",
"en",
"arxiv:2310.04484",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.36.2 | false | 32,001 | null | 0 | null | null | null | null |
TencentARC/LLaMA-Pro-8B-Instruct | null | null | "2024-01-06T02:12:26Z" | null | null | 1,005 | null | null | null | null | 62 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 40 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.33.3 | false | 32,000 | null | null | null | null | 1 | null |
liuhaotian/llava-v1.6-mistral-7b | null | null | "2024-01-31T04:20:00Z" | null | null | 56,875 | null | null | null | null | 225 | transformers | [
"transformers",
"safetensors",
"llava_mistral",
"text-generation",
"image-text-to-text",
"conversational",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 2 | [
"LlavaMistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | llava_mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.2 | true | 32,000 | null | 0 | null | null | null | null |
Himitsui/Kaiju-11B | null | null | "2024-02-13T12:33:38Z" | null | null | 109 | null | null | null | null | 14 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"en",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 4,096 | llama | 32 | 48 | 8 | 0.00001 | 10,000 | null | false | float16 | 4.37.2 | true | 32,000 | false | 0 | null | null | 1 | null |
prometheus-eval/prometheus-7b-v2.0 | null | null | "2024-02-13T17:18:13Z" | null | null | 6,567 | null | null | null | null | 77 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"text2text-generation",
"en",
"dataset:prometheus-eval/Feedback-Collection",
"dataset:prometheus-eval/Preference-Collection",
"arxiv:2405.01535",
"arxiv:2310.08491",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text2text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.35.2 | true | 32,000 | null | 0 | null | null | null | null |
NousResearch/Nous-Hermes-2-Mistral-7B-DPO | null | null | "2024-02-18T10:50:26Z" | null | null | 4,005 | null | null | null | null | 168 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"Mistral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.37.0.dev0 | false | 32,002 | null | 0 | null | null | null | null |
abideen/gemma-7b-openhermes | null | null | "2024-02-21T23:03:54Z" | null | null | 11 | null | null | null | null | 11 | transformers | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"generated_from_trainer",
"axolotl",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:mlabonne/chatml-OpenHermes2.5-dpo-binarized-alpha",
"base_model:google/gemma-7b-it",
"base_model:finetune:google/gemma-7b-it",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu | 3,072 | 0.02 | 24,576 | 8,192 | gemma | 16 | 28 | 16 | 0.000001 | 10,000 | null | null | float16 | 4.39.0.dev0 | true | 256,000 | false | 0 | 256 | null | null | null |
ND911/Fraken-Maid-TW-K-Slerp | null | null | "2024-02-26T00:02:51Z" | null | null | 14 | null | null | null | null | 2 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"SanjiWatsuki/Kunoichi-7B",
"ND911/Fraken-Maid-TW-Slerp",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 8,192 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.37.2 | true | 32,000 | null | 0 | null | null | null | null |
1bitLLM/bitnet_b1_58-large | null | null | "2024-03-29T11:10:00Z" | null | null | 19,392 | null | null | null | null | 72 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:2402.17764",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"BitnetForCausalLM"
] | 1 | 2 | silu | 1,536 | 0.02 | 4,096 | 2,048 | llama | 16 | 24 | 16 | 0.00001 | 10,000 | null | true | float16 | 4.39.0 | true | 32,002 | false | 0 | null | null | 1 | null |
Chat2DB/Chat2DB-SQL-7B | null | null | "2024-04-01T07:25:03Z" | null | null | 697 | null | null | null | null | 38 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"zh",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 16,384 | llama | 32 | 32 | 32 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.33.2 | true | 32,016 | null | null | null | null | 1 | null |
apple/OpenELM-270M | null | null | "2024-04-12T21:42:49Z" | null | null | 1,708 | null | null | null | null | 70 | transformers | [
"transformers",
"safetensors",
"openelm",
"text-generation",
"custom_code",
"arxiv:2404.14619",
"license:other",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"OpenELMForCausalLM"
] | 1 | 2 | null | null | 0.02 | null | null | openelm | null | null | null | null | null | null | null | float32 | 4.39.3 | true | 32,000 | null | null | 64 | null | null | null |
allenai/OLMo-7B-0424 | null | null | "2024-04-15T21:03:34Z" | null | null | 137 | null | null | null | null | 45 | transformers | [
"transformers",
"pytorch",
"olmo",
"text-generation",
"en",
"dataset:allenai/dolma",
"arxiv:2402.00838",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"OLMoForCausalLM"
] | null | 50,279 | null | null | null | null | null | olmo | null | null | null | null | null | null | null | null | 4.39.3 | true | 50,280 | null | 0 | null | null | null | null |
dreamgen/WizardLM-2-7B | null | null | "2024-04-16T05:08:27Z" | null | null | 14,472 | null | null | null | null | 37 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"arxiv:2304.12244",
"arxiv:2306.08568",
"arxiv:2308.09583",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.36.2 | false | 32,000 | null | 0 | null | null | null | null |
lmms-lab/LLaVA-NeXT-Video-7B | null | null | "2024-04-16T14:08:19Z" | null | null | 447 | null | null | null | null | 41 | transformers | [
"transformers",
"safetensors",
"llava",
"text-generation",
"video-text-to-text",
"license:llama2",
"autotrain_compatible",
"region:us"
] | video-text-to-text | null | null | 2 | [
"LlavaLlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llava | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.39.0.dev0 | false | 32,000 | false | 0 | null | null | 1 | null |
NousResearch/Meta-Llama-3-8B-Instruct | null | null | "2024-04-18T16:55:56Z" | null | null | 62,025 | null | null | null | null | 84 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"conversational",
"en",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0.dev0 | true | 128,256 | false | 0 | null | null | 1 | null |
sethuiyer/Medichat-Llama3-8B | null | null | "2024-04-22T05:07:13Z" | null | null | 95,455 | null | null | null | null | 18 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"medical",
"conversational",
"en",
"dataset:mlabonne/orpo-dpo-mix-40k",
"dataset:Open-Orca/SlimOrca-Dedup",
"dataset:jondurbin/airoboros-3.2",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:m-a-p/Code-Feedback",
"dataset:MaziyarPanahi/WizardLM_evol_instruct_V2_196k",
"dataset:ruslanmv/ai-medical-chatbot",
"base_model:Locutusque/llama-3-neural-chat-v1-8b",
"base_model:merge:Locutusque/llama-3-neural-chat-v1-8b",
"base_model:Undi95/Llama-3-Unholy-8B",
"base_model:merge:Undi95/Llama-3-Unholy-8B",
"base_model:ruslanmv/Medical-Llama3-8B",
"base_model:merge:ruslanmv/Medical-Llama3-8B",
"license:other",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0 | true | 128,256 | false | 0 | null | null | 1 | null |
beomi/Llama-3-Open-Ko-8B | null | null | "2024-04-22T09:53:13Z" | null | null | 50,458 | null | null | null | null | 130 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"llama-3-ko",
"conversational",
"en",
"ko",
"arxiv:2310.04799",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.1 | true | 128,256 | false | 0 | null | null | 1 | null |
Orenguteng/Llama-3-8B-Lexi-Uncensored | null | null | "2024-04-23T21:14:40Z" | null | null | 501,212 | null | null | null | null | 170 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"uncensored",
"llama3",
"instruct",
"open",
"conversational",
"license:llama3",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.39.3 | true | 128,256 | false | 0 | null | null | 1 | null |
aaditya/Llama3-OpenBioLLM-70B | null | null | "2024-04-24T02:21:12Z" | null | null | 7,055 | null | null | null | null | 349 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"llama-3",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"distillation",
"heathcare",
"medical",
"clinical",
"med",
"lifescience",
"Pharmaceutical",
"Pharma",
"conversational",
"en",
"arxiv:2305.18290",
"arxiv:2303.13375",
"arxiv:2212.13138",
"arxiv:2305.09617",
"arxiv:2402.07023",
"base_model:meta-llama/Meta-Llama-3-70B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-70B-Instruct",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 8,192 | 0.02 | 28,672 | 8,192 | llama | 64 | 80 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0.dev0 | false | 128,256 | false | 0 | null | null | 1 | null |
teddylee777/Llama-3-Open-Ko-8B-gguf | null | null | "2024-04-24T17:14:54Z" | null | null | 3,848 | null | null | null | null | 38 | transformers | [
"transformers",
"gguf",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"llama-3-ko",
"conversational",
"en",
"ko",
"arxiv:2310.04799",
"license:llama3",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.1 | true | 128,256 | false | 0 | null | null | 1 | null |
HPAI-BSC/Llama3-Aloe-8B-Alpha | null | null | "2024-04-26T16:43:36Z" | null | null | 4,670 | null | null | null | null | 52 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"biology",
"medical",
"question-answering",
"en",
"dataset:argilla/dpo-mix-7k",
"dataset:nvidia/HelpSteer",
"dataset:jondurbin/airoboros-3.2",
"dataset:hkust-nlp/deita-10k-v0",
"dataset:LDJnr/Capybara",
"dataset:HPAI-BSC/CareQA",
"dataset:GBaker/MedQA-USMLE-4-options",
"dataset:lukaemon/mmlu",
"dataset:bigbio/pubmed_qa",
"dataset:openlifescienceai/medmcqa",
"dataset:bigbio/med_qa",
"dataset:HPAI-BSC/better-safe-than-sorry",
"dataset:HPAI-BSC/pubmedqa-cot",
"dataset:HPAI-BSC/medmcqa-cot",
"dataset:HPAI-BSC/medqa-cot",
"arxiv:2405.01886",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | question-answering | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0.dev0 | false | 128,256 | false | 0 | null | null | 1 | null |
nvidia/Llama3-ChatQA-1.5-8B | null | null | "2024-04-28T21:32:50Z" | null | null | 14,764 | null | null | null | null | 545 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"nvidia",
"chatqa-1.5",
"chatqa",
"llama-3",
"pytorch",
"conversational",
"en",
"arxiv:2401.10225",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | float16 | 4.40.0.dev0 | true | 128,256 | false | 0 | null | null | 1 | null |
failspy/llama-3-70B-Instruct-abliterated | null | null | "2024-05-07T00:50:02Z" | null | null | 3,396 | null | null | null | null | 94 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 8,192 | 0.02 | 28,672 | 8,192 | llama | 64 | 80 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.1 | true | 128,256 | false | 0 | null | null | 1 | null |
defog/llama-3-sqlcoder-8b | null | null | "2024-05-09T15:58:49Z" | null | null | 143,101 | null | null | null | null | 191 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"code",
"conversational",
"license:cc-by-sa-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0 | true | 128,256 | false | 0 | null | null | 1 | null |
Miwa-Keita/zenz-v1 | null | null | "2024-05-12T15:48:46Z" | null | null | 18 | null | null | null | null | 7 | transformers | [
"transformers",
"gguf",
"gpt2",
"text-generation",
"japanese input",
"kana kanji conversion",
"ja",
"license:cc-by-sa-4.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"GPT2LMHeadModel"
] | 1 | 2 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.30.0 | true | 6,000 | null | null | null | null | null | null |
Henrychur/MMed-Llama-3-8B | null | null | "2024-05-22T09:17:24Z" | null | null | 1,353 | null | null | null | null | 20 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"medical",
"conversational",
"en",
"zh",
"ja",
"fr",
"ru",
"es",
"dataset:Henrychur/MMedC",
"arxiv:2402.13963",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | float32 | 4.39.2 | true | 128,256 | false | 0 | null | null | 1 | null |
Qwen/Qwen2-72B-Instruct | null | null | "2024-05-28T03:48:49Z" | null | null | 54,116 | null | null | null | null | 675 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"base_model:Qwen/Qwen2-72B",
"base_model:finetune:Qwen/Qwen2-72B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 8,192 | 0.02 | 29,568 | 32,768 | qwen2 | 64 | 80 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.40.1 | true | 152,064 | null | 0 | null | null | null | null |
failspy/Meta-Llama-3-70B-Instruct-abliterated-v3.5 | null | null | "2024-05-28T17:51:28Z" | null | null | 3,786 | null | null | null | null | 37 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 8,192 | 0.02 | 28,672 | 8,192 | llama | 64 | 80 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0.dev0 | true | 128,256 | false | 0 | null | null | 1 | null |
numind/NuExtract | null | null | "2024-05-31T09:53:13Z" | null | null | 2,091 | null | null | null | null | 203 | transformers | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"conversational",
"custom_code",
"en",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:finetune:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 2 | [
"Phi3ForCausalLM"
] | 1 | 32,000 | silu | 3,072 | 0.02 | 8,192 | 4,096 | phi3 | 32 | 32 | 32 | 0.00001 | 10,000 | 2,047 | false | float32 | 4.42.0.dev0 | true | 32,064 | null | 0 | null | null | null | null |