id
stringlengths 7
117
| author
stringclasses 6
values | sha
null | created_at
unknown | last_modified
null | disabled
null | downloads
int64 0
18.6M
| downloads_all_time
null | gated
bool 1
class | gguf
null | inference
null | likes
int64 0
4.77k
| library_name
stringclasses 36
values | tags
sequencelengths 1
430
| pipeline_tag
stringclasses 32
values | mask_token
null | model_index
null | trending_score
int64 0
132
| architectures
sequencelengths 1
5
⌀ | bos_token_id
int64 -1
256k
⌀ | eos_token_id
int64 -1
256k
⌀ | hidden_act
stringclasses 15
values | hidden_size
int64 1
20.5k
⌀ | initializer_range
float64 0
1
⌀ | intermediate_size
int64 1
98.3k
⌀ | max_position_embeddings
int64 8
1.05M
⌀ | model_type
stringclasses 530
values | num_attention_heads
int64 1
5k
⌀ | num_hidden_layers
int64 -1
8.93k
⌀ | num_key_value_heads
int64 1
160
⌀ | rms_norm_eps
float64 0
7
⌀ | rope_theta
float64 1k
1,000B
⌀ | sliding_window
int64 0
262k
⌀ | tie_word_embeddings
bool 2
classes | torch_dtype
stringclasses 8
values | transformers_version
stringclasses 207
values | use_cache
bool 2
classes | vocab_size
int64 -1
5.03M
⌀ | attention_bias
bool 2
classes | attention_dropout
float64 0
0.5
⌀ | head_dim
int64 2
256
⌀ | mlp_bias
bool 2
classes | pretraining_tp
int64 0
8
⌀ | rope_scaling
dict |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HuggingFaceTB/SmolLM2-1.7B-Instruct | null | null | "2024-10-31T13:42:06Z" | null | null | 43,720 | null | null | null | null | 307 | transformers | [
"transformers",
"tensorboard",
"onnx",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 132 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 8,192 | 8,192 | llama | 32 | 24 | 32 | 0.00001 | 130,000 | null | true | bfloat16 | 4.42.3 | true | 49,152 | false | 0 | null | false | 1 | null |
infly/OpenCoder-8B-Instruct | null | null | "2024-11-07T16:23:14Z" | null | null | 402 | null | null | null | null | 77 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"zh",
"dataset:OpenCoder-LLM/opencoder-sft-stage1",
"dataset:OpenCoder-LLM/opencoder-sft-stage2",
"arxiv:2411.04905",
"base_model:infly/OpenCoder-8B-Base",
"base_model:finetune:infly/OpenCoder-8B-Base",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 77 | [
"LlamaForCausalLM"
] | 96,540 | 96,539 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.37.0 | true | 96,640 | false | 0 | null | false | 1 | null |
Qwen/Qwen2.5-Coder-7B-Instruct | null | null | "2024-09-17T13:38:49Z" | null | null | 72,981 | null | null | null | null | 245 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"conversational",
"en",
"arxiv:2409.12186",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-Coder-7B",
"base_model:finetune:Qwen/Qwen2.5-Coder-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 54 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.44.0 | true | 152,064 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-72B-Instruct | null | null | "2024-09-16T11:56:31Z" | null | null | 416,529 | null | null | null | null | 452 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-72B",
"base_model:finetune:Qwen/Qwen2.5-72B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 32 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 8,192 | 0.02 | 29,568 | 32,768 | qwen2 | 64 | 80 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
facebook/MobileLLM-125M | null | null | "2024-10-30T22:48:34Z" | null | null | 4,281 | null | null | null | null | 80 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mobilellm",
"text-generation",
"custom_code",
"arxiv:2402.14905",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 29 | [
"MobileLLMForCausalLM"
] | 1 | 2 | silu | 576 | 0.02 | 1,536 | 2,048 | mobilellm | 9 | 30 | 3 | 0.00001 | 10,000 | null | false | float16 | 4.41.2 | true | 32,000 | false | 0 | 64 | false | 1 | null |
facebook/MobileLLM-1B | null | null | "2024-10-31T00:07:47Z" | null | null | 6,405 | null | null | null | null | 102 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mobilellm",
"text-generation",
"custom_code",
"arxiv:2402.14905",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 25 | [
"MobileLLMForCausalLM"
] | 1 | 2 | silu | 1,280 | 0.02 | 3,584 | 2,048 | mobilellm | 20 | 54 | 5 | 0.00001 | 10,000 | null | false | float16 | 4.41.2 | true | 32,000 | false | 0 | 64 | false | 1 | null |
infly/OpenCoder-1.5B-Instruct | null | null | "2024-11-07T16:22:28Z" | null | null | 186 | null | null | null | null | 21 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"zh",
"dataset:OpenCoder-LLM/opencoder-sft-stage1",
"dataset:OpenCoder-LLM/opencoder-sft-stage2",
"arxiv:2411.04905",
"base_model:infly/OpenCoder-1.5B-Base",
"base_model:finetune:infly/OpenCoder-1.5B-Base",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 21 | [
"LlamaForCausalLM"
] | 96,540 | 96,539 | silu | 2,240 | 0.02 | 6,144 | 4,096 | llama | 14 | 24 | 14 | 0.00001 | 10,000 | null | false | bfloat16 | 4.37.0 | true | 96,640 | false | 0 | null | false | 1 | null |
vikhyatk/moondream2 | null | null | "2024-03-04T18:03:06Z" | null | null | 197,554 | null | null | null | null | 693 | transformers | [
"transformers",
"safetensors",
"gguf",
"moondream1",
"text-generation",
"image-text-to-text",
"custom_code",
"doi:10.57967/hf/3219",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | image-text-to-text | null | null | 19 | [
"Moondream"
] | null | null | null | null | null | null | null | moondream1 | null | null | null | null | null | null | null | float16 | 4.44.0 | null | null | null | null | null | null | null | null |
Qwen/Qwen2.5-7B-Instruct | null | null | "2024-09-16T11:55:40Z" | null | null | 484,324 | null | null | null | null | 258 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-7B",
"base_model:finetune:Qwen/Qwen2.5-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 19 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
infly/OpenCoder-8B-Base | null | null | "2024-11-07T16:20:01Z" | null | null | 391 | null | null | null | null | 17 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"zh",
"arxiv:2411.04905",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 17 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.44.2 | true | 96,640 | false | 0 | null | false | 1 | null |
microsoft/Florence-2-large | null | null | "2024-06-15T00:34:55Z" | null | null | 1,658,548 | null | null | null | null | 1,207 | transformers | [
"transformers",
"pytorch",
"florence2",
"text-generation",
"vision",
"image-text-to-text",
"custom_code",
"arxiv:2311.06242",
"license:mit",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 16 | [
"Florence2ForConditionalGeneration"
] | 0 | 2 | null | null | null | null | null | florence2 | null | null | null | null | null | null | null | float16 | 4.41.0.dev0 | null | 51,289 | null | null | null | null | null | null |
arcee-ai/SuperNova-Medius | null | null | "2024-10-02T06:50:01Z" | null | null | 13,213 | null | null | null | null | 167 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:Qwen/Qwen2.5-14B",
"base_model:finetune:Qwen/Qwen2.5-14B",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 16 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 5,120 | 0.02 | 13,824 | 131,072 | qwen2 | 40 | 48 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.0.dev0 | true | 152,064 | null | 0 | null | null | null | null |
ibm-granite/granite-3.0-8b-instruct | null | null | "2024-10-02T21:16:23Z" | null | null | 29,211 | null | null | null | null | 155 | transformers | [
"transformers",
"safetensors",
"granite",
"text-generation",
"language",
"granite-3.0",
"conversational",
"arxiv:0000.00000",
"base_model:ibm-granite/granite-3.0-8b-base",
"base_model:finetune:ibm-granite/granite-3.0-8b-base",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 15 | [
"GraniteForCausalLM"
] | 0 | 0 | silu | 4,096 | 0.02 | 12,800 | 4,096 | granite | 32 | 40 | 8 | 0.00001 | 10,000 | null | true | bfloat16 | 4.46.0.dev0 | true | 49,155 | false | 0.1 | null | false | null | null |
rhymes-ai/Aria | null | null | "2024-09-26T02:58:52Z" | null | null | 26,676 | null | null | null | null | 581 | transformers | [
"transformers",
"safetensors",
"aria",
"text-generation",
"multimodal",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"arxiv:2410.05993",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 14 | [
"AriaForConditionalGeneration"
] | null | null | null | null | null | null | null | aria | null | null | null | null | null | null | null | bfloat16 | 4.45.0 | null | null | null | null | null | null | null | null |
HuggingFaceTB/SmolLM2-1.7B | null | null | "2024-10-30T22:50:10Z" | null | null | 7,262 | null | null | null | null | 63 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 14 | [
"LlamaForCausalLM"
] | 0 | 0 | silu | 2,048 | 0.02 | 8,192 | 8,192 | llama | 32 | 24 | 32 | 0.00001 | 130,000 | null | true | bfloat16 | 4.40.1 | true | 49,152 | false | 0 | null | null | 1 | null |
EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2 | null | null | "2024-11-05T05:36:22Z" | null | null | 425 | null | null | null | null | 13 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"conversational",
"dataset:anthracite-org/kalo-opus-instruct-22k-no-refusal",
"dataset:Nopm/Opus_WritingStruct",
"dataset:Gryphe/Sonnet3.5-SlimOrcaDedupCleaned",
"dataset:Gryphe/Sonnet3.5-Charcard-Roleplay",
"dataset:Gryphe/ChatGPT-4o-Writing-Prompts",
"dataset:Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned",
"dataset:Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned",
"dataset:nothingiisreal/Reddit-Dirty-And-WritingPrompts",
"dataset:allura-org/Celeste-1.x-data-mixture",
"dataset:cognitivecomputations/dolphin-2.9.3",
"base_model:Qwen/Qwen2.5-32B",
"base_model:finetune:Qwen/Qwen2.5-32B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 13 | [
"Qwen2ForCausalLM"
] | null | 151,643 | silu | 5,120 | 0.02 | 27,648 | 131,072 | qwen2 | 40 | 64 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.45.1 | false | 152,064 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-1.5B-Instruct | null | null | "2024-09-17T14:10:29Z" | null | null | 14,955,186 | null | null | null | null | 128 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-1.5B",
"base_model:finetune:Qwen/Qwen2.5-1.5B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 12 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 1,536 | 0.02 | 8,960 | 32,768 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.43.1 | true | 151,936 | null | 0 | null | null | null | null |
HuggingFaceTB/SmolLM2-135M-Instruct | null | null | "2024-10-31T13:41:10Z" | null | null | 9,151 | null | null | null | null | 53 | transformers | [
"transformers",
"tensorboard",
"onnx",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 12 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 576 | 0.041667 | 1,536 | 8,192 | llama | 9 | 30 | 3 | 0.00001 | 100,000 | null | true | bfloat16 | 4.42.3 | true | 49,152 | false | 0 | null | false | 1 | null |
jadechoghari/Ferret-UI-Llama8b | null | null | "2024-10-09T16:32:15Z" | null | null | 748 | null | null | null | null | 42 | transformers | [
"transformers",
"safetensors",
"ferret_llama",
"text-generation",
"image-text-to-text",
"conversational",
"custom_code",
"arxiv:2404.05719",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 11 | [
"FerretLlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | ferret_llama | null | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.39.0 | true | 128,258 | false | 0 | null | null | 1 | null |
BAAI/Aquila-VL-2B-llava-qwen | null | null | "2024-10-17T09:50:06Z" | null | null | 1,442 | null | null | null | null | 42 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"multimodal",
"image-text-to-text",
"conversational",
"en",
"zh",
"dataset:BAAI/Infinity-MM",
"dataset:BAAI/Infinity-Instruct",
"dataset:BAAI/Infinity-Preference",
"arxiv:2410.18558",
"base_model:Qwen/Qwen2.5-1.5B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-1.5B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | image-text-to-text | null | null | 11 | [
"LlavaQwenForCausalLM"
] | 151,643 | 151,645 | silu | 1,536 | 0.02 | 8,960 | 32,768 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.40.0 | true | 151,936 | null | 0 | null | null | null | null |
openai-community/gpt2 | null | null | "2022-03-02T23:29:04Z" | null | null | 16,737,409 | null | null | null | null | 2,354 | transformers | [
"transformers",
"pytorch",
"tf",
"jax",
"tflite",
"rust",
"onnx",
"safetensors",
"gpt2",
"text-generation",
"exbert",
"en",
"doi:10.57967/hf/0039",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 10 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | null | null | null | 50,257 | null | null | null | null | null | null |
liuhaotian/llava-v1.5-7b | null | null | "2023-10-05T18:25:51Z" | null | null | 793,191 | null | null | null | null | 363 | transformers | [
"transformers",
"pytorch",
"llava",
"text-generation",
"image-text-to-text",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 10 | [
"LlavaLlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llava | 32 | 32 | 32 | 0.00001 | null | null | false | float16 | 4.31.0 | true | 32,000 | null | null | null | null | 1 | null |
infly/OpenCoder-1.5B-Base | null | null | "2024-11-07T16:19:26Z" | null | null | 218 | null | null | null | null | 10 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"zh",
"arxiv:2411.04905",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 10 | [
"LlamaForCausalLM"
] | 1 | 96,539 | silu | 2,240 | 0.02 | 6,144 | 4,096 | llama | 14 | 24 | 14 | 0.00001 | 10,000 | null | false | bfloat16 | 4.44.2 | true | 96,640 | false | 0 | null | false | 1 | null |
Qwen/Qwen2.5-7B | null | null | "2024-09-15T12:17:40Z" | null | null | 52,542 | null | null | null | null | 64 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"arxiv:2407.10671",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 9 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 131,072 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.40.1 | true | 152,064 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-3B-Instruct | null | null | "2024-09-17T14:08:52Z" | null | null | 214,629 | null | null | null | null | 78 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-3B",
"base_model:finetune:Qwen/Qwen2.5-3B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 9 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.43.1 | true | 151,936 | null | 0 | null | null | null | null |
jadechoghari/Ferret-UI-Gemma2b | null | null | "2024-10-09T16:19:17Z" | null | null | 1,885 | null | null | null | null | 44 | transformers | [
"transformers",
"safetensors",
"ferret_gemma",
"text-generation",
"image-text-to-text",
"conversational",
"custom_code",
"arxiv:2404.05719",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 9 | [
"FerretGemmaForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 2,048 | 0.02 | 16,384 | 8,192 | ferret_gemma | 8 | 18 | 1 | 0.000001 | 10,000 | null | null | bfloat16 | 4.39.0 | true | 256,001 | false | 0 | 256 | null | null | null |
HuggingFaceTB/SmolLM2-360M-Instruct | null | null | "2024-10-31T13:41:35Z" | null | null | 11,297 | null | null | null | null | 43 | transformers | [
"transformers",
"tensorboard",
"onnx",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 9 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 960 | 0.02 | 2,560 | 8,192 | llama | 15 | 32 | 5 | 0.00001 | 100,000 | null | true | bfloat16 | 4.42.3 | true | 49,152 | false | 0 | null | false | 1 | null |
bigscience/bloom | null | null | "2022-05-19T11:53:33Z" | null | null | 10,169 | null | null | null | null | 4,766 | transformers | [
"transformers",
"pytorch",
"tensorboard",
"safetensors",
"bloom",
"text-generation",
"ak",
"ar",
"as",
"bm",
"bn",
"ca",
"code",
"en",
"es",
"eu",
"fon",
"fr",
"gu",
"hi",
"id",
"ig",
"ki",
"kn",
"lg",
"ln",
"ml",
"mr",
"ne",
"nso",
"ny",
"or",
"pa",
"pt",
"rn",
"rw",
"sn",
"st",
"sw",
"ta",
"te",
"tn",
"ts",
"tum",
"tw",
"ur",
"vi",
"wo",
"xh",
"yo",
"zh",
"zu",
"arxiv:2211.05100",
"arxiv:1909.08053",
"arxiv:2110.02861",
"arxiv:2108.12409",
"doi:10.57967/hf/0003",
"license:bigscience-bloom-rail-1.0",
"model-index",
"co2_eq_emissions",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 8 | [
"BloomForCausalLM"
] | 1 | 2 | null | null | 0.02 | null | null | bloom | 112 | null | null | null | null | null | null | null | 4.21.0 | true | 250,880 | null | 0 | null | null | 4 | null |
vikhyatk/moondream1 | null | null | "2024-01-20T18:10:04Z" | null | null | 298,246 | null | null | null | null | 484 | transformers | [
"transformers",
"pytorch",
"safetensors",
"moondream1",
"text-generation",
"custom_code",
"en",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 8 | [
"Moondream"
] | null | null | null | null | null | null | null | moondream1 | null | null | null | null | null | null | null | float16 | 4.36.2 | null | null | null | null | null | null | null | null |
dunzhang/stella_en_1.5B_v5 | null | null | "2024-07-12T15:52:09Z" | null | null | 86,749 | null | null | null | null | 145 | sentence-transformers | [
"sentence-transformers",
"pytorch",
"safetensors",
"qwen2",
"text-generation",
"mteb",
"transformers",
"sentence-similarity",
"custom_code",
"arxiv:2205.13147",
"license:mit",
"model-index",
"autotrain_compatible",
"text-embeddings-inference",
"endpoints_compatible",
"region:us"
] | sentence-similarity | null | null | 8 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 1,536 | 0.02 | 8,960 | 131,072 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | 131,072 | false | float32 | 4.42.3 | true | 151,646 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-Coder-7B | null | null | "2024-09-16T11:57:24Z" | null | null | 27,426 | null | null | null | null | 49 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"code",
"qwen",
"qwen-coder",
"codeqwen",
"conversational",
"en",
"arxiv:2409.12186",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-7B",
"base_model:finetune:Qwen/Qwen2.5-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 8 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.45.0.dev0 | true | 152,064 | null | 0 | null | null | null | null |
dfurman/CalmeRys-78B-Orpo-v0.1 | null | null | "2024-09-24T10:25:46Z" | null | null | 5,324 | null | null | null | null | 40 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"orpo",
"sft",
"chatml",
"conversational",
"en",
"dataset:mlabonne/orpo-dpo-mix-40k",
"base_model:MaziyarPanahi/calme-2.4-rys-78b",
"base_model:finetune:MaziyarPanahi/calme-2.4-rys-78b",
"license:mit",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | null | null | 8 | [
"Qwen2ForCausalLM"
] | 151,644 | 151,645 | silu | 8,192 | 0.02 | 29,568 | 32,768 | qwen2 | 64 | 86 | 8 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.44.2 | false | 151,646 | null | 0 | null | null | null | null |
rombodawg/Rombos-LLM-V2.6-Qwen-14b | null | null | "2024-10-12T20:19:50Z" | null | null | 4,214 | null | null | null | null | 41 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"base_model:Qwen/Qwen2.5-14B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-14B-Instruct",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 8 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 5,120 | 0.02 | 13,824 | 131,072 | qwen2 | 40 | 48 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.43.3 | true | 152,064 | null | 0 | null | null | null | null |
HuggingFaceH4/zephyr-7b-beta | null | null | "2023-10-26T11:25:49Z" | null | null | 732,427 | null | null | null | null | 1,604 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"conversational",
"en",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"arxiv:2305.18290",
"arxiv:2310.16944",
"arxiv:2305.14233",
"arxiv:2310.01377",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:mit",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 7 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.35.0 | true | 32,000 | null | null | null | null | null | null |
jinaai/reader-lm-1.5b | null | null | "2024-09-06T02:56:15Z" | null | null | 8,202 | null | null | null | null | 471 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"multilingual",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"region:eu"
] | text-generation | null | null | 7 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 1,536 | 0.02 | 8,960 | 256,000 | qwen2 | 12 | 28 | 2 | 0.000001 | 2,000,000 | null | true | bfloat16 | 4.43.3 | true | 151,936 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-0.5B | null | null | "2024-09-15T12:15:39Z" | null | null | 99,649 | null | null | null | null | 96 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"arxiv:2407.10671",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 7 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.40.1 | true | 151,936 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-14B-Instruct | null | null | "2024-09-16T11:56:10Z" | null | null | 120,731 | null | null | null | null | 107 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-14B",
"base_model:finetune:Qwen/Qwen2.5-14B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 7 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 5,120 | 0.02 | 13,824 | 32,768 | qwen2 | 40 | 48 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
EleutherAI/gpt-j-6b | null | null | "2022-03-02T23:29:04Z" | null | null | 291,371 | null | null | null | null | 1,440 | transformers | [
"transformers",
"pytorch",
"tf",
"jax",
"gptj",
"text-generation",
"causal-lm",
"en",
"dataset:EleutherAI/pile",
"arxiv:2104.09864",
"arxiv:2101.00027",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 6 | [
"GPTJForCausalLM"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gptj | null | null | null | null | null | null | false | null | 4.18.0.dev0 | true | 50,400 | null | null | null | null | null | null |
georgesung/llama2_7b_chat_uncensored | null | null | "2023-07-20T10:45:03Z" | null | null | 2,820 | null | null | null | null | 363 | transformers | [
"transformers",
"pytorch",
"tensorboard",
"safetensors",
"llama",
"text-generation",
"dataset:georgesung/wizard_vicuna_70k_unfiltered",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 6 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 2,048 | llama | 32 | 32 | null | 0.00001 | null | null | false | float32 | 4.30.2 | true | 32,000 | null | null | null | null | null | null |
Qwen/Qwen2-7B-Instruct | null | null | "2024-06-04T10:07:03Z" | null | null | 611,535 | null | null | null | null | 585 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 6 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.41.2 | true | 152,064 | null | 0 | null | null | null | null |
Alibaba-NLP/gte-Qwen2-7B-instruct | null | null | "2024-06-15T11:24:21Z" | null | null | 34,262 | null | null | null | null | 201 | sentence-transformers | [
"sentence-transformers",
"safetensors",
"qwen2",
"text-generation",
"mteb",
"transformers",
"Qwen2",
"sentence-similarity",
"custom_code",
"arxiv:2308.03281",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-embeddings-inference",
"endpoints_compatible",
"region:us"
] | sentence-similarity | null | null | 6 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 131,072 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | float32 | 4.41.2 | true | 151,646 | null | 0 | null | null | null | null |
upstage/solar-pro-preview-instruct | null | null | "2024-09-09T01:08:58Z" | null | null | 1,073 | null | null | null | null | 421 | transformers | [
"transformers",
"safetensors",
"solar",
"text-generation",
"nlp",
"conversational",
"custom_code",
"en",
"license:mit",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 6 | [
"SolarForCausalLM"
] | 1 | 32,007 | silu | 5,120 | 0.02 | 17,920 | 4,096 | solar | 40 | 64 | 10 | 0.00001 | 10,000 | 2,047 | false | bfloat16 | 4.44.2 | true | 32,128 | false | 0 | null | false | 1 | null |
Qwen/Qwen2.5-32B-Instruct | null | null | "2024-09-17T04:17:55Z" | null | null | 119,483 | null | null | null | null | 115 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-32B",
"base_model:finetune:Qwen/Qwen2.5-32B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 6 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 5,120 | 0.02 | 27,648 | 32,768 | qwen2 | 40 | 64 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-Coder-1.5B-Instruct | null | null | "2024-09-18T09:41:47Z" | null | null | 18,374 | null | null | null | null | 40 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"conversational",
"en",
"arxiv:2409.12186",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-Coder-1.5B",
"base_model:finetune:Qwen/Qwen2.5-Coder-1.5B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 6 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 1,536 | 0.02 | 8,960 | 32,768 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | 131,072 | true | bfloat16 | 4.44.0 | true | 151,936 | null | 0 | null | null | null | null |
allenai/Molmo-7B-D-0924 | null | null | "2024-09-25T01:48:22Z" | null | null | 73,568 | null | null | null | null | 424 | transformers | [
"transformers",
"safetensors",
"molmo",
"text-generation",
"multimodal",
"olmo",
"pixmo",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"arxiv:2409.17146",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 6 | [
"MolmoForCausalLM"
] | null | null | null | 3,584 | 0.02 | 37,888 | 4,096 | molmo | 28 | 28 | 4 | null | 1,000,000 | null | false | float32 | 4.43.3 | true | 152,064 | null | null | null | null | null | null |
prithivMLmods/Llama-3.2-8B-GGUF-200K | null | null | "2024-10-27T05:09:21Z" | null | null | 11,772 | null | null | null | null | 7 | transformers | [
"transformers",
"gguf",
"llama",
"text-generation-inference",
"unsloth",
"200K",
"text-generation",
"en",
"dataset:HuggingFaceH4/ultrachat_200k",
"license:creativeml-openrail-m",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 6 | null | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
HuggingFaceTB/SmolLM2-135M | null | null | "2024-10-31T00:46:04Z" | null | null | 7,803 | null | null | null | null | 27 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 6 | [
"LlamaForCausalLM"
] | 0 | 0 | silu | 576 | 0.041667 | 1,536 | 8,192 | llama | 9 | 30 | 3 | 0.00001 | 100,000 | null | true | bfloat16 | 4.40.1 | true | 49,152 | false | 0 | null | null | 1 | null |
cognitivecomputations/dolphin-2.9-llama3-8b | null | null | "2024-04-20T23:14:52Z" | null | null | 3,628 | null | null | null | null | 418 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:abacusai/SystemChat-1.1",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"LlamaForCausalLM"
] | 128,000 | 128,256 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0 | false | 128,258 | false | 0 | null | null | 1 | null |
Qwen/Qwen2-0.5B | null | null | "2024-05-31T08:38:11Z" | null | null | 1,267,361 | null | null | null | null | 108 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"pretrained",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 896 | 0.02 | 4,864 | 131,072 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | 131,072 | true | bfloat16 | 4.40.1 | true | 151,936 | null | 0 | null | null | null | null |
Sao10K/L3-8B-Stheno-v3.2 | null | null | "2024-06-05T10:30:57Z" | null | null | 2,737 | null | null | null | null | 238 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"dataset:Gryphe/Opus-WritingPrompts",
"dataset:Sao10K/Claude-3-Opus-Instruct-15K",
"dataset:Sao10K/Short-Storygen-v2",
"dataset:Sao10K/c2-Logs-Filtered",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.41.2 | true | 128,256 | false | 0 | null | false | 1 | null |
princeton-nlp/gemma-2-9b-it-SimPO | null | null | "2024-07-16T16:42:49Z" | null | null | 102,892 | null | null | null | null | 120 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"alignment-handbook",
"generated_from_trainer",
"conversational",
"dataset:princeton-nlp/gemma2-ultrafeedback-armorm",
"arxiv:2405.14734",
"arxiv:2310.01377",
"arxiv:2406.12845",
"base_model:google/gemma-2-9b-it",
"base_model:finetune:google/gemma-2-9b-it",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,584 | 0.02 | 14,336 | 8,192 | gemma2 | 16 | 42 | 8 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.42.4 | true | 256,000 | false | 0 | 256 | null | null | null |
mattshumer/Reflection-Llama-3.1-70B | null | null | "2024-09-05T18:29:50Z" | null | null | 2,503 | null | null | null | null | 1,702 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"base_model:meta-llama/Llama-3.1-70B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-70B-Instruct",
"license:llama3.1",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 8,192 | 0.02 | 28,672 | 8,192 | llama | 64 | 80 | 8 | 0.00001 | 500,000 | null | false | float32 | 4.40.0 | true | 128,262 | false | 0 | null | null | 1 | null |
anthracite-org/magnum-v4-72b | null | null | "2024-09-20T03:23:10Z" | null | null | 930 | null | null | null | null | 20 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"Qwen2ForCausalLM"
] | null | 151,645 | silu | 8,192 | 0.02 | 29,568 | 32,768 | qwen2 | 64 | 80 | 8 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.44.0 | false | 152,064 | null | 0 | null | null | null | null |
Vikhrmodels/Vikhr-Nemo-12B-Instruct-R-21-09-24 | null | null | "2024-09-20T13:32:03Z" | null | null | 13,734 | null | null | null | null | 81 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"en",
"ru",
"dataset:Vikhrmodels/GrandMaster-PRO-MAX",
"dataset:Vikhrmodels/Grounded-RAG-RU-v2",
"arxiv:2405.13929",
"base_model:mistralai/Mistral-Nemo-Instruct-2407",
"base_model:finetune:mistralai/Mistral-Nemo-Instruct-2407",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.2 | true | 131,074 | null | 0 | 128 | null | null | null |
VongolaChouko/Starcannon-Unleashed-12B-v1.0 | null | null | "2024-10-29T14:32:59Z" | null | null | 790 | null | null | null | null | 23 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"base_model:MarinaraSpaghetti/NemoMix-Unleashed-12B",
"base_model:merge:MarinaraSpaghetti/NemoMix-Unleashed-12B",
"base_model:nothingiisreal/MN-12B-Starcannon-v3",
"base_model:merge:nothingiisreal/MN-12B-Starcannon-v3",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.0 | true | 131,072 | null | 0 | 128 | null | null | null |
facebook/MobileLLM-600M | null | null | "2024-10-30T22:57:03Z" | null | null | 1,061 | null | null | null | null | 26 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mobilellm",
"text-generation",
"custom_code",
"arxiv:2402.14905",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"MobileLLMForCausalLM"
] | 1 | 2 | silu | 1,152 | 0.02 | 3,072 | 2,048 | mobilellm | 18 | 40 | 6 | 0.00001 | 10,000 | null | false | float16 | 4.41.2 | true | 32,000 | false | 0 | 64 | false | 1 | null |
amd/AMD-OLMo-1B | null | null | "2024-10-31T20:27:49Z" | null | null | 246 | null | null | null | null | 19 | null | [
"safetensors",
"olmo",
"text-generation",
"dataset:allenai/dolma",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 5 | [
"OlmoForCausalLM"
] | null | 50,279 | silu | 2,048 | 0.02 | 8,192 | 2,048 | olmo | 16 | 16 | 16 | null | 10,000 | null | true | float32 | 4.40.2 | true | 50,304 | false | 0 | null | null | null | null |
EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2 | null | null | "2024-11-06T19:49:06Z" | null | null | 53 | null | null | null | null | 5 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"conversational",
"dataset:anthracite-org/kalo-opus-instruct-22k-no-refusal",
"dataset:Nopm/Opus_WritingStruct",
"dataset:Gryphe/Sonnet3.5-SlimOrcaDedupCleaned",
"dataset:Gryphe/Sonnet3.5-Charcard-Roleplay",
"dataset:Gryphe/ChatGPT-4o-Writing-Prompts",
"dataset:Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned",
"dataset:Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned",
"dataset:nothingiisreal/Reddit-Dirty-And-WritingPrompts",
"dataset:allura-org/Celeste-1.x-data-mixture",
"dataset:cognitivecomputations/dolphin-2.9.3",
"base_model:Qwen/Qwen2.5-14B",
"base_model:finetune:Qwen/Qwen2.5-14B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 5 | [
"Qwen2ForCausalLM"
] | null | 151,643 | silu | 5,120 | 0.02 | 13,824 | 131,072 | qwen2 | 40 | 48 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.45.1 | false | 152,064 | null | 0 | null | null | null | null |
facebook/opt-1.3b | null | null | "2022-05-11T08:26:00Z" | null | null | 18,621,886 | null | null | null | null | 154 | transformers | [
"transformers",
"pytorch",
"tf",
"jax",
"opt",
"text-generation",
"en",
"arxiv:2205.01068",
"arxiv:2005.14165",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | null | null | 4 | [
"OPTForCausalLM"
] | 2 | 2 | null | 2,048 | null | null | 2,048 | opt | 32 | 24 | null | null | null | null | null | float16 | 4.21.0.dev0 | true | 50,272 | null | 0 | null | null | null | null |
huggyllama/llama-7b | null | null | "2023-04-03T23:16:48Z" | null | null | 154,247 | null | null | null | null | 294 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"conversational",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 2,048 | llama | 32 | 32 | null | 0.000001 | null | null | false | float16 | 4.28.0.dev0 | true | 32,000 | null | null | null | null | null | null |
ai-forever/ruGPT-3.5-13B | null | null | "2023-05-02T12:53:36Z" | null | null | 2,878 | null | null | null | null | 260 | transformers | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"gpt3",
"en",
"ru",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"GPT2LMHeadModel"
] | 2 | 3 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.27.1 | true | 50,272 | null | null | null | null | null | null |
TheBloke/dolphin-2.5-mixtral-8x7b-GPTQ | null | null | "2023-12-14T10:34:15Z" | null | null | 168 | null | null | null | null | 105 | transformers | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"en",
"dataset:ehartford/dolphin",
"dataset:jondurbin/airoboros-2.2.1",
"dataset:ehartford/dolphin-coder",
"dataset:migtissera/Synthia-v1.3",
"dataset:teknium/openhermes",
"dataset:ise-uiuc/Magicoder-OSS-Instruct-75K",
"dataset:ise-uiuc/Magicoder-Evol-Instruct-110K",
"dataset:LDJnr/Pure-Dove",
"base_model:cognitivecomputations/dolphin-2.5-mixtral-8x7b",
"base_model:quantized:cognitivecomputations/dolphin-2.5-mixtral-8x7b",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | null | null | 4 | [
"MixtralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mixtral | 32 | 32 | 8 | 0.00001 | 1,000,000 | 4,096 | false | bfloat16 | 4.37.0.dev0 | true | 32,002 | null | 0 | null | null | 1 | null |
defog/sqlcoder-7b-2 | null | null | "2024-02-05T14:36:51Z" | null | null | 117,028 | null | null | null | null | 288 | transformers | [
"transformers",
"safetensors",
"gguf",
"llama",
"text-generation",
"license:cc-by-sa-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 16,384 | llama | 32 | 32 | 32 | 0.00001 | 1,000,000 | null | false | float16 | 4.37.2 | true | 32,016 | false | 0 | null | null | 1 | null |
RUCKBReasoning/TableLLM-7b | null | null | "2024-02-06T12:07:20Z" | null | null | 512 | null | null | null | null | 12 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Table",
"QA",
"Code",
"en",
"dataset:RUCKBReasoning/TableLLM-SFT",
"arxiv:2403.19318",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 16,384 | llama | 32 | 32 | 32 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.36.2 | true | 32,016 | false | 0 | null | null | 1 | null |
BAAI/bge-reranker-v2-gemma | null | null | "2024-03-16T12:09:04Z" | null | null | 13,503 | null | null | null | null | 48 | sentence-transformers | [
"sentence-transformers",
"safetensors",
"gemma",
"text-generation",
"transformers",
"text-classification",
"multilingual",
"arxiv:2312.15503",
"arxiv:2402.03216",
"license:apache-2.0",
"region:us"
] | text-classification | null | null | 4 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu | 2,048 | 0.02 | 16,384 | 8,192 | gemma | 8 | 18 | 1 | 0.000001 | 10,000 | null | null | float32 | 4.38.1 | true | 256,000 | false | 0 | 256 | null | null | null |
IlyaGusev/saiga_llama3_8b | null | null | "2024-04-18T18:25:25Z" | null | null | 13,161 | null | null | null | null | 108 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"ru",
"dataset:IlyaGusev/saiga_scored",
"doi:10.57967/hf/2368",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.40.0 | true | 128,256 | false | 0 | null | null | 1 | null |
nothingiisreal/MN-12B-Celeste-V1.9 | null | null | "2024-07-31T04:55:12Z" | null | null | 565 | null | null | null | null | 119 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"en",
"dataset:nothingiisreal/c2-logs-cleaned",
"dataset:kalomaze/Opus_Instruct_25k",
"dataset:nothingiisreal/Reddit-Dirty-And-WritingPrompts",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.0.dev0 | false | 131,072 | null | 0 | 128 | null | null | null |
lmms-lab/LLaVA-Video-7B-Qwen2 | null | null | "2024-09-02T06:36:42Z" | null | null | 35,547 | null | null | null | null | 32 | transformers | [
"transformers",
"safetensors",
"llava",
"text-generation",
"multimodal",
"video-text-to-text",
"en",
"dataset:lmms-lab/LLaVA-OneVision-Data",
"dataset:lmms-lab/LLaVA-Video-178K",
"arxiv:2410.02713",
"base_model:lmms-lab/llava-onevision-qwen2-7b-si",
"base_model:finetune:lmms-lab/llava-onevision-qwen2-7b-si",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | video-text-to-text | null | null | 4 | [
"LlavaQwenForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | llava | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.40.0.dev0 | true | 152,064 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4 | null | null | "2024-09-17T12:51:51Z" | null | null | 74,308 | null | null | null | null | 9 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-7B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-7B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"gptq",
"region:us"
] | text-generation | null | null | 4 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | float16 | 4.39.3 | true | 152,064 | null | 0 | null | null | null | null |
Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4 | null | null | "2024-09-17T12:52:55Z" | null | null | 11,841 | null | null | null | null | 17 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-32B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-32B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"gptq",
"region:us"
] | text-generation | null | null | 4 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 5,120 | 0.02 | 27,648 | 32,768 | qwen2 | 40 | 64 | 8 | 0.000001 | 1,000,000 | 131,072 | false | float16 | 4.39.3 | true | 152,064 | null | 0 | null | null | null | null |
AIDC-AI/Ovis1.6-Gemma2-9B | null | null | "2024-09-18T12:29:44Z" | null | null | 7,812 | null | null | null | null | 241 | transformers | [
"transformers",
"safetensors",
"ovis",
"text-generation",
"MLLM",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"dataset:AIDC-AI/Ovis-dataset",
"arxiv:2405.20797",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 4 | [
"Ovis"
] | null | null | null | 3,584 | null | null | null | ovis | null | null | null | null | null | null | null | bfloat16 | 4.44.2 | true | null | null | null | null | null | null | null |
MaziyarPanahi/Qwen2.5-7B-Instruct-GGUF | null | null | "2024-09-18T19:44:20Z" | null | null | 893,193 | null | null | null | null | 5 | null | [
"gguf",
"mistral",
"quantized",
"2-bit",
"3-bit",
"4-bit",
"5-bit",
"6-bit",
"8-bit",
"GGUF",
"text-generation",
"base_model:Qwen/Qwen2.5-7B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-7B-Instruct",
"region:us",
"imatrix",
"conversational"
] | text-generation | null | null | 4 | null | null | null | null | null | null | null | null | mistral | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null |
huihui-ai/Qwen2.5-7B-Instruct-abliterated-v2 | null | null | "2024-09-22T19:02:37Z" | null | null | 2,726 | null | null | null | null | 15 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"abliterated",
"uncensored",
"conversational",
"en",
"base_model:Qwen/Qwen2.5-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-7B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
allenai/Molmo-72B-0924 | null | null | "2024-09-25T06:23:32Z" | null | null | 6,780 | null | null | null | null | 254 | transformers | [
"transformers",
"safetensors",
"molmo",
"text-generation",
"multimodal",
"olmo",
"pixmo",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"arxiv:2409.17146",
"base_model:Qwen/Qwen2-72B",
"base_model:finetune:Qwen/Qwen2-72B",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 4 | [
"MolmoForCausalLM"
] | null | null | null | 8,192 | 0.02 | 59,136 | 4,096 | molmo | 64 | 80 | 8 | null | 1,000,000 | null | false | float32 | 4.43.3 | true | 152,064 | null | null | null | null | null | null |
BAAI/Emu3-Gen | null | null | "2024-09-25T11:03:49Z" | null | null | 14,857 | null | null | null | null | 184 | transformers | [
"transformers",
"safetensors",
"Emu3",
"text-generation",
"any-to-any",
"custom_code",
"arxiv:2409.18869",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | any-to-any | null | null | 4 | [
"Emu3ForCausalLM"
] | 151,849 | 151,850 | silu | 4,096 | 0.02 | 14,336 | 9,216 | Emu3 | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float32 | 4.44.0 | true | 184,622 | null | 0.1 | null | null | 1 | null |
shuttleai/shuttle-3 | null | null | "2024-10-09T19:48:20Z" | null | null | 442 | null | null | null | null | 28 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"base_model:Qwen/Qwen2.5-72B",
"base_model:finetune:Qwen/Qwen2.5-72B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"Qwen2ForCausalLM"
] | null | 151,645 | silu | 8,192 | 0.02 | 29,568 | 32,768 | qwen2 | 64 | 80 | 8 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.45.0.dev0 | false | 152,064 | null | 0 | null | null | null | null |
h2oai/h2ovl-mississippi-2b | null | null | "2024-10-15T18:15:29Z" | null | null | 4,914 | null | null | null | null | 19 | transformers | [
"transformers",
"safetensors",
"h2ovl_chat",
"feature-extraction",
"gpt",
"llm",
"multimodal large language model",
"ocr",
"text-generation",
"conversational",
"custom_code",
"en",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 4 | [
"H2OVLChatModel"
] | null | null | null | null | null | null | null | h2ovl_chat | null | null | null | null | null | null | null | bfloat16 | null | null | null | null | null | null | null | null | null |
BSC-LT/salamandraTA-2B | null | null | "2024-10-28T08:43:09Z" | null | null | 304 | null | null | null | null | 4 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"translation",
"it",
"pt",
"de",
"en",
"es",
"eu",
"gl",
"fr",
"bg",
"cs",
"lt",
"hr",
"ca",
"nl",
"ro",
"da",
"el",
"fi",
"hu",
"sk",
"sl",
"et",
"pl",
"lv",
"mt",
"ga",
"sv",
"an",
"ast",
"oc",
"arxiv:1803.09010",
"arxiv:2010.11125",
"arxiv:2403.14009",
"arxiv:1907.05791",
"arxiv:1911.04944",
"arxiv:2207.04672",
"base_model:BSC-LT/salamandra-2b",
"base_model:finetune:BSC-LT/salamandra-2b",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:eu"
] | translation | null | null | 4 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 5,440 | 8,192 | llama | 16 | 24 | 16 | 0.00001 | 10,000 | null | false | bfloat16 | 4.42.4 | false | 256,000 | false | 0 | null | false | 1 | null |
arcee-ai/Arcee-VyLinh | null | null | "2024-10-29T20:49:46Z" | null | null | 761 | null | null | null | null | 17 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mergekit",
"merge",
"conversational",
"vi",
"base_model:Qwen/Qwen2.5-3B-Instruct",
"base_model:merge:Qwen/Qwen2.5-3B-Instruct",
"base_model:qnguyen3/VyLinh-3B",
"base_model:merge:qnguyen3/VyLinh-3B",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | bfloat16 | 4.46.1 | true | 151,936 | null | 0 | null | null | null | null |
amd/AMD-OLMo-1B-SFT | null | null | "2024-10-31T20:28:44Z" | null | null | 912 | null | null | null | null | 17 | transformers | [
"transformers",
"safetensors",
"olmo",
"text-generation",
"dataset:allenai/dolma",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"OlmoForCausalLM"
] | null | 50,279 | silu | 2,048 | 0.02 | 8,192 | 2,048 | olmo | 16 | 16 | 16 | null | 10,000 | null | true | float32 | 4.40.2 | true | 50,304 | false | 0 | null | null | null | null |
TechxGenus/Typst-Coder-9B | null | null | "2024-11-03T14:48:21Z" | null | null | 11 | null | null | null | null | 4 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"code",
"conversational",
"base_model:01-ai/Yi-Coder-9B",
"base_model:finetune:01-ai/Yi-Coder-9B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 131,072 | llama | 32 | 48 | 4 | 0.00001 | 10,000,000 | null | false | bfloat16 | 4.45.2 | false | 64,000 | false | 0 | 128 | false | 1 | null |
allura-org/G2-9B-Aletheia-v1 | null | null | "2024-11-03T15:12:20Z" | null | null | 297 | null | null | null | null | 8 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3",
"base_model:merge:UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3",
"base_model:allura-org/G2-9B-Sugarquill-v0",
"base_model:merge:allura-org/G2-9B-Sugarquill-v0",
"base_model:crestf411/gemma2-9B-sunfall-v0.5.2",
"base_model:merge:crestf411/gemma2-9B-sunfall-v0.5.2",
"license:gemma",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 4 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,584 | 0.02 | 14,336 | 8,192 | gemma2 | 16 | 42 | 8 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.45.2 | true | 256,000 | false | 0 | 256 | null | null | null |
theprint/ReWiz-Qwen-2.5-14B | null | null | "2024-11-05T10:01:22Z" | null | null | 256 | null | null | null | null | 4 | transformers | [
"transformers",
"safetensors",
"gguf",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"theprint",
"rewiz",
"en",
"dataset:theprint/ReWiz",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 4 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 5,120 | 0.02 | 13,824 | 131,072 | qwen2 | 40 | 48 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.2 | true | 152,064 | null | 0 | null | null | null | null |
openai-community/gpt2-large | null | null | "2022-03-02T23:29:04Z" | null | null | 1,207,892 | null | null | null | null | 270 | transformers | [
"transformers",
"pytorch",
"tf",
"jax",
"rust",
"onnx",
"safetensors",
"gpt2",
"text-generation",
"en",
"arxiv:1910.09700",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | null | null | null | 50,257 | null | null | null | null | null | null |
pierreguillou/gpt2-small-portuguese | null | null | "2022-03-02T23:29:05Z" | null | null | 8,870 | null | null | null | null | 40 | transformers | [
"transformers",
"pytorch",
"tf",
"jax",
"gpt2",
"text-generation",
"pt",
"dataset:wikipedia",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | null | null | null | 50,257 | null | null | null | null | null | null |
Qwen/Qwen-VL-Chat | null | null | "2023-08-20T04:45:22Z" | null | null | 21,054 | null | null | null | null | 337 | transformers | [
"transformers",
"pytorch",
"qwen",
"text-generation",
"custom_code",
"zh",
"en",
"arxiv:2308.12966",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"QWenLMHeadModel"
] | null | null | null | 4,096 | 0.02 | 22,016 | 8,192 | qwen | 32 | 32 | null | null | null | null | false | bfloat16 | 4.31.0 | true | 151,936 | null | null | null | null | null | null |
microsoft/phi-1 | null | null | "2023-09-10T04:10:57Z" | null | null | 9,179 | null | null | null | null | 207 | transformers | [
"transformers",
"safetensors",
"phi",
"text-generation",
"code",
"en",
"arxiv:2306.11644",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"PhiForCausalLM"
] | null | null | gelu_new | 2,048 | 0.02 | 8,192 | 2,048 | phi | 32 | 24 | null | null | 10,000 | null | false | float32 | 4.37.0 | true | 51,200 | null | 0 | null | null | null | null |
prometheus-eval/prometheus-13b-v1.0 | null | null | "2023-10-12T07:19:38Z" | null | null | 3,889 | null | null | null | null | 126 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"text2text-generation",
"en",
"dataset:kaist-ai/Feedback-Collection",
"arxiv:2310.08491",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text2text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llama | 40 | 40 | 40 | 0.00001 | 10,000 | null | false | float32 | 4.33.1 | true | 32,000 | null | null | null | null | 1 | null |
teknium/OpenHermes-2.5-Mistral-7B | null | null | "2023-10-29T20:36:39Z" | null | null | 144,479 | null | null | null | null | 813 | transformers | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.34.0.dev0 | false | 32,002 | null | null | null | null | null | null |
01-ai/Yi-34B-200K | null | null | "2023-11-06T01:46:54Z" | null | null | 5,178 | null | null | null | null | 317 | transformers | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"arxiv:2403.04652",
"arxiv:2311.16502",
"arxiv:2401.11944",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 7,168 | 0.02 | 20,480 | 200,000 | llama | 56 | 60 | 8 | 0.00001 | 10,000,000 | null | false | bfloat16 | 4.34.0 | true | 64,000 | null | null | null | null | 1 | null |
GOAT-AI/GOAT-70B-Storytelling | null | null | "2023-11-17T08:12:07Z" | null | null | 1,002 | null | null | null | null | 41 | transformers | [
"transformers",
"pytorch",
"llama",
"text-generation",
"facebook",
"meta",
"llama-2",
"Storywriter",
"license:llama2",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 8,192 | 0.02 | 28,672 | 4,096 | llama | 64 | 80 | 8 | 0.00001 | 10,000 | null | false | bfloat16 | 4.34.1 | true | 32,000 | false | null | null | null | 1 | null |
upstage/SOLAR-10.7B-v1.0 | null | null | "2023-12-12T14:57:41Z" | null | null | 32,758 | null | null | null | null | 291 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:2312.15166",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 4,096 | llama | 32 | 48 | 8 | 0.00001 | 10,000 | null | false | float16 | 4.35.2 | false | 32,000 | false | null | null | null | 1 | null |
microsoft/phi-2 | null | null | "2023-12-13T21:19:59Z" | null | null | 231,230 | null | null | null | null | 3,240 | transformers | [
"transformers",
"safetensors",
"phi",
"text-generation",
"nlp",
"code",
"en",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"PhiForCausalLM"
] | 50,256 | 50,256 | gelu_new | 2,560 | 0.02 | 10,240 | 2,048 | phi | 32 | 32 | 32 | null | 10,000 | null | false | float16 | 4.37.0 | true | 51,200 | null | 0 | null | null | null | null |
TinyLlama/TinyLlama-1.1B-Chat-v1.0 | null | null | "2023-12-30T06:27:30Z" | null | null | 1,255,617 | null | null | null | null | 1,087 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"dataset:cerebras/SlimPajama-627B",
"dataset:bigcode/starcoderdata",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 5,632 | 2,048 | llama | 32 | 22 | 4 | 0.00001 | 10,000 | null | false | bfloat16 | 4.35.0 | true | 32,000 | false | null | null | null | 1 | null |
Unbabel/TowerInstruct-13B-v0.1 | null | null | "2024-01-29T10:39:36Z" | null | null | 513 | null | null | null | null | 21 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"translation",
"en",
"de",
"fr",
"zh",
"pt",
"nl",
"ru",
"ko",
"it",
"es",
"arxiv:2402.17733",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | translation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 32,005 | silu | 5,120 | 0.02 | 13,824 | 4,096 | llama | 40 | 40 | 40 | 0.00001 | 10,000 | null | false | float32 | 4.36.2 | false | 32,007 | false | 0 | null | null | 1 | null |
liuhaotian/llava-v1.6-vicuna-7b | null | null | "2024-01-31T04:32:37Z" | null | null | 18,549 | null | null | null | null | 98 | transformers | [
"transformers",
"safetensors",
"llava",
"text-generation",
"image-text-to-text",
"autotrain_compatible",
"region:us"
] | image-text-to-text | null | null | 3 | [
"LlavaLlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llava | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.36.2 | true | 32,000 | false | 0 | null | null | 1 | null |
Ttimofeyka/MistralRP-Noromaid-NSFW-Mistral-7B-GGUF | null | null | "2024-02-07T14:19:28Z" | null | null | 19,888 | null | null | null | null | 29 | transformers | [
"transformers",
"safetensors",
"gguf",
"mistral",
"text-generation",
"mergekit",
"merge",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.37.2 | true | 32,000 | null | 0 | null | null | null | null |
wolfram/miquliz-120b-v2.0 | null | null | "2024-02-10T11:41:03Z" | null | null | 463 | null | null | null | null | 93 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"en",
"de",
"fr",
"es",
"it",
"arxiv:2203.05482",
"base_model:152334H/miqu-1-70b-sf",
"base_model:merge:152334H/miqu-1-70b-sf",
"base_model:lizpreciatior/lzlv_70b_fp16_hf",
"base_model:merge:lizpreciatior/lzlv_70b_fp16_hf",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 8,192 | 0.02 | 28,672 | 32,768 | llama | 64 | 140 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.37.2 | true | 32,000 | false | 0 | null | null | 1 | null |
BioMistral/BioMistral-7B | null | null | "2024-02-14T11:33:32Z" | null | null | 12,006 | null | null | null | null | 398 | transformers | [
"transformers",
"pytorch",
"tensorboard",
"mistral",
"text-generation",
"medical",
"biology",
"conversational",
"fr",
"en",
"de",
"nl",
"es",
"pt",
"pl",
"ro",
"it",
"arxiv:2402.10373",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.35.0 | false | 32,000 | null | null | null | null | null | null |
yanolja/EEVE-Korean-Instruct-10.8B-v1.0 | null | null | "2024-02-22T04:39:04Z" | null | null | 17,880 | null | null | null | null | 129 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"conversational",
"arxiv:2402.14714",
"arxiv:2310.01377",
"arxiv:2306.02707",
"base_model:yanolja/EEVE-Korean-10.8B-v1.0",
"base_model:finetune:yanolja/EEVE-Korean-10.8B-v1.0",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 3 | [
"LlamaForCausalLM"
] | 1 | 32,000 | silu | 4,096 | 0.02 | 14,336 | 4,096 | llama | 32 | 48 | 8 | 0.00001 | 10,000 | null | false | bfloat16 | 4.36.2 | false | 40,960 | false | 0 | null | null | 1 | null |