id
stringlengths
7
117
author
stringclasses
6 values
sha
null
created_at
unknown
last_modified
null
disabled
null
downloads
int64
0
18.6M
downloads_all_time
null
gated
bool
1 class
gguf
null
inference
null
likes
int64
0
4.77k
library_name
stringclasses
36 values
tags
sequencelengths
1
430
pipeline_tag
stringclasses
32 values
mask_token
null
model_index
null
trending_score
int64
0
132
architectures
sequencelengths
1
5
bos_token_id
int64
-1
256k
eos_token_id
int64
-1
256k
hidden_act
stringclasses
15 values
hidden_size
int64
1
20.5k
initializer_range
float64
0
1
intermediate_size
int64
1
98.3k
max_position_embeddings
int64
8
1.05M
model_type
stringclasses
530 values
num_attention_heads
int64
1
5k
num_hidden_layers
int64
-1
8.93k
num_key_value_heads
int64
1
160
rms_norm_eps
float64
0
7
rope_theta
float64
1k
1,000B
sliding_window
int64
0
262k
tie_word_embeddings
bool
2 classes
torch_dtype
stringclasses
8 values
transformers_version
stringclasses
207 values
use_cache
bool
2 classes
vocab_size
int64
-1
5.03M
attention_bias
bool
2 classes
attention_dropout
float64
0
0.5
head_dim
int64
2
256
mlp_bias
bool
2 classes
pretraining_tp
int64
0
8
rope_scaling
dict
NexaAIDev/Octopus-v2
null
null
"2024-03-16T17:54:22Z"
null
null
968
null
null
null
null
857
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "function calling", "on-device language model", "android", "conversational", "en", "arxiv:2404.19296", "arxiv:2404.11459", "arxiv:2404.01744", "base_model:google/gemma-2b", "base_model:finetune:google/gemma-2b", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
3
[ "GemmaForCausalLM" ]
2
1
gelu
2,048
0.02
16,384
8,192
gemma
8
18
1
0.000001
10,000
null
null
bfloat16
4.39.0.dev0
true
256,022
false
0
256
null
null
null
Qwen/CodeQwen1.5-7B-Chat
null
null
"2024-04-15T07:17:06Z"
null
null
47,065
null
null
null
null
314
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.16609", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
2
2
silu
4,096
0.02
13,440
65,536
qwen2
32
32
4
0.00001
1,000,000
65,536
false
bfloat16
4.39.3
true
92,416
null
0
null
null
null
null
aaditya/Llama3-OpenBioLLM-8B
null
null
"2024-04-20T00:53:57Z"
null
null
18,253
null
null
null
null
156
transformers
[ "transformers", "pytorch", "llama", "text-generation", "llama-3", "Mixtral", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "distillation", "en", "arxiv:2305.18290", "arxiv:2303.13375", "arxiv:2212.13138", "arxiv:2305.09617", "arxiv:2402.07023", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.0.dev0
false
128,256
false
0
null
null
1
null
shenzhi-wang/Llama3-8B-Chinese-Chat
null
null
"2024-04-21T15:11:45Z"
null
null
7,593
null
null
null
null
645
transformers
[ "transformers", "safetensors", "llama", "text-generation", "llama-factory", "orpo", "conversational", "en", "zh", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "doi:10.57967/hf/2316", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.0
true
128,256
false
0
null
null
1
null
microsoft/Phi-3-mini-4k-instruct
null
null
"2024-04-22T16:18:17Z"
null
null
1,347,345
null
null
null
null
1,070
transformers
[ "transformers", "safetensors", "phi3", "text-generation", "nlp", "code", "conversational", "custom_code", "en", "fr", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Phi3ForCausalLM" ]
1
32,000
silu
3,072
0.02
8,192
4,096
phi3
32
32
32
0.00001
10,000
2,047
false
bfloat16
4.40.2
true
32,064
false
0
null
null
null
null
MLP-KTLim/llama-3-Korean-Bllossom-8B
null
null
"2024-04-25T10:16:43Z"
null
null
89,597
null
null
null
null
273
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "en", "ko", "arxiv:2403.10882", "arxiv:2403.11399", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.43.1
false
128,256
false
0
null
false
1
null
01-ai/Yi-1.5-34B-Chat
null
null
"2024-05-10T06:47:21Z"
null
null
10,909
null
null
null
null
248
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:2403.04652", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
1
2
silu
7,168
0.02
20,480
4,096
llama
56
60
8
0.000001
5,000,000
null
false
bfloat16
4.40.0
false
64,000
false
0
null
null
1
null
mlabonne/NeuralDaredevil-8B-abliterated
null
null
"2024-05-27T19:33:23Z"
null
null
19,314
null
null
null
null
148
transformers
[ "transformers", "safetensors", "llama", "text-generation", "dpo", "conversational", "dataset:mlabonne/orpo-dpo-mix-40k", "license:llama3", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
float16
4.40.2
true
128,256
false
0
null
false
1
null
microsoft/Florence-2-large-ft
null
null
"2024-06-15T00:57:45Z"
null
null
66,725
null
null
null
null
298
transformers
[ "transformers", "pytorch", "florence2", "text-generation", "vision", "image-text-to-text", "custom_code", "arxiv:2311.06242", "license:mit", "autotrain_compatible", "region:us" ]
image-text-to-text
null
null
3
[ "Florence2ForConditionalGeneration" ]
0
2
null
null
null
null
null
florence2
null
null
null
null
null
null
null
float16
4.41.0.dev0
null
51,289
null
null
null
null
null
null
AI-MO/NuminaMath-7B-TIR
null
null
"2024-07-04T09:28:53Z"
null
null
6,769
null
null
null
null
318
transformers
[ "transformers", "safetensors", "llama", "text-generation", "alignment-handbook", "generated_from_trainer", "conversational", "arxiv:2309.17452", "base_model:deepseek-ai/deepseek-math-7b-base", "base_model:finetune:deepseek-ai/deepseek-math-7b-base", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
100,000
100,001
silu
4,096
0.02
11,008
4,096
llama
32
30
32
0.000001
10,000
null
false
bfloat16
4.42.3
true
102,400
false
0
null
false
1
null
MaziyarPanahi/calme-2.4-rys-78b
null
null
"2024-08-07T14:38:21Z"
null
null
484
null
null
null
null
42
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "qwen", "finetune", "chatml", "conversational", "en", "dataset:MaziyarPanahi/truthy-dpo-v0.1-axolotl", "dataset:Intel/orca_dpo_pairs", "base_model:MaziyarPanahi/calme-2.1-rys-78b", "base_model:finetune:MaziyarPanahi/calme-2.1-rys-78b", "license:mit", "model-index", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
null
151,645
silu
8,192
0.02
29,568
32,768
qwen2
64
86
8
0.000001
1,000,000
null
false
float32
4.44.0.dev0
false
151,646
null
0
null
null
null
null
ystemsrx/Qwen2-Boundless
null
null
"2024-08-19T06:49:29Z"
null
null
378
null
null
null
null
37
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "Qwen", "causal-lm", "fine-tuned", "ethics", "Chinese", "text2text-generation", "zh", "dataset:ystemsrx/Bad_Data_Alpaca", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
1,536
0.02
8,960
32,768
qwen2
12
28
2
0.000001
1,000,000
32,768
true
bfloat16
4.42.3
true
151,936
null
0
null
null
null
null
MarinaraSpaghetti/NemoMix-Unleashed-12B
null
null
"2024-08-21T16:42:43Z"
null
null
10,241
null
null
null
null
141
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.44.0
true
131,072
null
0
128
null
null
null
nbeerbower/Lyra-Gutenberg-mistral-nemo-12B
null
null
"2024-08-23T11:13:56Z"
null
null
176
null
null
null
null
18
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "dataset:jondurbin/gutenberg-dpo-v0.1", "base_model:Sao10K/MN-12B-Lyra-v1", "base_model:finetune:Sao10K/MN-12B-Lyra-v1", "license:cc-by-nc-4.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.44.2
true
131,072
null
0
128
null
null
null
speakleash/Bielik-11B-v2.3-Instruct
null
null
"2024-08-30T12:45:27Z"
null
null
6,926
null
null
null
null
22
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "conversational", "pl", "arxiv:2005.01643", "arxiv:2309.11235", "arxiv:2006.09092", "arxiv:2402.13228", "arxiv:2410.18565", "base_model:speakleash/Bielik-11B-v2", "base_model:merge:speakleash/Bielik-11B-v2", "base_model:speakleash/Bielik-11B-v2.0-Instruct", "base_model:merge:speakleash/Bielik-11B-v2.0-Instruct", "base_model:speakleash/Bielik-11B-v2.1-Instruct", "base_model:merge:speakleash/Bielik-11B-v2.1-Instruct", "base_model:speakleash/Bielik-11B-v2.2-Instruct", "base_model:merge:speakleash/Bielik-11B-v2.2-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
32,001
silu
4,096
0.02
14,336
32,768
mistral
32
50
8
0.00001
1,000,000
null
false
float16
4.44.2
true
32,128
null
0
null
null
null
null
HF1BitLLM/Llama3-8B-1.58-100B-tokens
null
null
"2024-09-10T10:35:45Z"
null
null
26,971
null
null
null
null
145
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:2402.17764", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "bitnet", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.44.0.dev0
true
128,256
false
0
null
false
1
null
ifable/gemma-2-Ifable-9B
null
null
"2024-09-10T11:48:51Z"
null
null
305
null
null
null
null
33
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "dataset:jondurbin/gutenberg-dpo-v0.1", "license:gemma", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.43.4
false
256,000
false
0
256
null
null
null
Qwen/Qwen2.5-3B
null
null
"2024-09-15T12:17:03Z"
null
null
97,632
null
null
null
null
33
null
[ "safetensors", "qwen2", "text-generation", "conversational", "en", "arxiv:2407.10671", "license:other", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
32,768
true
bfloat16
4.40.1
true
151,936
null
0
null
null
null
null
Qwen/Qwen2.5-32B
null
null
"2024-09-15T12:18:33Z"
null
null
23,072
null
null
null
null
28
null
[ "safetensors", "qwen2", "text-generation", "conversational", "en", "arxiv:2407.10671", "license:apache-2.0", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
5,120
0.02
27,648
131,072
qwen2
40
64
8
0.00001
1,000,000
131,072
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
inflatebot/MN-12B-Mag-Mell-R1
null
null
"2024-09-16T00:38:49Z"
null
null
154
null
null
null
null
18
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "arxiv:2311.03099", "arxiv:2306.01708", "base_model:Fizzarolli/MN-12b-Sunrose", "base_model:merge:Fizzarolli/MN-12b-Sunrose", "base_model:IntervitensInc/Mistral-Nemo-Base-2407-chatml", "base_model:merge:IntervitensInc/Mistral-Nemo-Base-2407-chatml", "base_model:anthracite-org/magnum-v2.5-12b-kto", "base_model:merge:anthracite-org/magnum-v2.5-12b-kto", "base_model:elinas/Chronos-Gold-12B-1.0", "base_model:merge:elinas/Chronos-Gold-12B-1.0", "base_model:nbeerbower/mistral-nemo-bophades-12B", "base_model:merge:nbeerbower/mistral-nemo-bophades-12B", "base_model:nbeerbower/mistral-nemo-gutenberg-12B-v4", "base_model:merge:nbeerbower/mistral-nemo-gutenberg-12B-v4", "base_model:nbeerbower/mistral-nemo-wissenschaft-12B", "base_model:merge:nbeerbower/mistral-nemo-wissenschaft-12B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
15
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.44.1
true
131,072
null
0
128
null
null
null
Qwen/Qwen2.5-0.5B-Instruct
null
null
"2024-09-16T11:52:46Z"
null
null
419,043
null
null
null
null
94
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-0.5B", "base_model:finetune:Qwen/Qwen2.5-0.5B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
32,768
true
bfloat16
4.43.1
true
151,936
null
0
null
null
null
null
thrishala/mental_health_chatbot
null
null
"2024-09-17T14:39:34Z"
null
null
689
null
null
null
null
7
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
float16
4.38.2
true
32,000
false
0
null
null
1
null
unsloth/Mistral-Small-Instruct-2409
null
null
"2024-09-17T17:48:49Z"
null
null
1,741
null
null
null
null
3
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
2
silu
6,144
0.02
16,384
131,072
mistral
48
56
8
0.00001
1,000,000
null
false
bfloat16
4.44.2
true
32,768
null
0
128
null
null
null
THUDM/cogvlm2-llama3-caption
null
null
"2024-09-18T11:21:11Z"
null
null
5,095
null
null
null
null
55
transformers
[ "transformers", "safetensors", "text-generation", "video-text-to-text", "custom_code", "en", "arxiv:2408.06072", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:other", "autotrain_compatible", "region:us" ]
video-text-to-text
null
null
3
[ "CogVLMVideoForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
2,048
null
32
32
null
0.00001
null
null
false
bfloat16
4.43.1
true
128,256
null
null
null
null
null
null
BAAI/Emu3-Chat
null
null
"2024-09-26T04:43:47Z"
null
null
4,600
null
null
null
null
71
transformers
[ "transformers", "safetensors", "Emu3", "text-generation", "custom_code", "arxiv:2409.18869", "license:apache-2.0", "autotrain_compatible", "region:us" ]
text-generation
null
null
3
[ "Emu3ForCausalLM" ]
151,849
151,850
silu
4,096
0.02
14,336
131,072
Emu3
32
32
8
0.00001
1,000,000
null
false
float32
4.44.0
true
184,622
null
0.1
null
null
1
null
ibm-granite/granite-3.0-2b-instruct
null
null
"2024-10-02T21:07:46Z"
null
null
15,000
null
null
null
null
34
transformers
[ "transformers", "safetensors", "granite", "text-generation", "language", "granite-3.0", "conversational", "arxiv:0000.00000", "base_model:ibm-granite/granite-3.0-2b-base", "base_model:finetune:ibm-granite/granite-3.0-2b-base", "license:apache-2.0", "model-index", "autotrain_compatible", "region:us" ]
text-generation
null
null
3
[ "GraniteForCausalLM" ]
0
0
silu
2,048
0.02
8,192
4,096
granite
32
40
8
0.00001
10,000
null
true
bfloat16
4.46.0.dev0
true
49,155
false
0.1
null
false
null
null
Steelskull/MSM-MS-Cydrion-22B
null
null
"2024-10-03T12:22:41Z"
null
null
113
null
null
null
null
10
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "conversational", "base_model:ArliAI/Mistral-Small-22B-ArliAI-RPMax-v1.1", "base_model:merge:ArliAI/Mistral-Small-22B-ArliAI-RPMax-v1.1", "base_model:TheDrummer/Cydonia-22B-v1.1", "base_model:merge:TheDrummer/Cydonia-22B-v1.1", "base_model:nbeerbower/Mistral-Small-Gutenberg-Doppel-22B", "base_model:merge:nbeerbower/Mistral-Small-Gutenberg-Doppel-22B", "base_model:rAIfle/Acolyte-22B", "base_model:merge:rAIfle/Acolyte-22B", "base_model:unsloth/Mistral-Small-Instruct-2409", "base_model:merge:unsloth/Mistral-Small-Instruct-2409", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
2
silu
6,144
0.02
16,384
131,072
mistral
48
56
8
0.00001
1,000,000
null
false
bfloat16
4.43.2
true
32,768
null
0
128
null
null
null
AIDC-AI/Ovis1.6-Gemma2-9B-GPTQ-Int4
null
null
"2024-10-15T12:10:13Z"
null
null
187
null
null
null
null
5
transformers
[ "transformers", "ovis", "text-generation", "MLLM", "image-text-to-text", "conversational", "custom_code", "en", "dataset:AIDC-AI/Ovis-dataset", "arxiv:2405.20797", "base_model:AIDC-AI/Ovis1.6-Gemma2-9B", "base_model:quantized:AIDC-AI/Ovis1.6-Gemma2-9B", "license:apache-2.0", "autotrain_compatible", "4-bit", "gptq", "region:us" ]
image-text-to-text
null
null
3
[ "Ovis" ]
null
null
null
3,584
null
null
null
ovis
null
null
null
null
null
null
null
bfloat16
4.44.2
true
null
null
null
null
null
null
null
ibm-granite/granite-guardian-3.0-8b
null
null
"2024-10-15T17:49:13Z"
null
null
1,076
null
null
null
null
27
transformers
[ "transformers", "safetensors", "granite", "text-generation", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "GraniteForCausalLM" ]
0
0
silu
4,096
0.02
12,800
8,192
granite
32
40
8
0.00001
10,000
null
true
bfloat16
4.45.0.dev0
false
49,155
false
0.1
null
false
null
null
AIDC-AI/Ovis1.6-Llama3.2-3B
null
null
"2024-10-16T13:01:34Z"
null
null
2,360
null
null
null
null
35
transformers
[ "transformers", "safetensors", "ovis", "text-generation", "MLLM", "image-text-to-text", "conversational", "custom_code", "en", "dataset:AIDC-AI/Ovis-dataset", "arxiv:2405.20797", "license:apache-2.0", "autotrain_compatible", "region:us" ]
image-text-to-text
null
null
3
[ "Ovis" ]
null
null
null
3,072
null
null
null
ovis
null
null
null
null
null
null
null
bfloat16
4.44.2
false
null
null
null
null
null
null
null
RefalMachine/ruadapt_qwen2.5_3B_ext_u48_instruct_v4
null
null
"2024-10-18T10:23:19Z"
null
null
1,597
null
null
null
null
21
null
[ "safetensors", "qwen2", "text-generation", "conversational", "ru", "dataset:IlyaGusev/saiga_scored", "dataset:IlyaGusev/saiga_preferences", "dataset:dichspace/darulm", "base_model:RefalMachine/ruadapt_qwen2.5_3B_ext_u48_full_lr5e4_peft_mlp_32_32_bs256", "base_model:finetune:RefalMachine/ruadapt_qwen2.5_3B_ext_u48_full_lr5e4_peft_mlp_32_32_bs256", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
null
147,077
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
bfloat16
4.45.2
true
147,097
null
0
null
null
null
null
anthracite-org/magnum-v4-12b
null
null
"2024-10-20T02:06:55Z"
null
null
806
null
null
null
null
21
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "chat", "conversational", "en", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.45.0.dev0
false
131,072
null
0
128
null
null
null
sarvamai/sarvam-1
null
null
"2024-10-23T13:18:53Z"
null
null
1,604
null
null
null
null
23
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "bn", "en", "gu", "hi", "kn", "ml", "mr", "or", "pa", "ta", "te", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
11,008
8,192
llama
16
28
8
0.000001
10,000
null
false
bfloat16
4.45.2
true
68,096
false
0
128
false
1
null
prithivMLmods/Llama-3.2-1B-GGUF
null
null
"2024-10-26T13:00:49Z"
null
null
1,339
null
null
null
null
5
transformers
[ "transformers", "gguf", "llama", "text-generation-inference", "ollama", "fast-llama", "1B", "meta", "meta-llama", "token", "text-generation", "en", "base_model:meta-llama/Llama-3.2-1B", "base_model:quantized:meta-llama/Llama-3.2-1B", "license:creativeml-openrail-m", "endpoints_compatible", "region:us", "conversational" ]
text-generation
null
null
3
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
fblgit/cybertron-v4-qw7B-MGS
null
null
"2024-10-29T12:44:49Z"
null
null
731
null
null
null
null
9
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "conversational", "en", "dataset:Magpie-Align/Magpie-Qwen2.5-Pro-1M-v0.1", "arxiv:2410.21228", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-7B-Instruct", "base_model:finetune:Qwen/Qwen2.5-7B-Instruct", "license:other", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.2
true
152,064
null
0
null
null
null
null
facebook/MobileLLM-350M
null
null
"2024-10-30T22:55:48Z"
null
null
929
null
null
null
null
24
transformers
[ "transformers", "pytorch", "safetensors", "mobilellm", "text-generation", "custom_code", "arxiv:2402.14905", "license:cc-by-nc-4.0", "autotrain_compatible", "region:us" ]
text-generation
null
null
3
[ "MobileLLMForCausalLM" ]
1
2
silu
960
0.02
2,560
2,048
mobilellm
15
32
5
0.00001
10,000
null
false
float16
4.41.2
true
32,000
false
0
64
false
1
null
HuggingFaceTB/SmolLM2-360M
null
null
"2024-10-31T00:43:11Z"
null
null
3,275
null
null
null
null
21
transformers
[ "transformers", "safetensors", "llama", "text-generation", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "LlamaForCausalLM" ]
0
0
silu
960
0.02
2,560
8,192
llama
15
32
5
0.00001
100,000
null
true
bfloat16
4.40.1
true
49,152
false
0
null
null
1
null
EVA-UNIT-01/EVA-Qwen2.5-32B-v0.1
null
null
"2024-10-31T05:13:25Z"
null
null
259
null
null
null
null
14
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "conversational", "dataset:anthracite-org/kalo-opus-instruct-22k-no-refusal", "dataset:Nopm/Opus_WritingStruct", "dataset:Gryphe/Sonnet3.5-SlimOrcaDedupCleaned", "dataset:Gryphe/Sonnet3.5-Charcard-Roleplay", "dataset:Gryphe/ChatGPT-4o-Writing-Prompts", "dataset:Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned", "dataset:Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned", "dataset:nothingiisreal/Reddit-Dirty-And-WritingPrompts", "dataset:allura-org/Celeste-1.x-data-mixture", "dataset:cognitivecomputations/dolphin-2.9.3", "base_model:Qwen/Qwen2.5-32B", "base_model:finetune:Qwen/Qwen2.5-32B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
null
151,643
silu
5,120
0.02
27,648
131,072
qwen2
40
64
8
0.00001
1,000,000
null
false
bfloat16
4.45.2
false
152,064
null
0
null
null
null
null
amd/AMD-OLMo-1B-SFT-DPO
null
null
"2024-10-31T20:29:17Z"
null
null
883
null
null
null
null
15
transformers
[ "transformers", "safetensors", "olmo", "text-generation", "conversational", "dataset:allenai/dolma", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "OlmoForCausalLM" ]
null
50,279
silu
2,048
0.02
8,192
2,048
olmo
16
16
16
null
10,000
null
true
float32
4.40.2
true
50,304
false
0
null
null
null
null
THU-KEG/Mistral-Crab-SFT
null
null
"2024-11-01T06:35:13Z"
null
null
28
null
null
null
null
4
null
[ "safetensors", "mistral", "alignment-handbook", "generated_from_trainer", "text-generation", "conversational", "en", "arxiv:2410.24175", "license:apache-2.0", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.44.0
true
32,768
null
0
128
null
null
null
fblgit/miniclaus-qw1.5B-UNAMGS
null
null
"2024-11-01T07:07:02Z"
null
null
4,865
null
null
null
null
4
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "conversational", "en", "dataset:Magpie-Align/Magpie-Pro-MT-300K-v0.1", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-1.5B-Instruct", "base_model:finetune:Qwen/Qwen2.5-1.5B-Instruct", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
1,536
0.02
8,960
131,072
qwen2
12
28
2
0.000001
1,000,000
null
true
bfloat16
4.45.2
true
151,936
null
0
null
null
null
null
MaziyarPanahi/SmolLM2-1.7B-Instruct-GGUF
null
null
"2024-11-01T16:40:28Z"
null
null
635,330
null
null
null
null
3
null
[ "gguf", "mistral", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:HuggingFaceTB/SmolLM2-1.7B-Instruct", "base_model:quantized:HuggingFaceTB/SmolLM2-1.7B-Instruct", "region:us", "imatrix", "conversational" ]
text-generation
null
null
3
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
allura-org/TQ2.5-14B-Sugarquill-v1
null
null
"2024-11-06T10:39:16Z"
null
null
51
null
null
null
null
3
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "en", "dataset:Mielikki/Erebus-87k", "dataset:allura-org/r_shortstories_24k", "base_model:arcee-ai/SuperNova-Medius", "base_model:finetune:arcee-ai/SuperNova-Medius", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
null
151,645
silu
5,120
0.02
13,824
131,072
qwen2
40
48
8
0.00001
1,000,000
null
false
bfloat16
4.45.1
false
152,064
null
0
null
null
null
null
FallenMerick/MN-Chunky-Lotus-12B
null
null
"2024-11-07T23:42:28Z"
null
null
33
null
null
null
null
3
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "storywriting", "text adventure", "creative", "story", "writing", "fiction", "roleplaying", "rp", "mergekit", "merge", "en", "arxiv:2306.01708", "base_model:Epiculous/Violet_Twilight-v0.2", "base_model:merge:Epiculous/Violet_Twilight-v0.2", "base_model:TheDrummer/Rocinante-12B-v1.1", "base_model:merge:TheDrummer/Rocinante-12B-v1.1", "base_model:flammenai/Mahou-1.5-mistral-nemo-12B", "base_model:merge:flammenai/Mahou-1.5-mistral-nemo-12B", "base_model:nbeerbower/mistral-nemo-gutenberg-12B-v4", "base_model:merge:nbeerbower/mistral-nemo-gutenberg-12B-v4", "license:cc-by-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
131,072
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.46.0
true
131,072
null
0
128
null
null
null
theo77186/Qwen2.5-Coder-7B-Instruct-20241106
null
null
"2024-11-08T10:00:12Z"
null
null
6
null
null
null
null
3
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "code", "codeqwen", "chat", "qwen", "qwen-coder", "conversational", "en", "arxiv:2409.12186", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-Coder-7B", "base_model:finetune:Qwen/Qwen2.5-Coder-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
3
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
131,072
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
EleutherAI/gpt-neo-2.7B
null
null
"2022-03-02T23:29:04Z"
null
null
191,834
null
null
null
null
438
transformers
[ "transformers", "pytorch", "jax", "rust", "safetensors", "gpt_neo", "text-generation", "text generation", "causal-lm", "en", "dataset:EleutherAI/pile", "arxiv:2101.00027", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "GPTNeoForCausalLM" ]
50,256
50,256
null
2,560
0.02
null
2,048
gpt_neo
null
null
null
null
null
null
null
null
4.9.0.dev0
true
50,257
null
0
null
null
null
null
Gustavosta/MagicPrompt-Stable-Diffusion
null
null
"2022-09-17T22:34:07Z"
null
null
321,164
null
null
null
null
704
transformers
[ "transformers", "pytorch", "coreml", "safetensors", "gpt2", "text-generation", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.23.0.dev0
true
50,257
null
null
null
null
null
null
bigscience/bloomz-560m
null
null
"2022-10-08T16:14:42Z"
null
null
15,791,704
null
null
null
null
107
transformers
[ "transformers", "pytorch", "tensorboard", "safetensors", "bloom", "text-generation", "ak", "ar", "as", "bm", "bn", "ca", "code", "en", "es", "eu", "fon", "fr", "gu", "hi", "id", "ig", "ki", "kn", "lg", "ln", "ml", "mr", "ne", "nso", "ny", "or", "pa", "pt", "rn", "rw", "sn", "st", "sw", "ta", "te", "tn", "ts", "tum", "tw", "ur", "vi", "wo", "xh", "yo", "zh", "zu", "dataset:bigscience/xP3", "arxiv:2211.01786", "license:bigscience-bloom-rail-1.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "BloomForCausalLM" ]
1
2
null
null
0.02
null
null
bloom
16
null
null
null
null
null
null
null
4.20.0
true
250,880
null
0
null
null
1
null
PygmalionAI/pygmalion-6b
null
null
"2023-01-07T18:43:33Z"
null
null
3,219
null
null
null
null
730
transformers
[ "transformers", "pytorch", "tensorboard", "gptj", "text-generation", "text generation", "conversational", "en", "license:creativeml-openrail-m", "autotrain_compatible", "region:us" ]
text-generation
null
null
2
[ "GPTJForCausalLM" ]
50,256
50,256
null
null
0.02
null
null
gptj
null
null
null
null
null
null
false
float16
4.26.0.dev0
true
50,400
null
null
null
null
null
null
medalpaca/medalpaca-7b
null
null
"2023-03-29T17:54:49Z"
null
null
4,771
null
null
null
null
71
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "medical", "en", "arxiv:2303.14070", "license:cc", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
0
1
silu
4,096
0.02
11,008
null
llama
32
32
null
0.000001
null
null
false
float32
4.28.0.dev0
true
32,001
null
null
null
null
null
null
tiiuae/falcon-7b-instruct
null
null
"2023-04-25T06:21:01Z"
null
null
169,796
null
null
null
null
919
transformers
[ "transformers", "pytorch", "coreml", "safetensors", "falcon", "text-generation", "conversational", "custom_code", "en", "dataset:tiiuae/falcon-refinedweb", "arxiv:2205.14135", "arxiv:1911.02150", "arxiv:2005.14165", "arxiv:2104.09864", "arxiv:2306.01116", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "FalconForCausalLM" ]
11
11
null
4,544
0.02
null
null
falcon
71
32
null
null
null
null
null
bfloat16
4.27.4
true
65,024
null
0
null
null
null
null
replit/replit-code-v1-3b
null
null
"2023-04-28T19:00:45Z"
null
null
379
null
null
null
null
722
transformers
[ "transformers", "pytorch", "mpt", "text-generation", "code", "custom_code", "dataset:bigcode/the-stack-dedup", "arxiv:2211.15533", "arxiv:2205.14135", "arxiv:2108.12409", "arxiv:2302.06675", "license:cc-by-sa-4.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MPTForCausalLM" ]
null
1
null
null
null
null
null
mpt
null
null
null
null
null
null
null
float32
4.28.1
true
32,768
null
null
null
null
null
null
alibidaran/medical_transcription_generator
null
null
"2023-06-03T10:16:09Z"
null
null
961
null
null
null
null
4
transformers
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "medical", "en", "base_model:openai-community/gpt2", "base_model:finetune:openai-community/gpt2", "doi:10.57967/hf/1204", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.29.2
true
50,257
null
null
null
null
null
null
TheBloke/Llama-2-7B-Chat-GPTQ
null
null
"2023-07-18T17:38:15Z"
null
null
136,936
null
null
null
null
261
transformers
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-2", "en", "arxiv:2307.09288", "base_model:meta-llama/Llama-2-7b-chat-hf", "base_model:quantized:meta-llama/Llama-2-7b-chat-hf", "license:llama2", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
null
0.000001
null
null
false
float16
4.30.0.dev0
true
32,000
null
null
null
null
null
null
lmsys/vicuna-7b-v1.5
null
null
"2023-07-29T04:42:33Z"
null
null
389,147
null
null
null
null
304
transformers
[ "transformers", "pytorch", "llama", "text-generation", "arxiv:2307.09288", "arxiv:2306.05685", "license:llama2", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
null
null
false
float16
4.31.0
true
32,000
null
null
null
null
1
null
TheBloke/Llama-2-7B-Chat-GGUF
null
null
"2023-09-04T16:38:41Z"
null
null
90,148
null
null
null
null
437
transformers
[ "transformers", "gguf", "llama", "facebook", "meta", "pytorch", "llama-2", "text-generation", "en", "arxiv:2307.09288", "base_model:meta-llama/Llama-2-7b-chat-hf", "base_model:quantized:meta-llama/Llama-2-7b-chat-hf", "license:llama2", "region:us" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
EleutherAI/llemma_7b
null
null
"2023-09-12T22:09:33Z"
null
null
7,883
null
null
null
null
98
transformers
[ "transformers", "pytorch", "llama", "text-generation", "math", "reasoning", "en", "dataset:EleutherAI/proof-pile-2", "dataset:open-web-math/open-web-math", "arxiv:2310.10631", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
null
null
false
bfloat16
4.31.0
true
32,016
null
null
null
null
1
null
thesephist/contra-bottleneck-t5-xl-wikipedia
null
null
"2023-09-30T21:48:13Z"
null
null
179
null
null
null
null
16
transformers
[ "transformers", "pytorch", "t5", "text-generation", "custom_code", "en", "dataset:wikipedia", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "BottleneckT5LMWithPerturb" ]
null
1
null
null
null
null
null
t5
null
null
null
null
null
null
false
float32
4.33.3
true
32,128
null
null
null
null
null
null
liuhaotian/llava-v1.5-13b
null
null
"2023-10-05T18:27:40Z"
null
null
464,536
null
null
null
null
481
transformers
[ "transformers", "pytorch", "llava", "text-generation", "image-text-to-text", "autotrain_compatible", "region:us" ]
image-text-to-text
null
null
2
[ "LlavaLlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
4,096
llava
40
40
40
0.00001
null
null
false
float16
4.31.0
true
32,000
null
null
null
null
1
null
HuggingFaceH4/zephyr-7b-alpha
null
null
"2023-10-09T08:45:10Z"
null
null
34,183
null
null
null
null
1,099
transformers
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "generated_from_trainer", "conversational", "en", "dataset:stingning/ultrachat", "dataset:openbmb/UltraFeedback", "arxiv:2305.18290", "arxiv:2310.16944", "arxiv:2305.14233", "arxiv:2310.01377", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.34.0
true
32,000
null
null
null
null
null
null
TheBloke/CausalLM-14B-GGUF
null
null
"2023-10-22T16:28:29Z"
null
null
2,983
null
null
null
null
151
transformers
[ "transformers", "gguf", "llama", "llama2", "qwen", "text-generation", "en", "zh", "dataset:JosephusCheung/GuanacoDataset", "dataset:Open-Orca/OpenOrca", "dataset:stingning/ultrachat", "dataset:meta-math/MetaMathQA", "dataset:liuhaotian/LLaVA-Instruct-150K", "dataset:jondurbin/airoboros-3.1", "dataset:WizardLM/WizardLM_evol_instruct_V2_196k", "dataset:RyokoAI/ShareGPT52K", "dataset:RyokoAI/Fandom23K", "dataset:milashkaarshif/MoeGirlPedia_wikitext_raw_archive", "dataset:wikipedia", "dataset:wiki_lingua", "dataset:fnlp/moss-003-sft-data", "dataset:garage-bAInd/Open-Platypus", "dataset:LDJnr/Puffin", "dataset:openbmb/llava_zh", "dataset:BAAI/COIG", "dataset:TigerResearch/tigerbot-zhihu-zh-10k", "dataset:liwu/MNBVC", "dataset:teknium/openhermes", "base_model:CausalLM/14B", "base_model:quantized:CausalLM/14B", "license:wtfpl", "region:us" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
01-ai/Yi-34B
null
null
"2023-11-01T07:03:50Z"
null
null
5,505
null
null
null
null
1,285
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "arxiv:2403.04652", "arxiv:2311.16502", "arxiv:2401.11944", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
7,168
0.02
20,480
4,096
llama
56
60
8
0.00001
5,000,000
null
false
bfloat16
4.34.0
true
64,000
null
null
null
null
1
null
maritaca-ai/sabia-7b
null
null
"2023-11-08T22:13:36Z"
null
null
1,104
null
null
null
null
104
transformers
[ "transformers", "safetensors", "llama", "text-generation", "pt", "arxiv:2304.07880", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
32
0.000001
10,000
null
false
bfloat16
4.37.2
true
32,000
false
0
null
null
1
null
reglab-rrc/mistral-rrc
null
null
"2023-11-23T06:42:21Z"
null
null
89
null
null
null
null
4
transformers
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "legal", "housing", "covenants", "property", "deed", "racial-covenant", "en", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float16
4.34.0
true
32,000
null
null
null
null
null
null
bigcode/starcoder2-3b
null
null
"2023-11-29T15:22:51Z"
null
null
189,981
null
null
null
null
146
transformers
[ "transformers", "safetensors", "starcoder2", "text-generation", "code", "dataset:bigcode/the-stack-v2-train", "arxiv:2305.13245", "arxiv:2205.14135", "arxiv:2004.05150", "arxiv:2207.14255", "arxiv:2402.19173", "license:bigcode-openrail-m", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Starcoder2ForCausalLM" ]
0
0
gelu_pytorch_tanh
3,072
0.018042
12,288
16,384
starcoder2
24
30
2
null
999,999.442036
4,096
null
null
4.37.0.dev0
true
49,152
null
0.1
null
null
null
null
TheBloke/Mistral-7B-Instruct-v0.2-GGUF
null
null
"2023-12-11T22:18:46Z"
null
null
131,340
null
null
null
null
398
transformers
[ "transformers", "gguf", "mistral", "finetuned", "text-generation", "arxiv:2310.06825", "base_model:mistralai/Mistral-7B-Instruct-v0.2", "base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2", "license:apache-2.0", "region:us", "conversational" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
upstage/SOLAR-10.7B-Instruct-v1.0
null
null
"2023-12-12T12:39:22Z"
null
null
130,262
null
null
null
null
614
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "en", "dataset:c-s-ale/alpaca-gpt4-data", "dataset:Open-Orca/OpenOrca", "dataset:Intel/orca_dpo_pairs", "dataset:allenai/ultrafeedback_binarized_cleaned", "arxiv:2312.15166", "arxiv:2403.19270", "base_model:upstage/SOLAR-10.7B-v1.0", "base_model:finetune:upstage/SOLAR-10.7B-v1.0", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
14,336
4,096
llama
32
48
8
0.00001
10,000
null
false
float16
4.35.0
true
32,000
false
null
null
null
1
null
cognitivecomputations/dolphin-2.5-mixtral-8x7b
null
null
"2023-12-14T00:45:47Z"
null
null
23,529
null
null
null
null
1,208
transformers
[ "transformers", "pytorch", "safetensors", "mixtral", "text-generation", "conversational", "en", "dataset:ehartford/dolphin", "dataset:jondurbin/airoboros-2.2.1", "dataset:ehartford/dolphin-coder", "dataset:migtissera/Synthia-v1.3", "dataset:teknium/openhermes", "dataset:ise-uiuc/Magicoder-OSS-Instruct-75K", "dataset:ise-uiuc/Magicoder-Evol-Instruct-110K", "dataset:LDJnr/Pure-Dove", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MixtralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.36.0.dev0
false
32,002
null
0
null
null
null
null
cognitivecomputations/dolphin-2.6-mixtral-8x7b
null
null
"2023-12-21T22:34:09Z"
null
null
4,920
null
null
null
null
195
transformers
[ "transformers", "pytorch", "mixtral", "text-generation", "conversational", "en", "dataset:ehartford/dolphin", "dataset:jondurbin/airoboros-2.2.1", "dataset:ehartford/dolphin-coder", "dataset:teknium/openhermes", "dataset:ise-uiuc/Magicoder-OSS-Instruct-75K", "dataset:ise-uiuc/Magicoder-Evol-Instruct-110K", "dataset:LDJnr/Capybara", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MixtralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mixtral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.37.0.dev0
false
32,002
null
0
null
null
null
null
MaralGPT/Maral-7B-alpha-1
null
null
"2023-12-24T17:10:55Z"
null
null
616
null
null
null
null
50
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "en", "fa", "dataset:sinarashidi/alpaca-persian", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.36.2
true
32,000
null
0
null
null
null
null
Vikhrmodels/Vikhr-7b-0.1
null
null
"2024-01-02T11:14:26Z"
null
null
280
null
null
null
null
53
transformers
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "ru", "en", "dataset:IlyaGusev/habr", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float16
4.37.2
true
40,000
false
0
null
null
1
null
NousResearch/Genstruct-7B
null
null
"2024-01-05T21:56:01Z"
null
null
231
null
null
null
null
367
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "Mistral", "instruct", "finetune", "synthetic", "conversational", "en", "arxiv:2310.04484", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.36.2
false
32,001
null
0
null
null
null
null
TencentARC/LLaMA-Pro-8B-Instruct
null
null
"2024-01-06T02:12:26Z"
null
null
1,005
null
null
null
null
62
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "conversational", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
40
32
0.00001
10,000
null
false
bfloat16
4.33.3
false
32,000
null
null
null
null
1
null
liuhaotian/llava-v1.6-mistral-7b
null
null
"2024-01-31T04:20:00Z"
null
null
56,875
null
null
null
null
225
transformers
[ "transformers", "safetensors", "llava_mistral", "text-generation", "image-text-to-text", "conversational", "license:apache-2.0", "autotrain_compatible", "region:us" ]
image-text-to-text
null
null
2
[ "LlavaMistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
llava_mistral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.36.2
true
32,000
null
0
null
null
null
null
Himitsui/Kaiju-11B
null
null
"2024-02-13T12:33:38Z"
null
null
109
null
null
null
null
14
transformers
[ "transformers", "safetensors", "llama", "text-generation", "en", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
14,336
4,096
llama
32
48
8
0.00001
10,000
null
false
float16
4.37.2
true
32,000
false
0
null
null
1
null
prometheus-eval/prometheus-7b-v2.0
null
null
"2024-02-13T17:18:13Z"
null
null
6,567
null
null
null
null
77
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "text2text-generation", "en", "dataset:prometheus-eval/Feedback-Collection", "dataset:prometheus-eval/Preference-Collection", "arxiv:2405.01535", "arxiv:2310.08491", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.35.2
true
32,000
null
0
null
null
null
null
NousResearch/Nous-Hermes-2-Mistral-7B-DPO
null
null
"2024-02-18T10:50:26Z"
null
null
4,005
null
null
null
null
168
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "Mistral", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "synthetic data", "distillation", "conversational", "en", "dataset:teknium/OpenHermes-2.5", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.37.0.dev0
false
32,002
null
0
null
null
null
null
abideen/gemma-7b-openhermes
null
null
"2024-02-21T23:03:54Z"
null
null
11
null
null
null
null
11
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "generated_from_trainer", "axolotl", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "distillation", "conversational", "en", "dataset:mlabonne/chatml-OpenHermes2.5-dpo-binarized-alpha", "base_model:google/gemma-7b-it", "base_model:finetune:google/gemma-7b-it", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "GemmaForCausalLM" ]
2
1
gelu
3,072
0.02
24,576
8,192
gemma
16
28
16
0.000001
10,000
null
null
float16
4.39.0.dev0
true
256,000
false
0
256
null
null
null
ND911/Fraken-Maid-TW-K-Slerp
null
null
"2024-02-26T00:02:51Z"
null
null
14
null
null
null
null
2
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "SanjiWatsuki/Kunoichi-7B", "ND911/Fraken-Maid-TW-Slerp", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
8,192
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.37.2
true
32,000
null
0
null
null
null
null
1bitLLM/bitnet_b1_58-large
null
null
"2024-03-29T11:10:00Z"
null
null
19,392
null
null
null
null
72
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:2402.17764", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "BitnetForCausalLM" ]
1
2
silu
1,536
0.02
4,096
2,048
llama
16
24
16
0.00001
10,000
null
true
float16
4.39.0
true
32,002
false
0
null
null
1
null
Chat2DB/Chat2DB-SQL-7B
null
null
"2024-04-01T07:25:03Z"
null
null
697
null
null
null
null
38
transformers
[ "transformers", "pytorch", "llama", "text-generation", "zh", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
16,384
llama
32
32
32
0.00001
1,000,000
null
false
bfloat16
4.33.2
true
32,016
null
null
null
null
1
null
apple/OpenELM-270M
null
null
"2024-04-12T21:42:49Z"
null
null
1,708
null
null
null
null
70
transformers
[ "transformers", "safetensors", "openelm", "text-generation", "custom_code", "arxiv:2404.14619", "license:other", "autotrain_compatible", "region:us" ]
text-generation
null
null
2
[ "OpenELMForCausalLM" ]
1
2
null
null
0.02
null
null
openelm
null
null
null
null
null
null
null
float32
4.39.3
true
32,000
null
null
64
null
null
null
allenai/OLMo-7B-0424
null
null
"2024-04-15T21:03:34Z"
null
null
137
null
null
null
null
45
transformers
[ "transformers", "pytorch", "olmo", "text-generation", "en", "dataset:allenai/dolma", "arxiv:2402.00838", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "OLMoForCausalLM" ]
null
50,279
null
null
null
null
null
olmo
null
null
null
null
null
null
null
null
4.39.3
true
50,280
null
0
null
null
null
null
dreamgen/WizardLM-2-7B
null
null
"2024-04-16T05:08:27Z"
null
null
14,472
null
null
null
null
37
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "arxiv:2304.12244", "arxiv:2306.08568", "arxiv:2308.09583", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.36.2
false
32,000
null
0
null
null
null
null
lmms-lab/LLaVA-NeXT-Video-7B
null
null
"2024-04-16T14:08:19Z"
null
null
447
null
null
null
null
41
transformers
[ "transformers", "safetensors", "llava", "text-generation", "video-text-to-text", "license:llama2", "autotrain_compatible", "region:us" ]
video-text-to-text
null
null
2
[ "LlavaLlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llava
32
32
32
0.00001
10,000
null
false
bfloat16
4.39.0.dev0
false
32,000
false
0
null
null
1
null
NousResearch/Meta-Llama-3-8B-Instruct
null
null
"2024-04-18T16:55:56Z"
null
null
62,025
null
null
null
null
84
transformers
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-3", "conversational", "en", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.0.dev0
true
128,256
false
0
null
null
1
null
sethuiyer/Medichat-Llama3-8B
null
null
"2024-04-22T05:07:13Z"
null
null
95,455
null
null
null
null
18
transformers
[ "transformers", "safetensors", "llama", "text-generation", "mergekit", "merge", "medical", "conversational", "en", "dataset:mlabonne/orpo-dpo-mix-40k", "dataset:Open-Orca/SlimOrca-Dedup", "dataset:jondurbin/airoboros-3.2", "dataset:microsoft/orca-math-word-problems-200k", "dataset:m-a-p/Code-Feedback", "dataset:MaziyarPanahi/WizardLM_evol_instruct_V2_196k", "dataset:ruslanmv/ai-medical-chatbot", "base_model:Locutusque/llama-3-neural-chat-v1-8b", "base_model:merge:Locutusque/llama-3-neural-chat-v1-8b", "base_model:Undi95/Llama-3-Unholy-8B", "base_model:merge:Undi95/Llama-3-Unholy-8B", "base_model:ruslanmv/Medical-Llama3-8B", "base_model:merge:ruslanmv/Medical-Llama3-8B", "license:other", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.0
true
128,256
false
0
null
null
1
null
beomi/Llama-3-Open-Ko-8B
null
null
"2024-04-22T09:53:13Z"
null
null
50,458
null
null
null
null
130
transformers
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-3", "llama-3-ko", "conversational", "en", "ko", "arxiv:2310.04799", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.1
true
128,256
false
0
null
null
1
null
Orenguteng/Llama-3-8B-Lexi-Uncensored
null
null
"2024-04-23T21:14:40Z"
null
null
501,212
null
null
null
null
170
transformers
[ "transformers", "safetensors", "llama", "text-generation", "uncensored", "llama3", "instruct", "open", "conversational", "license:llama3", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.39.3
true
128,256
false
0
null
null
1
null
aaditya/Llama3-OpenBioLLM-70B
null
null
"2024-04-24T02:21:12Z"
null
null
7,055
null
null
null
null
349
transformers
[ "transformers", "pytorch", "llama", "text-generation", "llama-3", "Mixtral", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "distillation", "heathcare", "medical", "clinical", "med", "lifescience", "Pharmaceutical", "Pharma", "conversational", "en", "arxiv:2305.18290", "arxiv:2303.13375", "arxiv:2212.13138", "arxiv:2305.09617", "arxiv:2402.07023", "base_model:meta-llama/Meta-Llama-3-70B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-70B-Instruct", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
8,192
0.02
28,672
8,192
llama
64
80
8
0.00001
500,000
null
false
bfloat16
4.40.0.dev0
false
128,256
false
0
null
null
1
null
teddylee777/Llama-3-Open-Ko-8B-gguf
null
null
"2024-04-24T17:14:54Z"
null
null
3,848
null
null
null
null
38
transformers
[ "transformers", "gguf", "llama", "text-generation", "facebook", "meta", "pytorch", "llama-3", "llama-3-ko", "conversational", "en", "ko", "arxiv:2310.04799", "license:llama3", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.1
true
128,256
false
0
null
null
1
null
HPAI-BSC/Llama3-Aloe-8B-Alpha
null
null
"2024-04-26T16:43:36Z"
null
null
4,670
null
null
null
null
52
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "biology", "medical", "question-answering", "en", "dataset:argilla/dpo-mix-7k", "dataset:nvidia/HelpSteer", "dataset:jondurbin/airoboros-3.2", "dataset:hkust-nlp/deita-10k-v0", "dataset:LDJnr/Capybara", "dataset:HPAI-BSC/CareQA", "dataset:GBaker/MedQA-USMLE-4-options", "dataset:lukaemon/mmlu", "dataset:bigbio/pubmed_qa", "dataset:openlifescienceai/medmcqa", "dataset:bigbio/med_qa", "dataset:HPAI-BSC/better-safe-than-sorry", "dataset:HPAI-BSC/pubmedqa-cot", "dataset:HPAI-BSC/medmcqa-cot", "dataset:HPAI-BSC/medqa-cot", "arxiv:2405.01886", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
question-answering
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.0.dev0
false
128,256
false
0
null
null
1
null
nvidia/Llama3-ChatQA-1.5-8B
null
null
"2024-04-28T21:32:50Z"
null
null
14,764
null
null
null
null
545
transformers
[ "transformers", "safetensors", "llama", "text-generation", "nvidia", "chatqa-1.5", "chatqa", "llama-3", "pytorch", "conversational", "en", "arxiv:2401.10225", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
float16
4.40.0.dev0
true
128,256
false
0
null
null
1
null
failspy/llama-3-70B-Instruct-abliterated
null
null
"2024-05-07T00:50:02Z"
null
null
3,396
null
null
null
null
94
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
8,192
0.02
28,672
8,192
llama
64
80
8
0.00001
500,000
null
false
bfloat16
4.40.1
true
128,256
false
0
null
null
1
null
defog/llama-3-sqlcoder-8b
null
null
"2024-05-09T15:58:49Z"
null
null
143,101
null
null
null
null
191
transformers
[ "transformers", "safetensors", "llama", "text-generation", "code", "conversational", "license:cc-by-sa-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.0
true
128,256
false
0
null
null
1
null
Miwa-Keita/zenz-v1
null
null
"2024-05-12T15:48:46Z"
null
null
18
null
null
null
null
7
transformers
[ "transformers", "gguf", "gpt2", "text-generation", "japanese input", "kana kanji conversion", "ja", "license:cc-by-sa-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "GPT2LMHeadModel" ]
1
2
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.30.0
true
6,000
null
null
null
null
null
null
Henrychur/MMed-Llama-3-8B
null
null
"2024-05-22T09:17:24Z"
null
null
1,353
null
null
null
null
20
transformers
[ "transformers", "safetensors", "llama", "text-generation", "medical", "conversational", "en", "zh", "ja", "fr", "ru", "es", "dataset:Henrychur/MMedC", "arxiv:2402.13963", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
float32
4.39.2
true
128,256
false
0
null
null
1
null
Qwen/Qwen2-72B-Instruct
null
null
"2024-05-28T03:48:49Z"
null
null
54,116
null
null
null
null
675
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "base_model:Qwen/Qwen2-72B", "base_model:finetune:Qwen/Qwen2-72B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
8,192
0.02
29,568
32,768
qwen2
64
80
8
0.000001
1,000,000
131,072
false
bfloat16
4.40.1
true
152,064
null
0
null
null
null
null
failspy/Meta-Llama-3-70B-Instruct-abliterated-v3.5
null
null
"2024-05-28T17:51:28Z"
null
null
3,786
null
null
null
null
37
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,001
silu
8,192
0.02
28,672
8,192
llama
64
80
8
0.00001
500,000
null
false
bfloat16
4.40.0.dev0
true
128,256
false
0
null
null
1
null
numind/NuExtract
null
null
"2024-05-31T09:53:13Z"
null
null
2,091
null
null
null
null
203
transformers
[ "transformers", "safetensors", "phi3", "text-generation", "conversational", "custom_code", "en", "base_model:microsoft/Phi-3-mini-4k-instruct", "base_model:finetune:microsoft/Phi-3-mini-4k-instruct", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Phi3ForCausalLM" ]
1
32,000
silu
3,072
0.02
8,192
4,096
phi3
32
32
32
0.00001
10,000
2,047
false
float32
4.42.0.dev0
true
32,064
null
0
null
null
null
null