id
stringlengths
7
117
author
stringclasses
6 values
sha
null
created_at
unknown
last_modified
null
disabled
null
downloads
int64
0
18.6M
downloads_all_time
null
gated
bool
1 class
gguf
null
inference
null
likes
int64
0
4.77k
library_name
stringclasses
36 values
tags
sequencelengths
1
430
pipeline_tag
stringclasses
32 values
mask_token
null
model_index
null
trending_score
int64
0
132
architectures
sequencelengths
1
5
bos_token_id
int64
-1
256k
eos_token_id
int64
-1
256k
hidden_act
stringclasses
15 values
hidden_size
int64
1
20.5k
initializer_range
float64
0
1
intermediate_size
int64
1
98.3k
max_position_embeddings
int64
8
1.05M
model_type
stringclasses
530 values
num_attention_heads
int64
1
5k
num_hidden_layers
int64
-1
8.93k
num_key_value_heads
int64
1
160
rms_norm_eps
float64
0
7
rope_theta
float64
1k
1,000B
sliding_window
int64
0
262k
tie_word_embeddings
bool
2 classes
torch_dtype
stringclasses
8 values
transformers_version
stringclasses
207 values
use_cache
bool
2 classes
vocab_size
int64
-1
5.03M
attention_bias
bool
2 classes
attention_dropout
float64
0
0.5
head_dim
int64
2
256
mlp_bias
bool
2 classes
pretraining_tp
int64
0
8
rope_scaling
dict
fireworks-ai/llama-3-firefunction-v2
null
null
"2024-06-05T04:44:29Z"
null
null
258
null
null
null
null
136
transformers
[ "transformers", "safetensors", "llama", "text-generation", "function-calling", "conversational", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
8,192
0.02
28,672
8,192
llama
64
80
8
0.00001
500,000
null
false
float16
4.41.2
true
128,256
false
0
null
false
1
null
Nitral-AI/Hathor_Stable-v0.2-L3-8B
null
null
"2024-06-09T18:53:59Z"
null
null
1,055
null
null
null
null
57
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "en", "license:other", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.38.1
false
128,256
false
0
null
false
1
null
microsoft/Florence-2-base
null
null
"2024-06-15T00:57:24Z"
null
null
611,009
null
null
null
null
172
transformers
[ "transformers", "pytorch", "florence2", "text-generation", "vision", "image-text-to-text", "custom_code", "arxiv:2311.06242", "license:mit", "autotrain_compatible", "region:us" ]
image-text-to-text
null
null
2
[ "Florence2ForConditionalGeneration" ]
0
2
null
null
null
null
null
florence2
null
null
null
null
null
null
null
float16
4.41.0.dev0
null
51,289
null
null
null
null
null
null
instruction-pretrain/instruction-synthesizer
null
null
"2024-06-18T06:30:53Z"
null
null
263
null
null
null
null
72
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "en", "dataset:instruction-pretrain/ft-instruction-synthesizer-collection", "arxiv:2406.14491", "arxiv:2309.09530", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float16
4.34.0.dev0
true
32,000
null
null
null
null
null
null
elyza/Llama-3-ELYZA-JP-8B
null
null
"2024-06-25T06:32:13Z"
null
null
48,916
null
null
null
null
76
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "ja", "en", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.41.2
true
128,256
false
0
null
false
1
null
cognitivecomputations/dolphin-vision-72b
null
null
"2024-06-28T14:49:21Z"
null
null
146
null
null
null
null
110
transformers
[ "transformers", "safetensors", "llava-qwen2", "text-generation", "generated_from_trainer", "axolotl", "conversational", "custom_code", "dataset:cognitivecomputations/Dolphin-2.9", "dataset:teknium/OpenHermes-2.5", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:cognitivecomputations/dolphin-coder", "dataset:cognitivecomputations/samantha-data", "dataset:microsoft/orca-math-word-problems-200k", "dataset:Locutusque/function-calling-chatml", "dataset:internlm/Agent-FLAN", "base_model:Qwen/Qwen2-72B", "base_model:finetune:Qwen/Qwen2-72B", "license:other", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "BunnyQwenForCausalLM" ]
null
151,645
silu
8,192
0.02
29,568
131,072
llava-qwen2
64
80
8
0.00001
1,000,000
4,096
false
bfloat16
4.41.2
true
152,064
null
0
null
null
null
null
lmms-lab/llava-onevision-qwen2-7b-ov
null
null
"2024-06-29T04:16:21Z"
null
null
116,214
null
null
null
null
35
transformers
[ "transformers", "safetensors", "llava", "text-generation", "multimodal", "conversational", "en", "zh", "dataset:lmms-lab/LLaVA-OneVision-Data", "arxiv:2408.03326", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlavaQwenForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
llava
28
28
4
0.000001
1,000,000
131,072
false
bfloat16
4.40.0.dev0
true
152,064
null
0
null
null
null
null
PatronusAI/Llama-3-Patronus-Lynx-70B-Instruct
null
null
"2024-07-03T18:34:38Z"
null
null
6,193
null
null
null
null
26
transformers
[ "transformers", "safetensors", "llama", "text-generation", "pytorch", "Lynx", "Patronus AI", "evaluation", "hallucination-detection", "conversational", "en", "arxiv:2407.08488", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
8,192
0.02
28,672
8,192
llama
64
80
8
0.00001
500,000
null
false
float16
4.42.3
false
128,256
false
0
null
false
1
null
TheDrummer/Tiger-Gemma-9B-v1
null
null
"2024-07-12T17:32:18Z"
null
null
358
null
null
null
null
34
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.42.3
false
256,000
false
0
256
null
null
null
HuggingFaceTB/SmolLM-1.7B
null
null
"2024-07-14T15:56:33Z"
null
null
9,987
null
null
null
null
160
transformers
[ "transformers", "onnx", "safetensors", "llama", "text-generation", "en", "dataset:HuggingFaceTB/smollm-corpus", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
0
0
silu
2,048
0.02
8,192
2,048
llama
32
24
32
0.00001
10,000
null
true
float32
4.39.3
true
49,152
false
0
null
null
1
null
HuggingFaceTB/SmolLM-135M
null
null
"2024-07-14T21:45:18Z"
null
null
34,824
null
null
null
null
171
transformers
[ "transformers", "onnx", "safetensors", "llama", "text-generation", "en", "dataset:HuggingFaceTB/smollm-corpus", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
0
0
silu
576
0.02
1,536
2,048
llama
9
30
3
0.00001
10,000
null
true
bfloat16
4.41.2
true
49,152
false
0
null
false
1
null
HuggingFaceTB/SmolLM-135M-Instruct
null
null
"2024-07-15T23:02:44Z"
null
null
25,090
null
null
null
null
97
transformers
[ "transformers", "onnx", "safetensors", "llama", "text-generation", "alignment-handbook", "trl", "sft", "conversational", "en", "dataset:Magpie-Align/Magpie-Pro-300K-Filtered", "dataset:bigcode/self-oss-instruct-sc2-exec-filter-50k", "dataset:teknium/OpenHermes-2.5", "dataset:HuggingFaceTB/everyday-conversations-llama3.1-2k", "base_model:HuggingFaceTB/SmolLM-135M", "base_model:quantized:HuggingFaceTB/SmolLM-135M", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
576
0.02
1,536
2,048
llama
9
30
3
0.00001
10,000
null
true
bfloat16
4.42.3
true
49,152
false
0
null
false
1
null
unsloth/Mistral-Nemo-Instruct-2407-bnb-4bit
null
null
"2024-07-18T17:40:54Z"
null
null
18,767
null
null
null
null
24
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "unsloth", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.44.2
true
131,072
null
0
128
null
null
null
Nexusflow/Athene-70B
null
null
"2024-07-19T07:32:05Z"
null
null
8,587
null
null
null
null
187
transformers
[ "transformers", "safetensors", "llama", "text-generation", "RLHF", "Nexusflow", "Athene", "Chat Model", "conversational", "en", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
8,192
0.02
28,672
8,192
llama
64
80
8
0.00001
500,000
null
false
bfloat16
4.42.4
false
128,256
false
0
null
false
1
null
rinna/llama-3-youko-8b-instruct
null
null
"2024-07-21T14:11:59Z"
null
null
13,283
null
null
null
null
11
transformers
[ "transformers", "safetensors", "llama", "text-generation", "llama-3", "conversational", "ja", "en", "dataset:CohereForAI/aya_dataset", "dataset:kunishou/databricks-dolly-15k-ja", "dataset:kunishou/HelpSteer-35k-ja", "dataset:kunishou/HelpSteer2-20k-ja", "dataset:kunishou/hh-rlhf-49k-ja", "dataset:kunishou/oasst1-chat-44k-ja", "dataset:kunishou/oasst2-chat-68k-ja", "dataset:meta-math/MetaMathQA", "dataset:OpenAssistant/oasst1", "dataset:OpenAssistant/oasst2", "dataset:sahil2801/CodeAlpaca-20k", "arxiv:2404.01657", "arxiv:2310.04799", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:merge:meta-llama/Meta-Llama-3-8B", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:merge:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:rinna/llama-3-youko-8b", "base_model:merge:rinna/llama-3-youko-8b", "license:llama3", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.40.0.dev0
true
128,256
false
0
null
null
1
null
dnhkng/RYS-XLarge
null
null
"2024-07-24T12:20:01Z"
null
null
2,673
null
null
null
null
74
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "license:mit", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
8,192
0.02
29,568
32,768
qwen2
64
86
8
0.000001
1,000,000
131,072
false
bfloat16
4.41.0
true
151,646
null
0
null
null
null
null
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B
null
null
"2024-07-25T05:04:52Z"
null
null
3,298
null
null
null
null
16
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "en", "dataset:BAAI/Infinity-Instruct", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
8,192
0.02
28,672
8,192
llama
64
80
8
0.00001
500,000
null
false
bfloat16
4.40.0.dev0
true
128,256
false
0
null
null
1
null
lemon07r/Gemma-2-Ataraxy-9B
null
null
"2024-08-14T11:33:33Z"
null
null
2,532
null
null
null
null
65
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:nbeerbower/gemma2-gutenberg-9B", "base_model:merge:nbeerbower/gemma2-gutenberg-9B", "base_model:princeton-nlp/gemma-2-9b-it-SimPO", "base_model:merge:princeton-nlp/gemma-2-9b-it-SimPO", "license:gemma", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.43.3
true
256,000
false
0
256
null
null
null
akjindal53244/Llama-3.1-Storm-8B-GGUF
null
null
"2024-08-16T03:12:29Z"
null
null
1,697
null
null
null
null
41
null
[ "gguf", "llama", "llama-3.1", "conversational", "instruction following", "reasoning", "function calling", "text-generation", "en", "de", "fr", "it", "pt", "hi", "es", "th", "arxiv:2406.06623", "arxiv:2311.07911", "arxiv:2311.12022", "arxiv:2406.01574", "arxiv:1803.05457", "arxiv:2310.16049", "arxiv:2210.09261", "arxiv:2109.07958", "base_model:akjindal53244/Llama-3.1-Storm-8B", "base_model:quantized:akjindal53244/Llama-3.1-Storm-8B", "license:llama3.1", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
nvidia/Mistral-NeMo-Minitron-8B-Base
null
null
"2024-08-19T21:39:52Z"
null
null
18,343
null
null
null
null
159
transformers
[ "transformers", "nemo", "safetensors", "mistral", "text-generation", "arxiv:2408.11796", "arxiv:2009.03300", "arxiv:2407.14679", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
11,520
8,192
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.40.2
true
131,072
null
0
128
null
null
null
01-ai/Yi-Coder-9B-Chat
null
null
"2024-08-21T02:11:52Z"
null
null
6,375
null
null
null
null
188
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:2403.04652", "base_model:01-ai/Yi-Coder-9B", "base_model:finetune:01-ai/Yi-Coder-9B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
7
silu
4,096
0.02
11,008
131,072
llama
32
48
4
0.00001
10,000,000
null
false
bfloat16
4.44.0
false
64,000
false
0
null
false
1
null
FluffyKaeloky/Luminum-v0.1-123B
null
null
"2024-08-28T23:44:23Z"
null
null
1,300
null
null
null
null
27
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
12,288
0.02
28,672
131,072
mistral
96
88
8
0.00001
1,000,000
null
false
bfloat16
4.44.2
true
32,768
null
0
128
null
null
null
DavidAU/MN-Instruct-2407-17.5B-BRAINSTORM-20x-FORM-8
null
null
"2024-08-29T07:48:15Z"
null
null
51
null
null
null
null
2
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
59
8
0.00001
1,000,000
null
false
float16
4.43.3
true
131,072
null
0
128
null
null
null
cyberagent/Mistral-Nemo-Japanese-Instruct-2408
null
null
"2024-08-30T03:57:43Z"
null
null
696
null
null
null
null
20
null
[ "safetensors", "mistral", "japanese", "text-generation", "conversational", "ja", "en", "base_model:mistralai/Mistral-Nemo-Base-2407", "base_model:finetune:mistralai/Mistral-Nemo-Base-2407", "license:apache-2.0", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
131,072
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.44.0
false
131,074
null
0
128
null
null
null
speakleash/Bielik-11B-v2.3-Instruct-GGUF
null
null
"2024-09-05T12:49:31Z"
null
null
5,279
null
null
null
null
14
transformers
[ "transformers", "gguf", "mistral", "text-generation", "finetuned", "pl", "base_model:speakleash/Bielik-11B-v2.3-Instruct", "base_model:quantized:speakleash/Bielik-11B-v2.3-Instruct", "license:apache-2.0", "autotrain_compatible", "region:us", "conversational" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
32,001
silu
4,096
0.02
14,336
32,768
mistral
32
50
8
0.00001
1,000,000
null
false
float16
4.44.2
true
32,128
null
0
null
null
null
null
jinaai/reader-lm-0.5b
null
null
"2024-09-06T02:53:09Z"
null
null
1,189
null
null
null
null
121
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "multilingual", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "region:eu" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
896
0.02
4,864
256,000
qwen2
14
24
2
0.000001
2,000,000
null
true
bfloat16
4.43.3
true
151,936
null
0
null
null
null
null
inflatebot/G2-9B-Blackout-R1
null
null
"2024-09-12T00:05:24Z"
null
null
35
null
null
null
null
7
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "arxiv:2212.04089", "base_model:IntervitensInc/gemma-2-9b-chatml", "base_model:merge:IntervitensInc/gemma-2-9b-chatml", "base_model:anthracite-org/magnum-v3-9b-chatml", "base_model:merge:anthracite-org/magnum-v3-9b-chatml", "base_model:crestf411/gemma2-9B-sunfall-v0.5.2", "base_model:merge:crestf411/gemma2-9B-sunfall-v0.5.2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Gemma2ForCausalLM" ]
2
8
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
float32
4.44.1
true
256,000
false
0
256
null
null
null
Qwen/Qwen2.5-1.5B
null
null
"2024-09-15T12:16:50Z"
null
null
55,428
null
null
null
null
37
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "en", "arxiv:2407.10671", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
1,536
0.02
8,960
131,072
qwen2
12
28
2
0.000001
1,000,000
131,072
true
bfloat16
4.40.1
true
151,936
null
0
null
null
null
null
Qwen/Qwen2.5-14B
null
null
"2024-09-15T12:18:18Z"
null
null
20,850
null
null
null
null
32
null
[ "safetensors", "qwen2", "text-generation", "conversational", "en", "arxiv:2407.10671", "license:apache-2.0", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
5,120
0.02
13,824
131,072
qwen2
40
48
8
0.00001
1,000,000
131,072
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
Qwen/Qwen2.5-Math-1.5B-Instruct
null
null
"2024-09-16T11:57:49Z"
null
null
39,905
null
null
null
null
17
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2409.12122", "base_model:Qwen/Qwen2.5-Math-1.5B", "base_model:finetune:Qwen/Qwen2.5-Math-1.5B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
1,536
0.02
8,960
4,096
qwen2
12
28
2
0.000001
10,000
4,096
true
bfloat16
4.43.1
true
151,936
null
0
null
null
null
null
Qwen/Qwen2.5-Math-72B-Instruct
null
null
"2024-09-16T11:58:10Z"
null
null
51,102
null
null
null
null
18
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2409.12122", "base_model:Qwen/Qwen2.5-Math-72B", "base_model:finetune:Qwen/Qwen2.5-Math-72B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
8,192
0.02
29,568
4,096
qwen2
64
80
8
0.000001
10,000
4,096
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
Azazelle/MN-Halide-12b-v1.0
null
null
"2024-09-17T01:00:03Z"
null
null
23
null
null
null
null
5
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "arxiv:2403.19522", "base_model:Epiculous/Azure_Dusk-v0.2", "base_model:merge:Epiculous/Azure_Dusk-v0.2", "base_model:Epiculous/Crimson_Dawn-v0.2", "base_model:merge:Epiculous/Crimson_Dawn-v0.2", "base_model:SillyTilly/mistralai_Mistral-Nemo-Base-2407", "base_model:merge:SillyTilly/mistralai_Mistral-Nemo-Base-2407", "base_model:TheDrummer/Rocinante-12B-v1", "base_model:merge:TheDrummer/Rocinante-12B-v1", "base_model:TheDrummer/Rocinante-12B-v1.1", "base_model:merge:TheDrummer/Rocinante-12B-v1.1", "base_model:UsernameJustAnother/Nemo-12B-Marlin-v8", "base_model:merge:UsernameJustAnother/Nemo-12B-Marlin-v8", "base_model:anthracite-org/magnum-v2-12b", "base_model:merge:anthracite-org/magnum-v2-12b", "base_model:anthracite-org/magnum-v2.5-12b-kto", "base_model:merge:anthracite-org/magnum-v2.5-12b-kto", "base_model:elinas/Chronos-Gold-12B-1.0", "base_model:merge:elinas/Chronos-Gold-12B-1.0", "base_model:jeiku/Aura-NeMo-12B", "base_model:merge:jeiku/Aura-NeMo-12B", "base_model:jtatman/mistral_nemo_12b_reasoning_psychology_lora", "base_model:merge:jtatman/mistral_nemo_12b_reasoning_psychology_lora", "base_model:mpasila/Mistral-freeLiPPA-LoRA-12B", "base_model:merge:mpasila/Mistral-freeLiPPA-LoRA-12B", "base_model:nbeerbower/Lyra4-Gutenberg-12B", "base_model:merge:nbeerbower/Lyra4-Gutenberg-12B", "base_model:nbeerbower/mistral-nemo-bophades-12B", "base_model:merge:nbeerbower/mistral-nemo-bophades-12B", "base_model:nbeerbower/mistral-nemo-cc-12B", "base_model:merge:nbeerbower/mistral-nemo-cc-12B", "base_model:nbeerbower/mistral-nemo-gutenberg-12B-v4", "base_model:merge:nbeerbower/mistral-nemo-gutenberg-12B-v4", "base_model:nbeerbower/mistral-nemo-wissenschaft-12B", "base_model:merge:nbeerbower/mistral-nemo-wissenschaft-12B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
float32
4.43.3
true
131,072
null
0
128
null
null
null
Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8
null
null
"2024-09-17T12:52:12Z"
null
null
8,015
null
null
null
null
8
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-7B-Instruct", "base_model:quantized:Qwen/Qwen2.5-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
131,072
false
float16
4.39.3
true
152,064
null
0
null
null
null
null
Qwen/Qwen2.5-32B-Instruct-GPTQ-Int8
null
null
"2024-09-17T12:53:06Z"
null
null
23,903
null
null
null
null
8
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-32B-Instruct", "base_model:quantized:Qwen/Qwen2.5-32B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
27,648
32,768
qwen2
40
64
8
0.000001
1,000,000
131,072
false
float16
4.39.3
true
152,064
null
0
null
null
null
null
Qwen/Qwen2.5-72B-Instruct-GPTQ-Int4
null
null
"2024-09-17T12:53:17Z"
null
null
28,874
null
null
null
null
22
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-72B-Instruct", "base_model:quantized:Qwen/Qwen2.5-72B-Instruct", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "gptq", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
8,192
0.02
29,696
32,768
qwen2
64
80
8
0.000001
1,000,000
131,072
false
float16
4.39.3
true
152,064
null
0
null
null
null
null
Qwen/Qwen2.5-72B-Instruct-GPTQ-Int8
null
null
"2024-09-17T12:53:28Z"
null
null
6,030
null
null
null
null
6
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-72B-Instruct", "base_model:quantized:Qwen/Qwen2.5-72B-Instruct", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
8,192
0.02
29,696
32,768
qwen2
64
80
8
0.000001
1,000,000
131,072
false
float16
4.39.3
true
152,064
null
0
null
null
null
null
Qwen/Qwen2.5-Coder-1.5B
null
null
"2024-09-18T09:41:38Z"
null
null
16,504
null
null
null
null
28
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "code", "qwen", "qwen-coder", "codeqwen", "conversational", "en", "arxiv:2409.12186", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-1.5B", "base_model:finetune:Qwen/Qwen2.5-1.5B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
1,536
0.02
8,960
32,768
qwen2
12
28
2
0.000001
1,000,000
131,072
true
bfloat16
4.44.0
true
151,936
null
0
null
null
null
null
unsloth/Qwen2.5-0.5B
null
null
"2024-09-18T21:15:04Z"
null
null
1,367
null
null
null
null
3
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "unsloth", "en", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-0.5B", "base_model:finetune:Qwen/Qwen2.5-0.5B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
null
true
bfloat16
4.46.1
true
151,936
null
0
null
null
null
null
unsloth/Qwen2.5-32B-Instruct-bnb-4bit
null
null
"2024-09-18T22:28:22Z"
null
null
5,637
null
null
null
null
6
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "unsloth", "conversational", "en", "arxiv:2309.00071", "base_model:Qwen/Qwen2.5-32B-Instruct", "base_model:quantized:Qwen/Qwen2.5-32B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
27,648
32,768
qwen2
40
64
8
0.000001
1,000,000
null
false
bfloat16
4.44.2
true
152,064
null
0
null
null
null
null
Qwen/Qwen2.5-Math-7B-Instruct
null
null
"2024-09-19T12:16:23Z"
null
null
133,524
null
null
null
null
34
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2409.12122", "base_model:Qwen/Qwen2.5-Math-7B", "base_model:finetune:Qwen/Qwen2.5-Math-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
4,096
qwen2
28
28
4
0.000001
10,000
4,096
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
AiCloser/Qwen2.5-32B-AGI
null
null
"2024-09-20T15:53:47Z"
null
null
389
null
null
null
null
56
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "zh", "en", "dataset:anthracite-org/kalo-opus-instruct-22k-no-refusal", "dataset:unalignment/toxic-dpo-v0.2", "dataset:Orion-zhen/dpo-toxic-zh", "base_model:Qwen/Qwen2.5-32B-Instruct", "base_model:finetune:Qwen/Qwen2.5-32B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
27,648
32,768
qwen2
40
64
8
0.000001
1,000,000
null
false
bfloat16
4.44.2
true
152,064
null
0
null
null
null
null
onnx-community/Qwen2.5-1.5B-Instruct
null
null
"2024-09-23T06:26:44Z"
null
null
245
null
null
null
null
3
transformers.js
[ "transformers.js", "onnx", "qwen2", "text-generation", "conversational", "base_model:Qwen/Qwen2.5-1.5B-Instruct", "base_model:quantized:Qwen/Qwen2.5-1.5B-Instruct", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
1,536
0.02
8,960
32,768
qwen2
12
28
2
0.000001
1,000,000
null
true
null
4.43.4
true
151,936
null
0
null
null
null
null
llm-jp/llm-jp-3-3.7b
null
null
"2024-09-23T12:25:31Z"
null
null
532
null
null
null
null
2
transformers
[ "transformers", "safetensors", "llama", "text-generation", "en", "ja", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
3,072
0.02
8,192
4,096
llama
24
28
24
0.00001
10,000
null
false
bfloat16
4.43.3
true
99,584
false
0
null
false
1
null
DiTy/gemma-2-27b-it-function-calling-GGUF
null
null
"2024-09-23T14:26:52Z"
null
null
264
null
null
null
null
5
transformers
[ "transformers", "safetensors", "gguf", "gemma2", "text-generation", "conversational", "function-calling", "trl", "en", "dataset:DiTy/function-calling", "base_model:google/gemma-2-27b-it", "base_model:quantized:google/gemma-2-27b-it", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
4,608
0.02
36,864
8,192
gemma2
32
46
16
0.000001
10,000
4,096
null
bfloat16
4.44.2
true
256,000
false
0
128
null
null
null
anthracite-org/magnum-v4-123b
null
null
"2024-09-27T00:25:27Z"
null
null
529
null
null
null
null
16
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "chat", "conversational", "en", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
12,288
0.02
28,672
131,072
mistral
96
88
8
0.00001
1,000,000
null
false
bfloat16
4.45.0.dev0
false
32,768
null
0
128
null
null
null
trollek/ImagePromptHelper-danube3-500M
null
null
"2024-09-27T19:40:20Z"
null
null
28
null
null
null
null
2
transformers
[ "transformers", "safetensors", "llama", "text-generation", "art", "conversational", "en", "dataset:k-mktr/improved-flux-prompts", "dataset:Gustavosta/Stable-Diffusion-Prompts", "dataset:ChrisGoringe/flux_prompts", "dataset:trollek/ImagePromptHelper-v01", "base_model:h2oai/h2o-danube3-500m-base", "base_model:finetune:h2oai/h2o-danube3-500m-base", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
32,000
32,001
silu
1,536
0.02
4,096
8,192
llama
16
16
8
0.00001
100,000
null
false
bfloat16
4.44.2
true
32,002
false
0
null
false
1
null
KBlueLeaf/TIPO-500M
null
null
"2024-09-28T17:59:56Z"
null
null
9,616
null
null
null
null
38
transformers
[ "transformers", "safetensors", "gguf", "llama", "text-generation", "en", "dataset:laion/conceptual-captions-12m-webdataset", "dataset:CaptionEmporium/coyo-hd-11m-llavanext", "dataset:KBlueLeaf/danbooru2023-metadata-database", "dataset:graph-based-captions/GBC10M", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
1,280
0.02
3,840
2,048
llama
20
20
20
0.00001
10,000
null
false
float32
4.44.2
false
32,013
false
0
null
false
1
null
SEOKDONG/llama3.1_korean_v0.1_sft_by_aidx
null
null
"2024-09-29T07:06:11Z"
null
null
2,731
null
null
null
null
2
null
[ "safetensors", "llama", "ktds", "ko", "ko_leaderboard", "korean", "text-generation", "conversational", "en", "dataset:AIDX-ktds/ko_leaderboard", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:apache-2.0", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
float32
4.44.2
true
128,256
false
0
null
false
1
null
abacusai/Dracarys2-72B-Instruct
null
null
"2024-09-30T15:33:37Z"
null
null
651
null
null
null
null
58
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
8,192
0.02
29,568
32,768
qwen2
64
80
8
0.000001
1,000,000
null
false
bfloat16
4.44.2
true
152,064
null
0
null
null
null
null
huihui-ai/Qwen2.5-Coder-1.5B-Instruct-abliterated
null
null
"2024-10-01T18:08:35Z"
null
null
80
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "abliterated", "uncensored", "conversational", "en", "base_model:Qwen/Qwen2.5-Coder-1.5B-Instruct", "base_model:finetune:Qwen/Qwen2.5-Coder-1.5B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
1,536
0.02
8,960
32,768
qwen2
12
28
2
0.000001
1,000,000
131,072
true
bfloat16
4.44.0
true
151,936
null
0
null
null
null
null
flammenai/Mahou-1.5-mistral-nemo-12B
null
null
"2024-10-06T17:12:41Z"
null
null
228
null
null
null
null
18
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "dataset:flammenai/MahouMix-v1", "base_model:flammenai/Flammades-Mistral-Nemo-12B", "base_model:finetune:flammenai/Flammades-Mistral-Nemo-12B", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
4
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.45.1
false
131,072
null
0
128
null
null
null
J-LAB/Florence-vl3
null
null
"2024-10-09T02:29:40Z"
null
null
471
null
null
null
null
2
transformers
[ "transformers", "safetensors", "florence2", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "region:us" ]
text-generation
null
null
2
[ "Florence2ForConditionalGeneration" ]
0
2
null
null
null
null
null
florence2
null
null
null
null
null
null
null
float16
4.41.0.dev0
null
51,289
null
null
null
null
null
null
tangledgroup/tangled-llama-p-128k-base-v0.1
null
null
"2024-10-09T09:12:41Z"
null
null
38
null
null
null
null
2
transformers
[ "transformers", "llama", "text-generation", "litgpt", "litdata", "conversational", "en", "am", "ar", "as", "az", "be", "bg", "bn", "br", "bs", "ca", "cs", "cy", "da", "de", "el", "eo", "es", "et", "eu", "fa", "ff", "fi", "fr", "fy", "ga", "gd", "gl", "gn", "gu", "ha", "he", "hi", "hr", "ht", "hu", "hy", "id", "ig", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "ku", "ky", "la", "lg", "li", "ln", "lo", "lt", "lv", "mg", "mk", "ml", "mn", "mr", "ms", "my", "ne", "nl", "no", "ns", "om", "or", "pa", "pl", "ps", "pt", "qu", "rm", "ro", "ru", "sa", "si", "sc", "sd", "sk", "sl", "so", "sq", "sr", "ss", "su", "sv", "sw", "ta", "te", "th", "tl", "tn", "tr", "ug", "uk", "ur", "uz", "vi", "wo", "xh", "yi", "yo", "zu", "dataset:yahma/alpaca-cleaned", "dataset:saillab/taco-datasets", "dataset:xu-song/cc100-samples", "dataset:badrex/llm-emoji-dataset", "dataset:pszemraj/simple_wikipedia", "dataset:AtlasUnified/Atlas-Reasoning", "dataset:fblgit/simple-math", "dataset:AtlasUnified/atlas-math-sets", "dataset:rvv-karma/Math-QA", "dataset:microsoft/orca-math-word-problems-200k", "dataset:meta-math/MetaMathQA", "dataset:TIGER-Lab/MathInstruct", "dataset:ChuGyouk/WebInstructSub-only-socratic", "dataset:thesven/gsm8k-reasoning", "dataset:AlgorithmicResearchGroup/math_reasoning_autoformalization_track", "dataset:KingNish/reasoning-base-20k", "dataset:fmars/wiki_stem", "dataset:ChuGyouk/WebInstructSub-only-sciencestackexchange", "dataset:bigcode/the-stack-smol-xs", "dataset:cognitivecomputations/dolphin-coder", "dataset:HuggingFaceH4/CodeAlpaca_20K", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:NuclearAi/Nuke-X-Glaive-Python-Dataset", "dataset:iamtarun/python_code_instructions_18k_alpaca", "dataset:kloodia/html_200k", "dataset:kloodia/json_200k", "dataset:kloodia/javascript_200k", "dataset:bleugreen/typescript-chunks", "dataset:SkunkworksAI/reasoning-0.01", "dataset:Magpie-Align/Magpie-Reasoning-150K", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
512
null
2,048
8,192
llama
32
16
8
0.00001
null
null
true
bfloat16
4.44.2
true
38,400
null
null
64
null
null
null
Zyphra/Zamba2-7B-Instruct
null
null
"2024-10-10T06:31:04Z"
null
null
2,206
null
null
null
null
82
transformers
[ "transformers", "safetensors", "zamba2", "text-generation", "conversational", "base_model:Zyphra/Zamba2-2.7B", "base_model:finetune:Zyphra/Zamba2-2.7B", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Zamba2ForCausalLM" ]
1
2
null
3,584
0.02
null
4,096
zamba2
32
81
32
0.00001
10,000
null
null
bfloat16
4.43.0.dev0
false
32,000
null
0
null
null
null
null
shibing624/chinese-text-correction-7b
null
null
"2024-10-12T10:47:36Z"
null
null
1,346
null
null
null
null
6
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "conversational", "zh", "dataset:shibing624/chinese_text_correction", "base_model:Qwen/Qwen2.5-7B-Instruct", "base_model:finetune:Qwen/Qwen2.5-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
131,072
false
bfloat16
4.37.2
true
152,064
null
0
null
null
null
null
Steelskull/ML-MS-Etheris-123B
null
null
"2024-10-13T22:52:42Z"
null
null
94
null
null
null
null
3
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "base_model:NeverSleep/Lumimaid-v0.2-123B", "base_model:merge:NeverSleep/Lumimaid-v0.2-123B", "base_model:SillyTilly/Mistral-Large-Instruct-2407", "base_model:merge:SillyTilly/Mistral-Large-Instruct-2407", "base_model:TheDrummer/Behemoth-123B-v1", "base_model:merge:TheDrummer/Behemoth-123B-v1", "base_model:anthracite-org/magnum-v2-123b", "base_model:merge:anthracite-org/magnum-v2-123b", "base_model:migtissera/Tess-3-Mistral-Large-2-123B", "base_model:merge:migtissera/Tess-3-Mistral-Large-2-123B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
12,288
0.02
28,672
32,768
mistral
96
88
8
0.00001
1,000,000
null
false
bfloat16
4.44.1
true
32,768
null
0
128
null
null
null
projecte-aina/salamandra-7b-aligned-EADOP
null
null
"2024-10-14T13:49:49Z"
null
null
287
null
null
null
null
2
transformers
[ "transformers", "safetensors", "llama", "text-generation", "legal", "conversational", "ca", "es", "dataset:alinia/EADOP-RAG-out-of-domain", "base_model:BSC-LT/salamandra-7b-instruct", "base_model:finetune:BSC-LT/salamandra-7b-instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
8,192
llama
32
32
8
0.000001
10,000
null
false
float32
4.45.2
false
256,000
false
0
128
false
1
null
ibm-granite/granite-guardian-3.0-2b
null
null
"2024-10-15T17:49:33Z"
null
null
2,244
null
null
null
null
16
transformers
[ "transformers", "safetensors", "granite", "text-generation", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "GraniteForCausalLM" ]
0
0
silu
2,048
0.02
8,192
8,192
granite
32
40
8
0.00001
10,000
null
true
bfloat16
4.45.0.dev0
false
49,155
false
0.1
null
false
null
null
Kortix/FastApply-1.5B-v1.0
null
null
"2024-10-18T11:55:22Z"
null
null
536
null
null
null
null
14
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "fast-apply", "instant-apply", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
1,536
0.02
8,960
32,768
qwen2
12
28
2
0.000001
1,000,000
null
true
bfloat16
4.44.2
true
151,936
null
0
null
null
null
null
arlineka/Isolde-12B
null
null
"2024-10-19T00:32:13Z"
null
null
55
null
null
null
null
2
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
3
4
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
float16
4.45.2
false
131,072
null
0
128
null
null
null
fblgit/TheBeagle-v2beta-32B-MGS
null
null
"2024-10-20T13:02:27Z"
null
null
266
null
null
null
null
7
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "en", "dataset:Magpie-Align/Magpie-Pro-300K-Filtered", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-32B-Instruct", "base_model:finetune:Qwen/Qwen2.5-32B-Instruct", "license:other", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
null
151,645
silu
5,120
0.02
27,648
32,768
qwen2
40
64
8
0.000001
1,000,000
null
false
bfloat16
4.45.2
false
152,064
null
0
null
null
null
null
Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4
null
null
"2024-10-21T18:23:48Z"
null
null
193
null
null
null
null
6
null
[ "safetensors", "qwen2", "chat", "text-generation", "conversational", "en", "de", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-14B-Instruct", "base_model:finetune:Qwen/Qwen2.5-14B-Instruct", "license:apache-2.0", "model-index", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
13,824
32,768
qwen2
40
48
8
0.000001
1,000,000
null
false
bfloat16
4.45.2
true
152,064
null
0
null
null
null
null
thwin27/Aria-sequential_mlp-bnb_nf4
null
null
"2024-10-23T09:01:54Z"
null
null
497
null
null
null
null
8
transformers
[ "transformers", "safetensors", "aria", "text-generation", "image-text-to-text", "conversational", "custom_code", "base_model:rhymes-ai/Aria", "base_model:quantized:rhymes-ai/Aria", "license:apache-2.0", "autotrain_compatible", "4-bit", "bitsandbytes", "region:us" ]
image-text-to-text
null
null
2
[ "AriaForConditionalGeneration" ]
null
null
null
null
null
null
null
aria
null
null
null
null
null
null
null
float16
4.45.0
null
null
null
null
null
null
null
null
eugenehp/Llama3-8B-1.58-100B-tokens-GGUF
null
null
"2024-10-24T13:07:33Z"
null
null
218
null
null
null
null
2
transformers
[ "transformers", "safetensors", "gguf", "llama", "text-generation", "conversational", "arxiv:2402.17764", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "bitnet", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.44.0.dev0
true
128,256
false
0
null
false
1
null
CultriX/Qwen2.5-14B-MegaMerge-pt2-Q5_K_M-GGUF
null
null
"2024-10-25T07:17:25Z"
null
null
75
null
null
null
null
2
transformers
[ "transformers", "gguf", "qwen2", "text-generation", "mergekit", "merge", "llama-cpp", "gguf-my-repo", "en", "base_model:CultriX/Qwen2.5-14B-MegaMerge-pt2", "base_model:quantized:CultriX/Qwen2.5-14B-MegaMerge-pt2", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us", "conversational" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
5,120
0.02
13,824
131,072
qwen2
40
48
8
0.00001
1,000,000
null
false
bfloat16
4.45.1
true
151,665
null
0
null
null
null
null
lemon07r/Gemma-2-Ataraxy-v4d-9B
null
null
"2024-10-25T10:39:29Z"
null
null
124
null
null
null
null
5
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "en", "base_model:lemon07r/Gemma-2-Ataraxy-v4c-9B", "base_model:merge:lemon07r/Gemma-2-Ataraxy-v4c-9B", "base_model:sam-paech/Darkest-muse-v1", "base_model:merge:sam-paech/Darkest-muse-v1", "license:gemma", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.43.3
true
256,000
false
0
256
null
null
null
ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca
null
null
"2024-10-27T10:06:47Z"
null
null
29
null
null
null
null
2
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "nlp", "code", "conversational", "multilingual", "dataset:tatsu-lab/alcapa", "base_model:google/gemma-2-2b-jpn-it", "base_model:finetune:google/gemma-2-2b-jpn-it", "license:gemma", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Gemma2ForCausalLM" ]
256,000
256,001
null
2,304
0.02
9,216
8,192
gemma2
8
26
4
0.000001
10,000
4,096
null
bfloat16
4.45.2
true
256,002
false
0
256
null
null
null
allknowingroger/Qwen-modelstock2-15B
null
null
"2024-10-27T18:18:36Z"
null
null
15
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2403.19522", "base_model:allknowingroger/Qwen-modelstock-15B", "base_model:merge:allknowingroger/Qwen-modelstock-15B", "base_model:allknowingroger/Qwen2.5-slerp-14B", "base_model:merge:allknowingroger/Qwen2.5-slerp-14B", "base_model:allknowingroger/Qwenslerp2-14B", "base_model:merge:allknowingroger/Qwenslerp2-14B", "base_model:allknowingroger/Qwenslerp3-14B", "base_model:merge:allknowingroger/Qwenslerp3-14B", "base_model:rombodawg/Rombos-LLM-V2.6-Qwen-14b", "base_model:merge:rombodawg/Rombos-LLM-V2.6-Qwen-14b", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
13,824
32,768
qwen2
40
48
8
0.000001
1,000,000
null
false
bfloat16
4.45.1
true
152,064
null
0
null
null
null
null
zetasepic/Qwen2.5-72B-Instruct-abliterated-v2
null
null
"2024-10-27T23:51:00Z"
null
null
132
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "base_model:Qwen/Qwen2.5-72B", "base_model:finetune:Qwen/Qwen2.5-72B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
8,192
0.02
29,568
32,768
qwen2
64
80
8
0.000001
1,000,000
null
false
bfloat16
4.44.2
true
152,064
null
0
null
null
null
null
knifeayumu/Magnum-v4-Cydonia-v1.2-22B
null
null
"2024-10-28T13:22:33Z"
null
null
40
null
null
null
null
3
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "base_model:TheDrummer/Cydonia-22B-v1.2", "base_model:merge:TheDrummer/Cydonia-22B-v1.2", "base_model:anthracite-org/magnum-v4-22b", "base_model:merge:anthracite-org/magnum-v4-22b", "license:other", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
6,144
0.02
16,384
32,768
mistral
48
56
8
0.00001
1,000,000
null
false
bfloat16
4.45.1
false
32,768
null
0
128
null
null
null
prithivMLmods/Meta-Llama-3.2-1B-GGUF-QX
null
null
"2024-10-28T16:51:09Z"
null
null
419
null
null
null
null
4
transformers
[ "transformers", "gguf", "llama", "text-generation-inference", "unsloth", "Llama-3.2-1B", "Quants", "text-generation", "en", "base_model:meta-llama/Llama-3.2-1B", "base_model:quantized:meta-llama/Llama-3.2-1B", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
bunnycore/Qwen2.5-3B-Loki
null
null
"2024-10-30T17:47:27Z"
null
null
17
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2306.01708", "base_model:Qwen/Qwen2.5-3B", "base_model:merge:Qwen/Qwen2.5-3B", "base_model:bunnycore/Qwen2.5-3B-MiniMix", "base_model:merge:bunnycore/Qwen2.5-3B-MiniMix", "base_model:bunnycore/Qwen2.5-3B-RP-Mix", "base_model:merge:bunnycore/Qwen2.5-3B-RP-Mix", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
float16
4.45.1
true
151,936
null
0
null
null
null
null
bunnycore/Qwen2.5-7B-Instruct-Fusion
null
null
"2024-10-31T10:18:52Z"
null
null
32
null
null
null
null
3
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2403.19522", "base_model:Qwen/Qwen2.5-7B", "base_model:merge:Qwen/Qwen2.5-7B", "base_model:fblgit/cybertron-v4-qw7B-MGS", "base_model:merge:fblgit/cybertron-v4-qw7B-MGS", "base_model:rombodawg/Rombos-LLM-V2.5-Qwen-7b", "base_model:merge:rombodawg/Rombos-LLM-V2.5-Qwen-7b", "base_model:sethuiyer/Qwen2.5-7B-Anvita", "base_model:merge:sethuiyer/Qwen2.5-7B-Anvita", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
float16
4.45.1
true
152,064
null
0
null
null
null
null
prithivMLmods/SmolLM2-135M-GGUF
null
null
"2024-11-01T09:04:47Z"
null
null
229
null
null
null
null
4
null
[ "gguf", "llama", "135M", "text-generation", "en", "dataset:yahma/alpaca-cleaned", "base_model:HuggingFaceTB/SmolLM2-135M", "base_model:quantized:HuggingFaceTB/SmolLM2-135M", "license:creativeml-openrail-m", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
prithivMLmods/SmolLM2-135M-Instruct-GGUF
null
null
"2024-11-01T10:34:09Z"
null
null
173
null
null
null
null
4
null
[ "gguf", "llama", "135M", "Instruct", "text-generation", "en", "dataset:yahma/alpaca-cleaned", "base_model:HuggingFaceTB/SmolLM2-135M-Instruct", "base_model:quantized:HuggingFaceTB/SmolLM2-135M-Instruct", "license:creativeml-openrail-m", "endpoints_compatible", "region:us", "conversational" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
Pretergeek/OpenChat-3.5-0106_32K-PoSE
null
null
"2024-11-02T04:14:59Z"
null
null
53
null
null
null
null
3
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "dataset:Yukang/LongAlpaca-12k", "arxiv:2309.10400", "base_model:openchat/openchat-3.5-0106", "base_model:finetune:openchat/openchat-3.5-0106", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
32,000
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.45.2
true
32,002
null
0
128
null
null
null
prithivMLmods/Qwen2.5-0.5B-200K-GGUF
null
null
"2024-11-02T05:24:12Z"
null
null
196
null
null
null
null
4
null
[ "gguf", "qwen2", "Qwen2.5", "200K", "5B", "Llama-cpp", "text-generation", "en", "dataset:HuggingFaceH4/ultrachat_200k", "base_model:Qwen/Qwen2.5-0.5B", "base_model:quantized:Qwen/Qwen2.5-0.5B", "license:creativeml-openrail-m", "endpoints_compatible", "region:us", "conversational" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
qwen2
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
gokaygokay/SmolLM2-Prompt-Enhance
null
null
"2024-11-02T11:04:58Z"
null
null
64
null
null
null
null
5
transformers
[ "transformers", "safetensors", "llama", "text-generation", "trl", "sft", "conversational", "en", "dataset:gokaygokay/prompt-enhancement-75k", "base_model:HuggingFaceTB/SmolLM2-135M-Instruct", "base_model:finetune:HuggingFaceTB/SmolLM2-135M-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
576
0.041667
1,536
8,192
llama
9
30
3
0.00001
100,000
null
true
float32
4.46.1
true
49,152
false
0
64
false
1
null
ZiweiChen/BioMistral-Clinical-7B
null
null
"2024-11-02T16:53:09Z"
null
null
47
null
null
null
null
2
null
[ "safetensors", "mistral", "clinical", "biology", "text-generation", "conversational", "en", "dataset:AGBonnet/augmented-clinical-notes", "base_model:BioMistral/BioMistral-7B", "base_model:finetune:BioMistral/BioMistral-7B", "license:mit", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float32
4.45.2
false
32,000
null
0
128
null
null
null
KBlueLeaf/TIPO-100M
null
null
"2024-11-03T09:40:05Z"
null
null
172
null
null
null
null
2
transformers
[ "transformers", "safetensors", "gguf", "llama", "text-generation", "en", "dataset:laion/conceptual-captions-12m-webdataset", "dataset:CaptionEmporium/coyo-hd-11m-llavanext", "dataset:KBlueLeaf/danbooru2023-metadata-database", "dataset:graph-based-captions/GBC10M", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
640
0.02
2,240
8,192
llama
10
10
10
0.00001
10,000
null
false
float32
4.44.2
false
32,013
false
0
null
false
1
null
Abhaykoul/1.7
null
null
"2024-11-03T14:01:56Z"
null
null
25
null
null
null
null
2
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "base_model:HuggingFaceTB/SmolLM2-360M-Instruct", "base_model:finetune:HuggingFaceTB/SmolLM2-360M-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "LlamaForCausalLM" ]
1
2
silu
960
0.02
2,560
8,192
llama
15
32
5
0.00001
100,000
null
true
bfloat16
4.47.0.dev0
true
49,152
false
0
64
false
1
null
CultriX/Qwen2.5-14B-UpToDate
null
null
"2024-11-03T20:54:42Z"
null
null
19
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "base_model:CultriX/Qwen2.5-14B-MegaMerge-pt2", "base_model:finetune:CultriX/Qwen2.5-14B-MegaMerge-pt2", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
5,120
0.02
13,824
131,072
qwen2
40
48
8
0.00001
1,000,000
null
false
bfloat16
4.46.1
true
151,665
null
0
null
null
null
null
SejongKRX/Sejong-Qwen-v5
null
null
"2024-11-04T12:28:48Z"
null
null
20
null
null
null
null
2
transformers
[ "transformers", "pytorch", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "krx", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
float16
4.44.2
true
152,064
null
0
null
null
null
null
bunnycore/Qwen2.5-7B-CyberRombos
null
null
"2024-11-04T16:34:04Z"
null
null
28
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2306.01708", "base_model:Qwen/Qwen2.5-7B", "base_model:merge:Qwen/Qwen2.5-7B", "base_model:fblgit/cybertron-v4-qw7B-MGS", "base_model:merge:fblgit/cybertron-v4-qw7B-MGS", "base_model:rombodawg/Rombos-LLM-V2.5-Qwen-7b", "base_model:merge:rombodawg/Rombos-LLM-V2.5-Qwen-7b", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
float16
4.45.1
true
152,064
null
0
null
null
null
null
EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1
null
null
"2024-11-06T11:14:10Z"
null
null
160
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "conversational", "dataset:anthracite-org/kalo-opus-instruct-22k-no-refusal", "dataset:Nopm/Opus_WritingStruct", "dataset:Gryphe/Sonnet3.5-SlimOrcaDedupCleaned", "dataset:Gryphe/Sonnet3.5-Charcard-Roleplay", "dataset:Gryphe/ChatGPT-4o-Writing-Prompts", "dataset:Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned", "dataset:Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned", "dataset:nothingiisreal/Reddit-Dirty-And-WritingPrompts", "dataset:allura-org/Celeste-1.x-data-mixture", "dataset:cognitivecomputations/dolphin-2.9.3", "base_model:Qwen/Qwen2.5-72B", "base_model:finetune:Qwen/Qwen2.5-72B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
null
151,643
silu
8,192
0.02
29,568
131,072
qwen2
64
80
8
0.00001
1,000,000
null
false
bfloat16
4.45.1
false
152,064
null
0
null
null
null
null
facebook/MobileLLM-350M-layer-share
null
null
"2024-11-07T00:27:46Z"
null
null
19
null
null
null
null
2
transformers
[ "transformers", "pytorch", "mobilellm", "text-generation", "custom_code", "arxiv:2402.14905", "license:cc-by-nc-4.0", "autotrain_compatible", "region:us" ]
text-generation
null
null
2
[ "MobileLLMForCausalLM" ]
1
2
silu
960
0.02
2,560
2,048
mobilellm
15
32
5
0.00001
10,000
null
false
float16
4.41.2
false
32,000
false
0
64
false
1
null
MaziyarPanahi/Mistral-Crab-DPO-GGUF
null
null
"2024-11-07T10:14:42Z"
null
null
69
null
null
null
null
2
null
[ "gguf", "mistral", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:THU-KEG/Mistral-Crab-DPO", "base_model:quantized:THU-KEG/Mistral-Crab-DPO", "region:us", "imatrix", "conversational" ]
text-generation
null
null
2
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
mlx-community/Qwen2.5.1-Coder-7B-Instruct-8bit
null
null
"2024-11-07T18:05:46Z"
null
null
71
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "code", "codeqwen", "chat", "qwen", "qwen-coder", "mlx", "conversational", "en", "base_model:Qwen/Qwen2.5-Coder-7B-Instruct", "base_model:quantized:Qwen/Qwen2.5-Coder-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
131,072
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
MaziyarPanahi/calme-3.3-qwenloi-3b
null
null
"2024-11-07T22:26:36Z"
null
null
18
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "qwen", "qwen2.5", "finetune", "french", "legal", "loi", "conversational", "fr", "en", "dataset:MaziyarPanahi/calme-legalkit-v0.2", "base_model:Qwen/Qwen2.5-3B", "base_model:finetune:Qwen/Qwen2.5-3B", "license:other", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
bfloat16
4.45.2
true
151,665
null
0
null
null
null
null
KaraKaraWitch/EurobeatVARemix-Qwen2.5-72b
null
null
"2024-11-08T07:10:37Z"
null
null
44
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2403.19522", "base_model:EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1", "base_model:merge:EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1", "base_model:Qwen/Qwen2.5-72B", "base_model:merge:Qwen/Qwen2.5-72B", "base_model:ZeusLabs/Chronos-Platinum-72B", "base_model:merge:ZeusLabs/Chronos-Platinum-72B", "base_model:abacusai/Dracarys2-72B-Instruct", "base_model:merge:abacusai/Dracarys2-72B-Instruct", "base_model:m8than/banana-2-b-72b", "base_model:merge:m8than/banana-2-b-72b", "base_model:rombodawg/Rombos-LLM-V2.5-Qwen-72b", "base_model:merge:rombodawg/Rombos-LLM-V2.5-Qwen-72b", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
8,192
0.02
29,568
131,072
qwen2
64
80
8
0.00001
1,000,000
null
false
bfloat16
4.44.2
true
152,064
null
0
null
null
null
null
0x7o/poemma-12B
null
null
"2024-11-08T09:46:09Z"
null
null
22
null
null
null
null
2
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "orpo", "conversational", "ru", "dataset:0x7o/poemma-10k", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.46.2
true
131,072
null
0
128
null
null
null
Ttimofeyka/Tissint-14B-128k-RP
null
null
"2024-11-09T16:36:09Z"
null
null
5
null
null
null
null
2
transformers
[ "transformers", "pytorch", "qwen2", "text-generation", "unsloth", "trl", "sft", "conversational", "base_model:arcee-ai/SuperNova-Medius", "base_model:finetune:arcee-ai/SuperNova-Medius", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
5,120
0.02
13,824
131,072
qwen2
40
48
8
0.00001
1,000,000
null
false
float16
4.46.2
true
152,064
null
0
null
null
null
null
BenevolenceMessiah/Qwen2.5-Coder-7B-3x-Instruct-TIES-v1.2
null
null
"2024-11-10T23:58:21Z"
null
null
2
null
null
null
null
2
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2306.01708", "base_model:BenevolenceMessiah/Qwen2.5-Coder-7B-Chat-Instruct-TIES-v1.2", "base_model:merge:BenevolenceMessiah/Qwen2.5-Coder-7B-Chat-Instruct-TIES-v1.2", "base_model:MadeAgents/Hammer2.0-7b", "base_model:merge:MadeAgents/Hammer2.0-7b", "base_model:Qwen/Qwen2.5-Coder-7B", "base_model:merge:Qwen/Qwen2.5-Coder-7B", "base_model:huihui-ai/Qwen2.5-Coder-7B-Instruct-abliterated", "base_model:merge:huihui-ai/Qwen2.5-Coder-7B-Instruct-abliterated", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
2
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
151,665
null
0
null
null
null
null
distilbert/distilgpt2
null
null
"2022-03-02T23:29:04Z"
null
null
3,841,453
null
null
null
null
442
transformers
[ "transformers", "pytorch", "tf", "jax", "tflite", "rust", "coreml", "safetensors", "gpt2", "text-generation", "exbert", "en", "dataset:openwebtext", "arxiv:1910.01108", "arxiv:2201.08542", "arxiv:2203.12574", "arxiv:1910.09700", "arxiv:1503.02531", "license:apache-2.0", "model-index", "co2_eq_emissions", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
null
null
null
50,257
null
null
null
null
null
null
openai-community/gpt2-medium
null
null
"2022-03-02T23:29:04Z"
null
null
513,438
null
null
null
null
154
transformers
[ "transformers", "pytorch", "tf", "jax", "rust", "onnx", "safetensors", "gpt2", "text-generation", "en", "arxiv:1910.09700", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
null
null
null
50,257
null
null
null
null
null
null
openai-community/gpt2-xl
null
null
"2022-03-02T23:29:04Z"
null
null
210,884
null
null
null
null
310
transformers
[ "transformers", "pytorch", "tf", "jax", "rust", "safetensors", "gpt2", "text-generation", "en", "arxiv:1910.09700", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
null
null
null
50,257
null
null
null
null
null
null
openai-community/openai-gpt
null
null
"2022-03-02T23:29:04Z"
null
null
40,728
null
null
null
null
242
transformers
[ "transformers", "pytorch", "tf", "rust", "safetensors", "openai-gpt", "text-generation", "en", "arxiv:1705.11168", "arxiv:1803.02324", "arxiv:1910.09700", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "OpenAIGPTLMHeadModel" ]
null
null
null
null
0.02
null
null
openai-gpt
null
null
null
null
null
null
null
null
null
null
40,478
null
null
null
null
null
null
transfo-xl/transfo-xl-wt103
null
null
"2022-03-02T23:29:04Z"
null
null
2,883
null
null
null
null
14
transformers
[ "transformers", "pytorch", "tf", "transfo-xl", "text-generation", "en", "dataset:wikitext-103", "arxiv:1901.02860", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "TransfoXLLMHeadModel" ]
null
0
null
null
null
null
null
transfo-xl
null
null
null
null
null
null
null
null
null
null
267,735
null
null
null
null
null
null
EleutherAI/gpt-neo-1.3B
null
null
"2022-03-02T23:29:04Z"
null
null
164,241
null
null
null
null
263
transformers
[ "transformers", "pytorch", "jax", "rust", "safetensors", "gpt_neo", "text-generation", "text generation", "causal-lm", "en", "dataset:EleutherAI/pile", "arxiv:2101.00027", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoForCausalLM" ]
50,256
50,256
null
2,048
0.02
null
2,048
gpt_neo
null
null
null
null
null
null
null
null
4.9.0.dev0
true
50,257
null
0
null
null
null
null
EleutherAI/gpt-neo-125m
null
null
"2022-03-02T23:29:04Z"
null
null
359,545
null
null
null
null
182
transformers
[ "transformers", "pytorch", "jax", "rust", "safetensors", "gpt_neo", "text-generation", "text generation", "causal-lm", "en", "dataset:EleutherAI/pile", "arxiv:2101.00027", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoForCausalLM" ]
50,256
50,256
null
768
0.02
null
2,048
gpt_neo
null
null
null
null
null
null
null
null
4.9.0.dev0
true
50,257
null
0
null
null
null
null