id
stringlengths
7
117
author
stringclasses
6 values
sha
null
created_at
unknown
last_modified
null
disabled
null
downloads
int64
0
18.6M
downloads_all_time
null
gated
bool
1 class
gguf
null
inference
null
likes
int64
0
4.77k
library_name
stringclasses
36 values
tags
sequencelengths
1
430
pipeline_tag
stringclasses
32 values
mask_token
null
model_index
null
trending_score
int64
0
132
architectures
sequencelengths
1
5
bos_token_id
int64
-1
256k
eos_token_id
int64
-1
256k
hidden_act
stringclasses
15 values
hidden_size
int64
1
20.5k
initializer_range
float64
0
1
intermediate_size
int64
1
98.3k
max_position_embeddings
int64
8
1.05M
model_type
stringclasses
530 values
num_attention_heads
int64
1
5k
num_hidden_layers
int64
-1
8.93k
num_key_value_heads
int64
1
160
rms_norm_eps
float64
0
7
rope_theta
float64
1k
1,000B
sliding_window
int64
0
262k
tie_word_embeddings
bool
2 classes
torch_dtype
stringclasses
8 values
transformers_version
stringclasses
207 values
use_cache
bool
2 classes
vocab_size
int64
-1
5.03M
attention_bias
bool
2 classes
attention_dropout
float64
0
0.5
head_dim
int64
2
256
mlp_bias
bool
2 classes
pretraining_tp
int64
0
8
rope_scaling
dict
DeepPavlov/rudialogpt3_medium_based_on_gpt2_v2
null
null
"2022-11-07T13:26:22Z"
null
null
1,334
null
null
null
null
4
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.17.0
true
50,257
null
null
null
null
null
null
facebook/galactica-1.3b
null
null
"2022-11-16T13:37:55Z"
null
null
5,675
null
null
null
null
65
transformers
[ "transformers", "pytorch", "opt", "text-generation", "galactica", "arxiv:1810.03993", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "OPTForCausalLM" ]
0
2
null
2,048
null
null
2,048
opt
32
24
null
null
null
null
null
float16
4.21.0.dev0
true
50,000
null
0.1
null
null
null
null
facebook/galactica-6.7b
null
null
"2022-11-16T14:20:00Z"
null
null
801
null
null
null
null
96
transformers
[ "transformers", "pytorch", "opt", "text-generation", "galactica", "arxiv:1810.03993", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "OPTForCausalLM" ]
0
2
null
4,096
null
null
2,048
opt
32
32
null
null
null
null
null
float16
4.21.0.dev0
true
50,000
null
0.1
null
null
null
null
facebook/galactica-120b
null
null
"2022-11-16T14:49:36Z"
null
null
899
null
null
null
null
152
transformers
[ "transformers", "pytorch", "opt", "text-generation", "galactica", "arxiv:1810.03993", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "OPTForCausalLM" ]
0
2
null
10,240
null
null
2,048
opt
80
96
null
null
null
null
null
float16
4.21.0.dev0
true
50,000
null
0.1
null
null
null
null
microsoft/biogpt
null
null
"2022-11-20T13:20:45Z"
null
null
61,377
null
null
null
null
226
transformers
[ "transformers", "pytorch", "biogpt", "text-generation", "en", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BioGptForCausalLM" ]
0
2
gelu
1,024
0.02
4,096
1,024
biogpt
16
24
null
null
null
null
null
null
4.25.0.dev0
true
42,384
null
null
null
null
null
null
KPEKEP/rugpt_chitchat
null
null
"2022-11-29T14:48:34Z"
null
null
147
null
null
null
null
1
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "PyTorch", "Transformers", "ru", "license:unlicense", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.21.1
true
50,258
null
null
null
null
null
null
Den4ikAI/DLM_CHITCHAT_700M
null
null
"2022-12-02T16:36:28Z"
null
null
42
null
null
null
null
4
transformers
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "ru", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.21.1
true
50,258
null
null
null
null
null
null
Den4ikAI/DLM_CHITCHAT_500M
null
null
"2022-12-03T13:16:40Z"
null
null
22
null
null
null
null
1
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "ru", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.19.4
true
50,258
null
null
null
null
null
null
FredZhang7/distilgpt2-stable-diffusion-v2
null
null
"2022-12-10T06:49:07Z"
null
null
3,758
null
null
null
null
99
transformers
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "stable-diffusion", "prompt-generator", "arxiv:2210.14140", "dataset:FredZhang7/stable-diffusion-prompts-2.47M", "dataset:poloclub/diffusiondb", "dataset:Gustavosta/Stable-Diffusion-Prompts", "dataset:bartman081523/stable-diffusion-discord-prompts", "license:creativeml-openrail-m", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.21.2
true
50,257
null
null
null
null
null
null
Reverb/GPyT
null
null
"2022-12-11T20:14:26Z"
null
null
18
null
null
null
null
1
transformers
[ "transformers", "pytorch", "tf", "gpt2", "text-generation", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
0
2
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
null
4.8.2
true
52,000
null
null
null
null
null
null
stanford-crfm/BioMedLM
null
null
"2022-12-14T08:14:59Z"
null
null
2,903
null
null
null
null
395
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "dataset:pubmed", "arxiv:2403.18421", "license:bigscience-bloom-rail-1.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
28,895
28,895
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.21.3
false
28,896
null
null
null
null
null
null
microsoft/Promptist
null
null
"2022-12-19T07:43:54Z"
null
null
1,611
null
null
null
null
63
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
null
null
null
50,257
null
null
null
null
null
null
PygmalionAI/pygmalion-350m
null
null
"2022-12-20T22:04:32Z"
null
null
1,832
null
null
null
null
54
transformers
[ "transformers", "pytorch", "opt", "text-generation", "convAI", "conversational", "en", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "OPTForCausalLM" ]
2
2
null
1,024
null
null
2,048
opt
16
24
null
null
null
null
null
float16
4.20.0.dev0
true
50,272
null
0
null
null
null
null
robowaifudev/megatron-gpt2-345m
null
null
"2022-12-21T00:55:48Z"
null
null
4,065
null
null
null
null
9
transformers
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "en", "dataset:wikitext", "dataset:openwebtext", "dataset:spacemanidol/cc-stories", "arxiv:1909.08053", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
true
null
4.15.0
true
50,257
null
null
null
null
null
null
sander-wood/tunesformer
null
null
"2023-01-07T17:07:32Z"
null
null
29
null
null
null
null
5
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "music", "dataset:sander-wood/irishman", "arxiv:2301.02884", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.18.0
true
128
null
null
null
null
null
null
AUTOMATIC/promptgen-lexart
null
null
"2023-01-18T03:38:39Z"
null
null
2,836
null
null
null
null
48
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "en", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.19.2
true
50,257
null
null
null
null
null
null
hakurei/lotus-12B
null
null
"2023-01-25T00:01:32Z"
null
null
951
null
null
null
null
27
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "causal-lm", "en", "license:agpl-3.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
5,120
0.02
20,480
2,048
gpt_neox
40
36
null
null
null
null
false
float16
4.24.0
true
50,688
null
null
null
null
null
null
facebook/opt-iml-1.3b
null
null
"2023-01-26T00:08:49Z"
null
null
720
null
null
null
null
29
transformers
[ "transformers", "pytorch", "opt", "text-generation", "arxiv:2212.12017", "license:other", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "OPTForCausalLM" ]
2
2
null
2,048
null
null
2,048
opt
32
24
null
null
null
null
null
float16
4.22.2
true
50,272
null
0
null
null
null
null
Dmitriy007/rugpt2_gen_news
null
null
"2023-02-01T08:08:55Z"
null
null
58
null
null
null
null
4
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "PyTorch", "Transformers", "text generation", "ru", "dataset:Lenta2", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.26.0
true
50,264
null
null
null
null
null
null
microsoft/BioGPT-Large
null
null
"2023-02-03T16:17:26Z"
null
null
2,276
null
null
null
null
170
transformers
[ "transformers", "pytorch", "biogpt", "text-generation", "medical", "en", "dataset:pubmed", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BioGptForCausalLM" ]
0
2
gelu
1,600
0.02
6,400
2,048
biogpt
25
48
null
null
null
null
null
float32
4.27.0.dev0
true
57,717
null
null
null
null
null
null
microsoft/BioGPT-Large-PubMedQA
null
null
"2023-02-03T20:33:43Z"
null
null
628
null
null
null
null
103
transformers
[ "transformers", "pytorch", "biogpt", "text-generation", "medical", "en", "dataset:pubmed_qa", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BioGptForCausalLM" ]
0
2
gelu
1,600
0.02
6,400
2,048
biogpt
25
48
null
null
null
null
null
float32
4.27.0.dev0
true
57,726
null
null
null
null
null
null
FredZhang7/anime-anything-promptgen-v2
null
null
"2023-02-09T07:29:25Z"
null
null
3,974
null
null
null
null
58
transformers
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "stable-diffusion", "anime", "anything-v4", "art", "arxiv:2210.14140", "en", "dataset:FredZhang7/anime-prompts-180K", "license:creativeml-openrail-m", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.25.1
true
50,257
null
null
null
null
null
null
EleutherAI/pythia-1.4b
null
null
"2023-02-09T14:08:20Z"
null
null
23,157
null
null
null
null
22
transformers
[ "transformers", "pytorch", "safetensors", "gpt_neox", "text-generation", "causal-lm", "pythia", "en", "dataset:EleutherAI/the_pile", "arxiv:2304.01373", "arxiv:2101.00027", "arxiv:2201.07311", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.24.0
true
50,304
null
null
null
null
null
null
EleutherAI/pythia-410m
null
null
"2023-02-13T18:45:00Z"
null
null
36,217
null
null
null
null
21
transformers
[ "transformers", "pytorch", "safetensors", "gpt_neox", "text-generation", "causal-lm", "pythia", "en", "dataset:EleutherAI/pile", "arxiv:2304.01373", "arxiv:2101.00027", "arxiv:2201.07311", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.24.0
true
50,304
null
null
null
null
null
null
EleutherAI/pythia-1b-deduped
null
null
"2023-02-14T00:07:42Z"
null
null
11,164
null
null
null
null
17
transformers
[ "transformers", "pytorch", "safetensors", "gpt_neox", "text-generation", "causal-lm", "pythia", "en", "dataset:EleutherAI/the_pile_deduplicated", "arxiv:2304.01373", "arxiv:2101.00027", "arxiv:2201.07311", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.24.0
true
50,304
null
null
null
null
null
null
EleutherAI/pythia-6.9b
null
null
"2023-02-14T04:18:48Z"
null
null
19,965
null
null
null
null
48
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "causal-lm", "pythia", "en", "dataset:EleutherAI/pile", "arxiv:2304.01373", "arxiv:2101.00027", "arxiv:2201.07311", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.24.0
true
50,432
null
null
null
null
null
null
TurkuNLP/gpt3-finnish-small
null
null
"2023-02-15T10:08:16Z"
null
null
2,320
null
null
null
null
11
transformers
[ "transformers", "pytorch", "bloom", "feature-extraction", "text-generation", "fi", "arxiv:2203.02155", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BloomModel" ]
1
2
null
768
0.02
null
null
bloom
null
null
null
null
null
null
null
null
4.26.0.dev0
true
131,072
null
0
null
null
1
null
TurkuNLP/gpt3-finnish-large
null
null
"2023-02-15T10:26:29Z"
null
null
1,471
null
null
null
null
5
transformers
[ "transformers", "pytorch", "bloom", "feature-extraction", "text-generation", "fi", "arxiv:2203.02155", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BloomModel" ]
1
2
null
1,536
0.02
null
null
bloom
null
null
null
null
null
null
null
null
4.26.0.dev0
true
131,072
null
0
null
null
1
null
digitous/GPT-R
null
null
"2023-02-16T16:02:46Z"
null
null
832
null
null
null
null
11
transformers
[ "transformers", "pytorch", "gptj", "text-generation", "en", "license:bigscience-openrail-m", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTJForCausalLM" ]
50,256
50,256
null
null
0.02
null
null
gptj
null
null
null
null
null
null
false
float16
4.24.0
true
50,400
null
null
null
null
null
null
vicclab/FolkGPT
null
null
"2023-02-26T10:43:03Z"
null
null
28
null
null
null
null
1
transformers
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "en", "dataset:vicclab/fairy_tales", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.26.1
true
50,257
null
null
null
null
null
null
togethercomputer/GPT-NeoXT-Chat-Base-20B
null
null
"2023-03-03T00:24:29Z"
null
null
1,063
null
null
null
null
695
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu_fast
6,144
0.02
24,576
2,048
gpt_neox
64
44
null
null
null
null
false
float16
4.21.1
true
50,432
null
null
null
null
null
null
boboto/LLaMA-65B-HF
null
null
"2023-03-05T19:02:51Z"
null
null
25
null
null
null
null
19
transformers
[ "transformers", "pytorch", "llama", "text-generation", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LLaMAForCausalLM" ]
0
1
silu
8,192
0.02
22,016
null
llama
64
80
null
0.00001
null
null
null
float16
4.27.0.dev0
true
32,000
null
null
null
null
null
null
EleutherAI/pythia-1b
null
null
"2023-03-10T21:42:46Z"
null
null
65,739
null
null
null
null
32
transformers
[ "transformers", "pytorch", "safetensors", "gpt_neox", "text-generation", "causal-lm", "pythia", "en", "dataset:the_pile", "arxiv:2304.01373", "arxiv:2101.00027", "arxiv:2201.07311", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.24.0
true
50,304
null
null
null
null
null
null
eminecg/Lawsuit-Petition-TextGen-Gpt2-Preprocess-Dataset
null
null
"2023-03-13T06:13:52Z"
null
null
10
null
null
null
null
1
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
0
2
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.26.1
true
50,000
null
null
null
null
null
null
ozcur/alpaca-native-4bit
null
null
"2023-03-20T03:19:09Z"
null
null
18
null
null
null
null
58
transformers
[ "transformers", "llama", "text-generation", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LLaMAForCausalLM" ]
0
1
silu
4,096
0.02
11,008
null
llama
32
32
null
0.000001
null
null
false
float32
4.27.0.dev0
true
32,001
null
null
null
null
null
null
IlyaGusev/rugpt_medium_turbo_instructed
null
null
"2023-03-20T14:40:43Z"
null
null
27
null
null
null
null
5
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "ru", "dataset:IlyaGusev/ru_turbo_alpaca", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
1
2
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.27.3
true
50,258
null
null
null
null
null
null
cerebras/Cerebras-GPT-1.3B
null
null
"2023-03-20T20:43:21Z"
null
null
1,233
null
null
null
null
48
transformers
[ "transformers", "pytorch", "gpt2", "causal-lm", "text-generation", "en", "dataset:the_pile", "arxiv:2304.03208", "arxiv:2203.15556", "arxiv:2101.00027", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
null
null
null
null
null
null
null
null
gpt2
null
null
null
null
null
null
true
null
null
null
50,257
null
null
null
null
null
null
cerebras/Cerebras-GPT-13B
null
null
"2023-03-20T20:45:54Z"
null
null
2,411
null
null
null
null
645
transformers
[ "transformers", "pytorch", "gpt2", "feature-extraction", "causal-lm", "text-generation", "en", "dataset:the_pile", "arxiv:2304.03208", "arxiv:2203.15556", "arxiv:2101.00027", "license:apache-2.0", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "GPT2Model" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.27.2
true
50,257
null
null
null
null
null
null
Dm271/Gptsmall
null
null
"2023-03-22T18:55:30Z"
null
null
16
null
null
null
null
1
transformers
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "PyTorch", "Transformers", "ru", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
null
null
null
50,257
null
null
null
null
null
null
trl-internal-testing/tiny-random-LlamaForCausalLM
null
null
"2023-03-29T07:11:13Z"
null
null
1,247,370
null
null
null
null
6
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
0
1
silu
16
0.02
64
null
llama
4
2
null
0.000001
null
null
false
float32
4.28.0.dev0
true
32,000
null
null
null
null
null
null
circulus/alpaca-base-13b
null
null
"2023-03-30T05:24:07Z"
null
null
26
null
null
null
null
2
transformers
[ "transformers", "pytorch", "llama", "text-generation", "license:gpl-3.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
0
1
silu
5,120
0.02
13,824
null
llama
40
40
null
0.000001
null
null
false
float16
4.28.0.dev0
true
32,000
null
null
null
null
null
null
medalpaca/medalpaca-13b
null
null
"2023-03-30T13:43:05Z"
null
null
1,225
null
null
null
null
86
transformers
[ "transformers", "pytorch", "llama", "text-generation", "medical", "en", "arxiv:2303.14070", "license:cc", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
0
1
silu
5,120
0.02
13,824
null
llama
40
40
null
0.000001
null
null
false
float32
4.28.0.dev0
true
32,001
null
null
null
null
null
null
IlyaGusev/llama_7b_ru_turbo_alpaca_lora_merged
null
null
"2023-03-30T16:39:01Z"
null
null
61
null
null
null
null
9
transformers
[ "transformers", "pytorch", "llama", "text-generation", "text2text-generation", "ru", "dataset:IlyaGusev/ru_turbo_alpaca", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
null
llama
32
32
null
0.000001
null
null
false
float16
4.28.0.dev0
true
32,000
null
null
null
null
null
null
OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5
null
null
"2023-04-03T20:06:28Z"
null
null
560,300
null
null
null
null
361
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "sft", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
5,120
0.02
20,480
2,048
gpt_neox
40
36
null
null
null
null
false
float16
4.28.0.dev0
true
50,288
null
null
null
null
null
null
huggyllama/llama-30b
null
null
"2023-04-04T00:50:30Z"
null
null
5,257
null
null
null
null
46
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
6,656
0.02
17,920
null
llama
52
60
null
0.000001
null
null
false
float16
4.28.0.dev0
true
32,000
null
null
null
null
null
null
wptoux/bloom-7b-chunhua
null
null
"2023-04-06T06:35:13Z"
null
null
28
null
null
null
null
17
transformers
[ "transformers", "pytorch", "bloom", "text-generation", "art", "zh", "dataset:BelleGroup/train_0.5M_CN", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BloomForCausalLM" ]
1
2
null
4,096
0.02
null
null
bloom
null
null
null
null
null
null
null
float16
4.27.1
true
250,880
null
0
null
null
4
null
vicgalle/gpt2-open-instruct-v1
null
null
"2023-04-08T07:59:46Z"
null
null
991
null
null
null
null
30
transformers
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "instruction-tuning", "alpaca", "en", "dataset:hakurei/open-instruct-v1", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.25.1
false
50,260
null
null
null
null
null
null
yahma/llama-7b-hf
null
null
"2023-04-08T14:39:35Z"
null
null
13,880
null
null
null
null
81
transformers
[ "transformers", "pytorch", "llama", "text-generation", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
null
llama
32
32
null
0.000001
null
null
false
float16
4.28.0.dev0
true
32,000
null
null
null
null
null
null
yahma/llama-13b-hf
null
null
"2023-04-08T15:16:07Z"
null
null
787
null
null
null
null
23
transformers
[ "transformers", "pytorch", "llama", "text-generation", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
null
llama
40
40
null
0.000001
null
null
false
float16
4.28.0.dev0
true
32,000
null
null
null
null
null
null
YeungNLP/firefly-bloom-2b6
null
null
"2023-04-10T03:32:25Z"
null
null
19
null
null
null
null
18
transformers
[ "transformers", "pytorch", "bloom", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BloomForCausalLM" ]
1
2
null
null
0.02
null
null
bloom
32
null
null
null
null
null
null
float16
4.20.0
true
46,145
null
0
null
null
4
null
baffo32/decapoda-research-llama-7B-hf
null
null
"2023-04-10T12:49:58Z"
null
null
17,944
null
null
null
null
52
transformers
[ "transformers", "pytorch", "llama", "text-generation", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
0
1
silu
4,096
0.02
11,008
null
llama
32
32
null
0.000001
null
null
null
float16
4.27.0.dev0
true
32,000
null
null
null
null
null
null
nomic-ai/gpt4all-j
null
null
"2023-04-11T15:39:16Z"
null
null
1,638
null
null
null
null
296
transformers
[ "transformers", "pytorch", "safetensors", "gptj", "text-generation", "en", "dataset:nomic-ai/gpt4all-j-prompt-generations", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTJForCausalLM" ]
50,256
50,256
null
null
0.02
null
null
gptj
null
null
null
null
null
null
false
bfloat16
4.28.0.dev0
false
50,400
null
null
null
null
null
null
databricks/dolly-v2-12b
null
null
"2023-04-11T16:10:54Z"
null
null
8,365
null
null
null
null
1,948
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "en", "dataset:databricks/databricks-dolly-15k", "license:mit", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
5,120
0.02
20,480
2,048
gpt_neox
40
36
null
null
null
null
false
bfloat16
4.25.1
true
50,280
null
null
null
null
null
null
chaoyi-wu/PMC_LLAMA_7B
null
null
"2023-04-12T12:59:34Z"
null
null
1,707
null
null
null
null
60
transformers
[ "transformers", "pytorch", "llama", "text-generation", "medical", "dataset:allenai/s2orc", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
0
1
silu
4,096
0.02
11,008
null
llama
32
32
null
0.000001
null
null
false
float32
4.28.0.dev0
true
32,000
null
null
null
null
null
null
databricks/dolly-v2-7b
null
null
"2023-04-13T05:19:39Z"
null
null
18,763
null
null
null
null
148
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "en", "dataset:databricks/databricks-dolly-15k", "license:mit", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
bfloat16
4.25.1
true
50,280
null
null
null
null
null
null
databricks/dolly-v2-3b
null
null
"2023-04-13T05:20:15Z"
null
null
82,041
null
null
null
null
288
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "en", "dataset:databricks/databricks-dolly-15k", "license:mit", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
bfloat16
4.25.1
true
50,280
null
null
null
null
null
null
reeducator/vicuna-13b-free
null
null
"2023-04-14T12:57:52Z"
null
null
1,347
null
null
null
null
131
transformers
[ "transformers", "llama", "text-generation", "en", "dataset:anon8231489123/ShareGPT_Vicuna_unfiltered", "dataset:gozfarb/ShareGPT_Vicuna_unfiltered", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float32
4.28.0.dev0
true
32,000
null
null
null
null
null
null
autobots/gpt-j-fourchannel-4bit
null
null
"2023-04-15T02:36:00Z"
null
null
45
null
null
null
null
23
transformers
[ "transformers", "gptj", "text-generation", "en", "license:unknown", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTJForCausalLM" ]
50,256
50,256
null
null
0.02
null
null
gptj
null
null
null
null
null
null
false
null
4.18.0.dev0
true
50,400
null
null
null
null
null
null
liyuesen/druggpt
null
null
"2023-04-16T02:46:23Z"
null
null
477
null
null
null
null
18
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "chemistry", "biology", "medical", "doi:10.57967/hf/0627", "license:gpl-3.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.27.2
true
53,083
null
null
null
null
null
null
MBZUAI/LaMini-Neo-1.3B
null
null
"2023-04-16T13:16:07Z"
null
null
34
null
null
null
null
13
transformers
[ "transformers", "pytorch", "gpt_neo", "text-generation", "en", "arxiv:2304.14402", "license:cc-by-nc-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoForCausalLM" ]
50,256
50,256
null
2,048
0.02
null
2,048
gpt_neo
null
null
null
null
null
null
null
float32
4.28.0.dev0
false
50,258
null
0
null
null
null
null
kalpeshk2011/dipper-paraphraser-xxl
null
null
"2023-04-17T06:12:38Z"
null
null
1,499
null
null
null
null
36
transformers
[ "transformers", "pytorch", "safetensors", "t5", "text2text-generation", "xxl", "rewriting", "paraphrase", "paraphrase-generation", "paraphrasing", "text-generation", "en", "arxiv:2303.13408", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "T5ForConditionalGeneration" ]
null
1
null
null
null
null
null
t5
null
null
null
null
null
null
false
float32
4.20.1
true
32,128
null
null
null
null
null
null
fnlp/moss-moon-003-sft
null
null
"2023-04-19T07:46:26Z"
null
null
273
null
null
null
null
126
transformers
[ "transformers", "pytorch", "moss", "text-generation", "llm", "custom_code", "en", "zh", "dataset:fnlp/moss-002-sft-data", "arxiv:2203.13474", "license:agpl-3.0", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "MossForCausalLM" ]
106,028
106,068
null
null
0.02
null
null
moss
null
null
null
null
null
null
false
float16
4.25.1
true
107,008
null
null
null
null
null
null
tekkonetes/fineneo
null
null
"2023-04-19T11:38:11Z"
null
null
17
null
null
null
null
2
transformers
[ "transformers", "pytorch", "gpt_neo", "text-generation", "gpt2", "en", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoForCausalLM" ]
50,256
50,256
null
768
0.02
null
2,048
gpt_neo
null
null
null
null
null
null
null
float32
4.28.1
true
50,257
null
0
null
null
null
null
luodian/llama-7b-hf
null
null
"2023-04-21T18:58:59Z"
null
null
1,005
null
null
null
null
32
transformers
[ "transformers", "pytorch", "llama", "text-generation", "en", "zh", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
null
0.000001
null
null
false
float16
4.29.0.dev0
true
32,000
null
null
null
null
null
null
wangrongsheng/MiniGPT-4-LLaMA-7B
null
null
"2023-04-22T09:06:20Z"
null
null
6,184
null
null
null
null
17
transformers
[ "transformers", "pytorch", "llama", "text-generation", "LLMs", "MiniGPT-4", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
null
0.000001
null
null
false
float16
4.29.0.dev0
true
32,001
null
null
null
null
null
null
tiiuae/falcon-7b
null
null
"2023-04-24T16:36:24Z"
null
null
116,173
null
null
null
null
1,076
transformers
[ "transformers", "pytorch", "safetensors", "falcon", "text-generation", "custom_code", "en", "dataset:tiiuae/falcon-refinedweb", "arxiv:2205.14135", "arxiv:1911.02150", "arxiv:2101.00027", "arxiv:2005.14165", "arxiv:2104.09864", "arxiv:2306.01116", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "FalconForCausalLM" ]
11
11
null
4,544
0.02
null
null
falcon
71
32
null
null
null
null
null
bfloat16
4.27.4
true
65,024
null
0
null
null
null
null
tiiuae/falcon-rw-1b
null
null
"2023-04-26T09:25:36Z"
null
null
95,609
null
null
null
null
104
transformers
[ "transformers", "pytorch", "falcon", "text-generation", "custom_code", "en", "dataset:tiiuae/falcon-refinedweb", "arxiv:2306.01116", "arxiv:2005.14165", "arxiv:2108.12409", "arxiv:2205.14135", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "FalconForCausalLM" ]
1
2
null
2,048
0.02
null
null
falcon
32
24
null
null
null
null
null
bfloat16
4.27.4
true
50,304
null
0
null
null
null
null
TheBloke/stable-vicuna-13B-GPTQ
null
null
"2023-04-28T21:14:29Z"
null
null
81
null
null
null
null
219
transformers
[ "transformers", "safetensors", "llama", "text-generation", "causal-lm", "en", "dataset:OpenAssistant/oasst1", "dataset:nomic-ai/gpt4all_prompt_generations", "dataset:tatsu-lab/alpaca", "arxiv:2302.13971", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float16
4.28.1
true
32,001
null
null
null
null
null
null
jinxuewen/vicuna-13b
null
null
"2023-04-29T06:19:00Z"
null
null
20
null
null
null
null
6
transformers
[ "transformers", "pytorch", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float16
4.28.1
true
32,000
null
null
null
null
null
null
Narrativaai/BioGPT-Large-finetuned-chatdoctor
null
null
"2023-04-29T09:30:48Z"
null
null
160
null
null
null
null
36
transformers
[ "transformers", "pytorch", "tensorboard", "safetensors", "biogpt", "text-generation", "medical", "doctor", "chat", "qa", "question-answering", "en", "dataset:LinhDuong/chatdoctor-200k", "arxiv:2303.14070", "doi:10.57967/hf/0601", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "BioGptForCausalLM" ]
0
2
gelu
1,600
0.02
6,400
2,048
biogpt
25
48
null
null
null
null
null
float32
4.28.1
false
57,717
null
null
null
null
null
null
TehVenom/Pygmalion-7b-4bit-GPTQ-Safetensors
null
null
"2023-04-30T07:45:16Z"
null
null
48
null
null
null
null
35
transformers
[ "transformers", "llama", "text-generation", "text generation", "conversational", "en", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
null
0.000001
null
null
false
bfloat16
4.28.0.dev0
true
32,000
null
null
null
null
null
null
Multi-Domain-Expert-Learning/expert-github
null
null
"2023-05-01T12:36:08Z"
null
null
17
null
null
null
null
1
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float32
4.28.1
true
50,304
null
null
null
null
null
null
liuhaotian/LLaVA-Lightning-7B-delta-v1-1
null
null
"2023-05-03T00:16:20Z"
null
null
1,964
null
null
null
null
16
transformers
[ "transformers", "pytorch", "llava", "text-generation", "license:apache-2.0", "autotrain_compatible", "region:us" ]
text-generation
null
null
1
[ "LlavaLlamaForCausalLM" ]
0
1
silu
4,096
0.02
11,008
2,048
llava
32
32
null
0.000001
null
null
false
float16
4.28.0.dev0
false
32,003
null
null
null
null
null
null
togethercomputer/RedPajama-INCITE-7B-Base
null
null
"2023-05-04T05:50:06Z"
null
null
1,086
null
null
null
null
94
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "en", "dataset:togethercomputer/RedPajama-Data-1T", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.28.1
true
50,432
null
null
null
null
null
null
Norod78/Hebrew-GPT2-345M-Stage
null
null
"2023-05-04T11:57:42Z"
null
null
884
null
null
null
null
1
transformers
[ "transformers", "safetensors", "gguf", "gpt2", "text-generation", "he", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.41.2
true
50,257
null
null
null
null
null
null
cognitivecomputations/WizardLM-7B-Uncensored
null
null
"2023-05-04T20:31:51Z"
null
null
1,388
null
null
null
null
432
transformers
[ "transformers", "pytorch", "llama", "text-generation", "uncensored", "dataset:ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
null
0.000001
null
null
false
float16
4.29.0.dev0
true
32,001
null
null
null
null
null
null
mosaicml/mpt-7b-storywriter
null
null
"2023-05-04T22:55:02Z"
null
null
2,136
null
null
null
null
822
transformers
[ "transformers", "pytorch", "mpt", "text-generation", "Composer", "MosaicML", "llm-foundry", "custom_code", "dataset:the_pile_books3", "arxiv:2108.12409", "arxiv:2205.14135", "arxiv:2302.06675", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "MPTForCausalLM" ]
null
null
null
null
null
null
null
mpt
null
null
null
null
null
null
null
bfloat16
4.28.1
false
50,432
null
null
null
null
null
null
mosaicml/mpt-7b-chat
null
null
"2023-05-04T23:56:17Z"
null
null
8,949
null
null
null
null
512
transformers
[ "transformers", "pytorch", "mpt", "text-generation", "Composer", "MosaicML", "llm-foundry", "custom_code", "dataset:jeffwan/sharegpt_vicuna", "dataset:Hello-SimpleAI/HC3", "dataset:tatsu-lab/alpaca", "dataset:Anthropic/hh-rlhf", "dataset:victor123/evol_instruct_70k", "arxiv:2205.14135", "arxiv:2108.12409", "arxiv:2010.04245", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "MPTForCausalLM" ]
null
null
null
null
null
null
null
mpt
null
null
null
null
null
null
null
bfloat16
4.28.1
false
50,432
null
null
null
null
null
null
mosaicml/mpt-7b
null
null
"2023-05-05T00:48:02Z"
null
null
40,755
null
null
null
null
1,160
transformers
[ "transformers", "pytorch", "mpt", "text-generation", "Composer", "MosaicML", "llm-foundry", "StreamingDatasets", "custom_code", "dataset:mc4", "dataset:c4", "dataset:togethercomputer/RedPajama-Data-1T", "dataset:bigcode/the-stack", "dataset:allenai/s2orc", "arxiv:2108.12409", "arxiv:2302.13971", "arxiv:2205.14135", "arxiv:2010.04245", "arxiv:1909.08053", "arxiv:2302.06675", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "MPTForCausalLM" ]
null
null
null
null
null
null
null
mpt
null
null
null
null
null
null
null
bfloat16
4.28.1
false
50,432
null
null
null
null
null
null
mosaicml/mpt-7b-instruct
null
null
"2023-05-05T00:52:12Z"
null
null
7,902
null
null
null
null
467
transformers
[ "transformers", "pytorch", "mpt", "text-generation", "Composer", "MosaicML", "llm-foundry", "custom_code", "dataset:mosaicml/dolly_hhrlhf", "arxiv:2205.14135", "arxiv:2108.12409", "arxiv:2010.04245", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "MPTForCausalLM" ]
null
null
null
null
null
null
null
mpt
null
null
null
null
null
null
null
bfloat16
4.28.1
false
50,432
null
null
null
null
null
null
togethercomputer/RedPajama-INCITE-Chat-3B-v1
null
null
"2023-05-05T05:06:37Z"
null
null
8,367
null
null
null
null
152
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "en", "dataset:togethercomputer/RedPajama-Data-1T", "dataset:OpenAssistant/oasst1", "dataset:databricks/databricks-dolly-15k", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.28.1
true
50,432
null
null
null
null
null
null
TheBloke/WizardLM-7B-uncensored-GPTQ
null
null
"2023-05-05T08:09:36Z"
null
null
620
null
null
null
null
189
transformers
[ "transformers", "safetensors", "llama", "text-generation", "uncensored", "dataset:ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered", "base_model:cognitivecomputations/WizardLM-7B-Uncensored", "base_model:quantized:cognitivecomputations/WizardLM-7B-Uncensored", "license:other", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
null
0.000001
null
null
false
float16
4.29.0.dev0
true
32,001
null
null
null
null
null
null
cognitivecomputations/WizardLM-13B-Uncensored
null
null
"2023-05-09T18:56:32Z"
null
null
479
null
null
null
null
554
transformers
[ "transformers", "pytorch", "llama", "text-generation", "uncensored", "dataset:ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float16
4.29.0.dev0
true
32,001
null
null
null
null
null
null
coffeeee/nsfw-story-generator2
null
null
"2023-05-10T20:48:26Z"
null
null
1,323
null
null
null
null
37
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.28.1
true
50,257
null
null
null
null
null
null
roneneldan/TinyStories-1M
null
null
"2023-05-12T19:01:50Z"
null
null
23,178
null
null
null
null
40
transformers
[ "transformers", "pytorch", "gpt_neo", "text-generation", "arxiv:2305.07759", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoForCausalLM" ]
50,256
50,256
null
64
0.02
null
2,048
gpt_neo
null
null
null
null
null
null
null
float32
4.28.0
true
50,257
null
0
null
null
null
null
roneneldan/TinyStories-Instruct-2Layers-33M
null
null
"2023-05-12T21:56:46Z"
null
null
694
null
null
null
null
7
transformers
[ "transformers", "pytorch", "gpt_neo", "text-generation", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoForCausalLM" ]
50,256
50,256
null
1,024
0.02
null
2,048
gpt_neo
null
null
null
null
null
null
null
float32
4.28.1
true
50,257
null
0
null
null
null
null
TheBloke/Wizard-Vicuna-13B-Uncensored-GPTQ
null
null
"2023-05-13T08:18:23Z"
null
null
8,429
null
null
null
null
319
transformers
[ "transformers", "safetensors", "llama", "text-generation", "uncensored", "en", "dataset:ehartford/wizard_vicuna_70k_unfiltered", "base_model:cognitivecomputations/Wizard-Vicuna-13B-Uncensored", "base_model:quantized:cognitivecomputations/Wizard-Vicuna-13B-Uncensored", "license:other", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float16
4.28.1
true
32,000
null
null
null
null
null
null
TheBloke/Wizard-Vicuna-13B-Uncensored-HF
null
null
"2023-05-13T09:24:35Z"
null
null
1,722
null
null
null
null
210
transformers
[ "transformers", "pytorch", "llama", "text-generation", "uncensored", "en", "dataset:ehartford/wizard_vicuna_70k_unfiltered", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float16
4.29.2
true
32,000
null
null
null
null
null
null
AlexWortega/wortegaLM-1b
null
null
"2023-05-13T20:41:44Z"
null
null
27
null
null
null
null
2
transformers
[ "transformers", "pytorch", "safetensors", "gpt_neox", "text-generation", "finance", "code", "ru", "dataset:IlyaGusev/rulm", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
2
3
gelu
2,048
0.02
8,192
1,024
gpt_neox
8
16
null
null
null
null
false
float32
4.25.1
true
30,000
null
null
null
null
null
null
shawmoon/EkattorBloom_3b_lora_squad_bn
null
null
"2023-05-14T11:21:38Z"
null
null
17
null
null
null
null
1
transformers
[ "transformers", "bloom", "text-generation", "question-answering", "bn", "en", "dataset:csebuetnlp/squad_bn", "license:openrail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
question-answering
null
null
1
[ "BloomForCausalLM" ]
1
2
null
null
0.02
null
null
bloom
32
null
null
null
null
null
null
null
4.20.0
true
250,880
null
0
null
null
4
null
xzuyn/LLaMa-1-MedicWizard-7B
null
null
"2023-05-14T16:30:20Z"
null
null
20
null
null
null
null
14
transformers
[ "transformers", "pytorch", "llama", "text-generation", "alpaca", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
2,048
llama
32
32
null
0.000001
null
null
false
float16
4.28.1
true
32,001
null
null
null
null
null
null
openaccess-ai-collective/wizard-mega-13b
null
null
"2023-05-14T21:48:43Z"
null
null
1,643
null
null
null
null
107
transformers
[ "transformers", "pytorch", "llama", "text-generation", "en", "dataset:anon8231489123/ShareGPT_Vicuna_unfiltered", "dataset:ehartford/wizard_vicuna_70k_unfiltered", "dataset:ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
bfloat16
4.30.0.dev0
true
32,000
null
null
null
null
null
null
cyberagent/open-calm-3b
null
null
"2023-05-15T07:14:36Z"
null
null
1,783
null
null
null
null
18
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "japanese", "causal-lm", "ja", "dataset:wikipedia", "dataset:cc100", "dataset:mc4", "license:cc-by-sa-4.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.27.0.dev0
true
52,224
null
null
null
null
null
null
bigcode/tiny_starcoder_py
null
null
"2023-05-15T07:43:22Z"
null
null
18,569
null
null
null
null
72
transformers
[ "transformers", "pytorch", "safetensors", "gpt_bigcode", "text-generation", "code", "dataset:bigcode/the-stack-dedup", "license:bigcode-openrail-m", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "GPTBigCodeForCausalLM" ]
0
0
null
null
0.02
null
null
gpt_bigcode
null
null
null
null
null
null
null
float32
4.28.1
true
49,152
null
null
null
null
null
null
cyberagent/open-calm-7b
null
null
"2023-05-15T07:53:34Z"
null
null
7,050
null
null
null
null
205
transformers
[ "transformers", "pytorch", "gpt_neox", "text-generation", "japanese", "causal-lm", "ja", "dataset:wikipedia", "dataset:cc100", "dataset:mc4", "license:cc-by-sa-4.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.27.0.dev0
true
52,224
null
null
null
null
null
null
rinna/japanese-gpt-neox-3.6b
null
null
"2023-05-17T02:16:45Z"
null
null
28,632
null
null
null
null
97
transformers
[ "transformers", "pytorch", "safetensors", "gpt_neox", "text-generation", "lm", "nlp", "ja", "dataset:cc100", "dataset:wikipedia", "dataset:mc4", "arxiv:2404.01657", "license:mit", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "GPTNeoXForCausalLM" ]
2
3
gelu
2,816
0.02
11,264
2,048
gpt_neox
22
36
null
null
null
null
false
float16
null
true
32,000
null
null
null
null
null
null
notstoic/pygmalion-13b-4bit-128g
null
null
"2023-05-18T22:25:44Z"
null
null
1,006
null
null
null
null
146
transformers
[ "transformers", "pytorch", "llama", "text-generation", "en", "license:other", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float16
4.30.0.dev0
true
32,000
null
null
null
null
null
null
digitous/13B-HyperMantis
null
null
"2023-05-19T23:13:32Z"
null
null
827
null
null
null
null
29
transformers
[ "transformers", "pytorch", "llama", "text-generation", "alpaca", "vicuna", "mix", "merge", "model merge", "roleplay", "chat", "instruct", "en", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
2,048
llama
40
40
null
0.000001
null
null
false
float16
4.28.1
true
32,000
null
null
null
null
null
null
Vision-CAIR/vicuna-7b
null
null
"2023-05-22T09:56:55Z"
null
null
1,413
null
null
null
null
19
transformers
[ "transformers", "pytorch", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
0
1
silu
4,096
0.02
11,008
2,048
llama
32
32
null
0.000001
null
null
false
float16
4.29.0.dev0
true
32,001
null
null
null
null
null
null
TheBloke/WizardLM-30B-Uncensored-GPTQ
null
null
"2023-05-22T12:56:35Z"
null
null
92
null
null
null
null
118
transformers
[ "transformers", "safetensors", "llama", "text-generation", "uncensored", "dataset:ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered", "base_model:cognitivecomputations/WizardLM-30B-Uncensored", "base_model:quantized:cognitivecomputations/WizardLM-30B-Uncensored", "license:other", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
null
null
1
[ "LlamaForCausalLM" ]
1
2
silu
6,656
0.02
17,920
2,048
llama
52
60
null
0.000001
null
null
false
float16
4.30.0.dev0
true
32,001
null
null
null
null
null
null