Search is not available for this dataset
pipeline_tag
stringclasses
48 values
library_name
stringclasses
205 values
text
stringlengths
0
18.3M
metadata
stringlengths
2
1.07B
id
stringlengths
5
122
last_modified
null
tags
listlengths
1
1.84k
sha
null
created_at
stringlengths
25
25
feature-extraction
transformers
{}
Shauli/RE-metric-model-siamese-spike
null
[ "transformers", "pytorch", "jax", "bert", "feature-extraction", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
feature-extraction
transformers
{}
Shauli/RE-metric-model-spike
null
[ "transformers", "pytorch", "jax", "bert", "feature-extraction", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ShauryaP24/DialoGPT-small-pratyush
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
null
## RusEnQA QA for Russian and English based on the [rugpt3xl](https://huggingface.co/sberbank-ai/rugpt3xl) model ### Fine-tuning format: ``` "<s>paragraph: "+eng_context+"\nlang: rus\nquestion: "+rus_question+' answer: '+ rus_answer+"</s>" ``` ### About ruGPT-3 XL model Model was trained with 512 sequence length using [Deepspeed](https://github.com/microsoft/DeepSpeed) and [Megatron](https://github.com/NVIDIA/Megatron-LM) code by [SberDevices](https://sberdevices.ru/) team, on 80B tokens dataset for 4 epochs. After that model was finetuned 1 epoch with sequence length 2048. *Note! Model has sparse attention blocks.* Total training time was around 10 days on 256 GPUs. Final perplexity on test set is 12.05. Model parameters: 1.3B.
{"language": ["ru", "en"], "tags": ["PyTorch", "Transformers", "gpt2", "squad", "lm-head", "casual-lm"], "pipeline_tag": "text2text-generation", "thumbnail": "https://github.com/RussianNLP/RusEnQA"}
Shavrina/RusEnQA
null
[ "PyTorch", "Transformers", "gpt2", "squad", "lm-head", "casual-lm", "text2text-generation", "ru", "en", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
Shay77/Mine
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
Shaya/POS_pretraining
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
Shaya007/Bert_hindi
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
Shayo/pickle_test
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ShayoGun/DialoGPT-medium-shay0
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ShayoGun/DialoGPT-medium-shayo
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
# SHAY0 Dialo GPT Model
{"tags": ["conversational"]}
ShayoGun/DialoGPT-small-shayo
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ShayoGun/policy-distilbert-7d
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
# Harry Potter DialGPT Model
{"tags": ["conversational"]}
Sheel/DialoGPT-small-harrypotter
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
# Mikasa DialoGPT Model
{"tags": ["conversational"]}
Sheerwin02/DialoGPT-medium-mikasa
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
#isla DialoGPT Model
{"tags": ["conversational"]}
Sheerwin02/DialoGPT-small-isla
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
This is a repo with gather thoughts and experiments on the state-of-the-art techniques in NLP.
{}
ShenSeanchen/NLP
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
Shendy/Teste
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ShengdingHu/97834603
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_roberta-base_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_roberta-large_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_cola
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_mnli
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_mrpc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_qnli
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_qqp
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_sst2
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_stsb
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_superglue-boolq
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_superglue-cb
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_superglue-copa
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_superglue-multirc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_superglue-record
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_superglue-wic
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/adapter_t5-base_superglue-wsc.fixed
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/autodelta_try
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_roberta-base_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_cola
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_mnli
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_mrpc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_qnli
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_qqp
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_sst2
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_stsb
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_superglue-boolq
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_superglue-cb
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_superglue-copa
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_superglue-multirc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_superglue-record
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_superglue-wic
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/bitfit_t5-base_superglue-wsc.fixed
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/cola
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/compacter_t5-base_cola
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/compacter_t5-base_mnli
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/compacter_t5-base_mrpc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/compactor_roberta-base_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_roberta-base_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_cola
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_mnli
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_qnli
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_qqp
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_sst2
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_stsb
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_superglue-cb
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_superglue-copa
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_superglue-multirc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_superglue-record
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_superglue-wic
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/lora_t5-base_superglue-wsc.fixed
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/low_rank_adapter_roberta-base_rte
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/mnli
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/mrpc
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/prefix_roberta-base_mrpc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/prefix_t5-base_mrpc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/qnli
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/qqp
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/rte
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/soft_prompt_t5-base_cola
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/soft_prompt_t5-base_mrpc
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/soft_prompt_t5-base_sst2
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
fill-mask
transformers
{}
ShengdingHu/sst2
null
[ "transformers", "pytorch", "tensorboard", "big_bird", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/stsb
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/superglue-boolq-multig
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # superglue-boolq This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.2098 - Accuracy: 76.7584 - Average Metrics: 76.7584 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Average Metrics | |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------------:| | No log | 0.34 | 100 | 0.2293 | 73.2722 | 73.2722 | | No log | 0.68 | 200 | 0.2098 | 76.7584 | 76.7584 | ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.2+cu111 - Datasets 1.17.0 - Tokenizers 0.12.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "model-index": [{"name": "superglue-boolq", "results": []}]}
ShengdingHu/superglue-boolq
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/superglue-cb
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/superglue-copa
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/superglue-multirc
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/superglue-record
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/superglue-wic
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text2text-generation
transformers
{}
ShengdingHu/superglue-wsc.fixed
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ShengdingHu/t5-base_bitfit
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ShengdingHu/t5-base_bitfit_mrpc
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
{}
ShengdingHu/test_delta_model
null
[ "transformers", "pytorch", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-ner This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0620 - Precision: 0.9267 - Recall: 0.9371 - F1: 0.9319 - Accuracy: 0.9838 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.2462 | 1.0 | 878 | 0.0714 | 0.9052 | 0.9223 | 0.9137 | 0.9803 | | 0.0535 | 2.0 | 1756 | 0.0615 | 0.9188 | 0.9331 | 0.9259 | 0.9827 | | 0.0315 | 3.0 | 2634 | 0.0620 | 0.9267 | 0.9371 | 0.9319 | 0.9838 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["conll2003"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "distilbert-base-uncased-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "conll2003", "type": "conll2003", "args": "conll2003"}, "metrics": [{"type": "precision", "value": 0.9266592920353982, "name": "Precision"}, {"type": "recall", "value": 0.9371294328224634, "name": "Recall"}, {"type": "f1", "value": 0.9318649535569274, "name": "F1"}, {"type": "accuracy", "value": 0.9838117781625813, "name": "Accuracy"}]}]}]}
Shenyancheng/distilbert-base-uncased-finetuned-ner
null
[ "transformers", "pytorch", "tensorboard", "distilbert", "token-classification", "generated_from_trainer", "dataset:conll2003", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
null
# My Awesome Model
{"tags": ["conversational"]}
Sherman/DialoGPT-medium-joey
null
[ "conversational", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
# Harry Potter DialoGPT Model
{"tags": ["conversational"]}
Shike/DialoGPT_medium_harrypotter
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
# My Hero Academia DialoGPT Model
{"tags": ["conversational"]}
Shinx/DialoGPT-medium-myheroacademia
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
# My Awesome Model
{"tags": ["conversational"]}
NaturesDisaster/DialoGPT-large-Neku
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00