Search is not available for this dataset
pipeline_tag
stringclasses
48 values
library_name
stringclasses
205 values
text
stringlengths
0
18.3M
metadata
stringlengths
2
1.07B
id
stringlengths
5
122
last_modified
null
tags
listlengths
1
1.84k
sha
null
created_at
stringlengths
25
25
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical-imbalanced-PubMedBERT-base-uncased-abstract_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical-imbalanced-biobert-v1.1_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical-imbalanced-biobert
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical-imbalanced-bluebert_pubmed_uncased_L-12_H-768_A-12_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical-imbalanced-pubmedbert
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical-imbalanced-scibert_scivocab_uncased_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghadeermobasher/BC5CDR-Chemical_Imbalanced-biobert-large-cased-v1.1
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical_Imbalanced-scibert_scivocab_cased
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical_ImbalancedBioM-ELECTRA-Base-Discriminator
null
[ "transformers", "pytorch", "electra", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical_Modified_BioM-ELECTRA-Base-Discriminator
null
[ "transformers", "pytorch", "electra", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical_Modified_PubMedBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Chemical_Modified_scibert_scivocab_cased
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease-Modified_BiomedNLP-PubMedBERT-base-uncased-abstract
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/WELT-BioBERT-BC5CDRDisease
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease-Modified_bluebert_pubmed_uncased_L-12_H-768_A-12_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease-Modified_scibert_scivocab_uncased
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease-balancedPubMedBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease-imbalanced-BiomedNLP-PubMedBERT-base-uncased-abstract
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Original-BioBERT-BC5CDRDisease
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease-imbalanced-bluebert_pubmed_uncased_L-12_H-768_A-12_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease-imbalanced-scibert_scivocab_uncased
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease_Modified_BioM-ELECTRA-Base-Discriminator
null
[ "transformers", "pytorch", "electra", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease_Modified_PubMedBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghadeermobasher/BC5CDR-Disease_Modified_biobert-large-cased-v1.1
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Disease_Modified_scibert_scivocab_cased
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Imbalanced-PubMedBERT
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Imbalanced-SapBERT-from-PubMedBERT-fulltext
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Imbalanced-biobert-base-cased-v1.2
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghadeermobasher/BC5CDR-Imbalanced-biobert-large-cased-v1.1
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BC5CDR-Imbalanced-scibert_scivocab_cased
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BCHEM4-Modified-BioBERT-v1
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified-Bioformers
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified-Bioformers_2
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified-PubMedBert-abstract-3
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified-PubMedBert-full-3
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified-pubmedabstract_latest
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified_BioM-ELECTRA-Base-Discriminator
null
[ "transformers", "pytorch", "electra", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified_PubMedBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified_SciBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem-Modified_biobert-large
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem_Imbalanced-biobert
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem_Imbalanced-scibert_scivocab_cased
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem_ImbalancedBioM-ELECTRA-Base-Discriminator
null
[ "transformers", "pytorch", "electra", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Chem_ImbalancedPubMedBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Modified-biobert-v1.1_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Modified-bluebert_pubmed_uncased_L-12_H-768_A-12_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Modified-pubmedabstract_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Modified-scibert-uncased_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Original-PubmedBert-abstract-latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Original-biobert-v1.1_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Original-bluebert_pubmed_uncased_L-12_H-768_A-12_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/BioNLP13CG-Original-scibert_latest
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified-Bioformers
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified-BiomedNLP-PubMedBERT-base-uncased-abstract
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified-biobert-v1.1
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified-bluebert_pubmed_uncased_L-12_H-768_A-12
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified-scibert_scivocab_uncased
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified_BioM-ELECTRA-Base-Discriminator
null
[ "transformers", "pytorch", "electra", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified_PubMedBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified_SciBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem-Modified_biobert-large-cased
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem_Imbalanced-SciBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghadeermobasher/CRAFT-Chem_Imbalanced-biobert-large-cased-v1.1
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem_ImbalancedBioM-ELECTRA-Base-Discriminator
null
[ "transformers", "pytorch", "electra", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem_ImbalancedPubMedBERT
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem_Original-BiomedNLP-PubMedBERT-base-uncased-abstract
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem_Original-biobert-v1.1
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem_Original-bluebert_pubmed_uncased_L-12_H-768_A-12
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/CRAFT-Chem_Original-scibert_scivocab_uncased
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghadeermobasher/Model1
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghadeermobasher/Model2
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Model_co_imb
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Model_col-mod
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Model_imb
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Model_imb_1
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Model_imb_2
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghadeermobasher/Model_imb_3
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Model_org
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Model_org_1
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/Model_org_2
null
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghadeermobasher/Model_org_3
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
token-classification
transformers
{}
ghadeermobasher/bc4chemd-imbalanced-biobert-base-casesd-v1.1
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-classification
transformers
A fake news detector using RoBERTa. Dataset: https://www.kaggle.com/clmentbisaillon/fake-and-real-news-dataset Training involved using hyperparameter search with 10 trials.
{}
ghanashyamvtatti/roberta-fake-news
null
[ "transformers", "pytorch", "tf", "jax", "roberta", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-classification
transformers
{}
ghazikhanihamed/A-TCDB-BERT-C
null
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-classification
transformers
{"license": "afl-3.0"}
ghazikhanihamed/MembraneBERT
null
[ "transformers", "pytorch", "bert", "text-classification", "license:afl-3.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-classification
transformers
{}
ghazikhanihamed/TCDB-BERT-C
null
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
fill-mask
transformers
{}
ghazikhanihamed/TCDB-BERT
null
[ "transformers", "pytorch", "bert", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-classification
transformers
{}
ghazikhanihamed/IonchannelBERT
null
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-classification
transformers
{}
ghazikhanihamed/TooT-BERT-M
null
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-classification
transformers
{}
ghazikhanihamed/TransporterBERT
null
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
transformers
This repository belongs to TransportersBERT from ActTrans publication. Taju, Semmy Wellem, Syed Muazzam Ali Shah, and Yu-Yen Ou. “ActTRANS: Functional Classification in Active Transport Proteins Based on Transfer Learning and Contextual Representations.” Computational Biology and Chemistry 93 (August 1, 2021): 107537. https://doi.org/10.1016/j.compbiolchem.2021.107537.
{}
ghazikhanihamed/TransportersBERT
null
[ "transformers", "pytorch", "bert", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
# Connor
{"tags": ["conversational"]}
ghhostboy/DialoGPT-medium-connorDBH3-1
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
text-generation
transformers
# Connor
{"tags": ["conversational"]}
ghhostboy/DialoGPT-medium-connorDBH3-21
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # common6 This model is a fine-tuned version of [common6/checkpoint-3500](https://huggingface.co/common6/checkpoint-3500) on the COMMON_VOICE - FA dataset. It achieves the following results on the evaluation set: - Loss: 0.3706 - Wer: 0.3421 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 6e-05 - train_batch_size: 32 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 256 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 200.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 1.0344 | 10.0 | 500 | 0.4043 | 0.4511 | | 0.9651 | 20.0 | 1000 | 0.3793 | 0.4159 | | 0.9125 | 30.0 | 1500 | 0.3756 | 0.4046 | | 0.8831 | 40.0 | 2000 | 0.3650 | 0.3876 | | 0.8399 | 50.0 | 2500 | 0.3605 | 0.3772 | | 0.819 | 60.0 | 3000 | 0.3622 | 0.3714 | | 0.8029 | 70.0 | 3500 | 0.3561 | 0.3664 | | 0.8104 | 80.0 | 4000 | 0.3595 | 0.3660 | | 0.8118 | 90.0 | 4500 | 0.3460 | 0.3592 | | 0.7831 | 100.0 | 5000 | 0.3566 | 0.3593 | | 0.744 | 110.0 | 5500 | 0.3578 | 0.3535 | | 0.7388 | 120.0 | 6000 | 0.3538 | 0.3520 | | 0.714 | 130.0 | 6500 | 0.3682 | 0.3506 | | 0.7291 | 140.0 | 7000 | 0.3625 | 0.3505 | | 0.697 | 150.0 | 7500 | 0.3619 | 0.3479 | | 0.6811 | 160.0 | 8000 | 0.3631 | 0.3440 | | 0.6841 | 170.0 | 8500 | 0.3672 | 0.3460 | | 0.6616 | 180.0 | 9000 | 0.3677 | 0.3410 | | 0.6471 | 190.0 | 9500 | 0.3707 | 0.3420 | | 0.6759 | 200.0 | 10000 | 0.3706 | 0.3421 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2 - Datasets 1.18.3.dev0 - Tokenizers 0.10.3
{"language": ["fa"], "tags": ["automatic-speech-recognition", "common_voice", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "common6", "results": []}]}
ghofrani/common6
null
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "common_voice", "generated_from_trainer", "fa", "dataset:common_voice", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # common7 This model is a fine-tuned version of [common7/checkpoint-18500](https://huggingface.co/common7/checkpoint-18500) on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - FA dataset. It achieves the following results on the evaluation set: - Loss: 0.3448 - Wer: 0.3478 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 6e-05 - train_batch_size: 32 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 150.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:-----:|:---------------:|:------:| | 2.957 | 3.29 | 500 | 2.9503 | 1.0 | | 1.7225 | 6.58 | 1000 | 0.8860 | 0.7703 | | 1.4907 | 9.86 | 1500 | 0.6555 | 0.6673 | | 1.4177 | 13.16 | 2000 | 0.5784 | 0.6076 | | 1.3425 | 16.45 | 2500 | 0.5379 | 0.5718 | | 1.33 | 19.73 | 3000 | 0.4962 | 0.5245 | | 1.4378 | 23.03 | 3500 | 0.4699 | 0.5098 | | 1.1894 | 26.31 | 4000 | 0.4527 | 0.4848 | | 1.1844 | 29.6 | 4500 | 0.4309 | 0.4651 | | 1.1795 | 32.89 | 5000 | 0.4131 | 0.4524 | | 1.1471 | 36.18 | 5500 | 0.4052 | 0.4435 | | 1.1337 | 39.47 | 6000 | 0.3927 | 0.4363 | | 1.1896 | 42.76 | 6500 | 0.3811 | 0.4254 | | 1.1847 | 46.05 | 7000 | 0.3855 | 0.4129 | | 0.9954 | 49.34 | 7500 | 0.3729 | 0.3981 | | 1.0293 | 52.63 | 8000 | 0.3637 | 0.4014 | | 1.0224 | 55.92 | 8500 | 0.3578 | 0.3885 | | 1.012 | 59.21 | 9000 | 0.3629 | 0.3930 | | 1.0772 | 62.5 | 9500 | 0.3635 | 0.3906 | | 1.0344 | 65.79 | 10000 | 0.3469 | 0.3771 | | 0.9457 | 69.08 | 10500 | 0.3435 | 0.3735 | | 0.9307 | 72.37 | 11000 | 0.3519 | 0.3762 | | 0.9523 | 75.65 | 11500 | 0.3443 | 0.3666 | | 0.9523 | 78.94 | 12000 | 0.3502 | 0.3757 | | 0.9475 | 82.24 | 12500 | 0.3509 | 0.3643 | | 0.9971 | 85.52 | 13000 | 0.3502 | 0.3626 | | 0.9058 | 88.81 | 13500 | 0.3472 | 0.3605 | | 0.8922 | 92.1 | 14000 | 0.3530 | 0.3618 | | 0.9 | 95.39 | 14500 | 0.3500 | 0.3574 | | 0.9051 | 98.68 | 15000 | 0.3456 | 0.3535 | | 0.9304 | 101.97 | 15500 | 0.3438 | 0.3578 | | 0.9433 | 105.26 | 16000 | 0.3396 | 0.3530 | | 0.8988 | 108.55 | 16500 | 0.3436 | 0.3539 | | 0.8789 | 111.84 | 17000 | 0.3426 | 0.3516 | | 0.8667 | 115.13 | 17500 | 0.3438 | 0.3506 | | 0.8895 | 118.42 | 18000 | 0.3434 | 0.3503 | | 0.8888 | 121.71 | 18500 | 0.3425 | 0.3494 | | 0.9453 | 125.0 | 19000 | 0.3415 | 0.3480 | | 0.9267 | 128.29 | 19500 | 0.3477 | 0.3503 | | 0.8315 | 131.58 | 20000 | 0.3476 | 0.3505 | | 0.8542 | 134.86 | 20500 | 0.3475 | 0.3506 | | 0.8478 | 138.16 | 21000 | 0.3430 | 0.3481 | | 0.8643 | 141.45 | 21500 | 0.3451 | 0.3485 | | 0.8705 | 144.73 | 22000 | 0.3444 | 0.3474 | | 0.9869 | 148.03 | 22500 | 0.3441 | 0.3493 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2 - Datasets 1.18.3.dev0 - Tokenizers 0.10.3
{"language": ["fa"], "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_7_0", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "common7", "results": []}]}
ghofrani/common7
null
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "mozilla-foundation/common_voice_7_0", "generated_from_trainer", "fa", "dataset:common_voice", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # common8 This model is a fine-tuned version of [wghts/checkpoint-20000](https://huggingface.co/wghts/checkpoint-20000) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - FA dataset. It achieves the following results on the evaluation set: - Loss: 0.3174 - Wer: 0.3022 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-06 - train_batch_size: 32 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 6 - total_train_batch_size: 192 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 250.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:-----:|:---------------:|:------:| | 3.5847 | 1.93 | 500 | 3.5104 | 1.0 | | 2.7858 | 3.86 | 1000 | 2.9601 | 1.0001 | | 1.6827 | 5.79 | 1500 | 0.7853 | 0.7030 | | 1.4656 | 7.72 | 2000 | 0.6076 | 0.6014 | | 1.3693 | 9.65 | 2500 | 0.5114 | 0.5307 | | 1.379 | 11.58 | 3000 | 0.4666 | 0.4940 | | 1.2832 | 13.51 | 3500 | 0.4257 | 0.4593 | | 1.1931 | 15.44 | 4000 | 0.4039 | 0.4427 | | 1.2911 | 17.37 | 4500 | 0.3956 | 0.4295 | | 1.1577 | 19.3 | 5000 | 0.3705 | 0.4114 | | 1.1135 | 21.24 | 5500 | 0.3740 | 0.4010 | | 1.19 | 23.17 | 6000 | 0.3611 | 0.3935 | | 1.1008 | 25.1 | 6500 | 0.3503 | 0.3880 | | 1.0805 | 27.03 | 7000 | 0.3427 | 0.3781 | | 1.1556 | 28.96 | 7500 | 0.3442 | 0.3727 | | 1.0596 | 30.89 | 8000 | 0.3398 | 0.3646 | | 1.0219 | 32.82 | 8500 | 0.3312 | 0.3660 | | 1.1042 | 34.75 | 9000 | 0.3287 | 0.3612 | | 1.0273 | 36.68 | 9500 | 0.3236 | 0.3556 | | 1.0383 | 38.61 | 10000 | 0.3217 | 0.3558 | | 1.0498 | 40.54 | 10500 | 0.3205 | 0.3520 | | 0.9969 | 42.47 | 11000 | 0.3125 | 0.3504 | | 1.0658 | 44.4 | 11500 | 0.3120 | 0.3493 | | 0.992 | 46.33 | 12000 | 0.3137 | 0.3476 | | 0.9737 | 48.26 | 12500 | 0.3085 | 0.3413 | | 1.0817 | 50.19 | 13000 | 0.3091 | 0.3418 | | 0.9414 | 52.12 | 13500 | 0.3072 | 0.3344 | | 0.9295 | 54.05 | 14000 | 0.3039 | 0.3322 | | 1.0248 | 55.98 | 14500 | 0.2991 | 0.3325 | | 0.9474 | 57.91 | 15000 | 0.3032 | 0.3348 | | 0.928 | 59.85 | 15500 | 0.2999 | 0.3285 | | 1.0321 | 61.78 | 16000 | 0.2982 | 0.3253 | | 0.9255 | 63.71 | 16500 | 0.2970 | 0.3231 | | 0.8928 | 65.64 | 17000 | 0.2993 | 0.3250 | | 1.008 | 67.57 | 17500 | 0.2985 | 0.3222 | | 0.9371 | 69.5 | 18000 | 0.2968 | 0.3216 | | 0.9077 | 71.43 | 18500 | 0.3011 | 0.3299 | | 1.0044 | 73.36 | 19000 | 0.3053 | 0.3306 | | 0.9625 | 75.29 | 19500 | 0.3159 | 0.3295 | | 0.9816 | 77.22 | 20000 | 0.3080 | 0.3304 | | 0.9587 | 119.19 | 20500 | 0.3088 | 0.3284 | | 0.9178 | 122.09 | 21000 | 0.3132 | 0.3320 | | 1.0282 | 125.0 | 21500 | 0.3099 | 0.3266 | | 0.9337 | 127.9 | 22000 | 0.3110 | 0.3317 | | 0.8822 | 130.81 | 22500 | 0.3037 | 0.3247 | | 0.9644 | 133.72 | 23000 | 0.3037 | 0.3238 | | 0.9214 | 136.62 | 23500 | 0.3040 | 0.3234 | | 0.9167 | 139.53 | 24000 | 0.3079 | 0.3203 | | 0.9047 | 142.44 | 24500 | 0.3018 | 0.3177 | | 0.8909 | 145.35 | 25000 | 0.3053 | 0.3181 | | 0.9646 | 148.25 | 25500 | 0.3095 | 0.3229 | | 0.8802 | 151.16 | 26000 | 0.3111 | 0.3192 | | 0.8411 | 154.07 | 26500 | 0.3068 | 0.3123 | | 0.9235 | 156.97 | 27000 | 0.3090 | 0.3177 | | 0.8943 | 159.88 | 27500 | 0.3115 | 0.3179 | | 0.8854 | 162.79 | 28000 | 0.3052 | 0.3157 | | 0.8734 | 165.69 | 28500 | 0.3077 | 0.3124 | | 0.8515 | 168.6 | 29000 | 0.3117 | 0.3128 | | 0.912 | 171.51 | 29500 | 0.3039 | 0.3121 | | 0.8669 | 174.42 | 30000 | 0.3120 | 0.3123 | | 0.823 | 177.32 | 30500 | 0.3148 | 0.3118 | | 0.9129 | 180.23 | 31000 | 0.3179 | 0.3101 | | 0.8255 | 183.14 | 31500 | 0.3164 | 0.3114 | | 0.8948 | 186.05 | 32000 | 0.3128 | 0.3101 | | 0.8397 | 188.95 | 32500 | 0.3143 | 0.3068 | | 0.8341 | 191.86 | 33000 | 0.3127 | 0.3136 | | 0.873 | 194.76 | 33500 | 0.3149 | 0.3124 | | 0.8232 | 197.67 | 34000 | 0.3166 | 0.3086 | | 0.8002 | 200.58 | 34500 | 0.3149 | 0.3061 | | 0.8621 | 203.49 | 35000 | 0.3160 | 0.3093 | | 0.8123 | 206.39 | 35500 | 0.3141 | 0.3063 | | 0.7995 | 209.3 | 36000 | 0.3174 | 0.3075 | | 0.8271 | 212.21 | 36500 | 0.3173 | 0.3043 | | 0.8059 | 215.12 | 37000 | 0.3176 | 0.3079 | | 0.8835 | 218.02 | 37500 | 0.3169 | 0.3062 | | 0.8027 | 220.93 | 38000 | 0.3203 | 0.3098 | | 0.775 | 223.83 | 38500 | 0.3159 | 0.3068 | | 0.8487 | 226.74 | 39000 | 0.3161 | 0.3072 | | 0.7929 | 229.65 | 39500 | 0.3143 | 0.3037 | | 0.7653 | 232.56 | 40000 | 0.3160 | 0.3048 | | 0.8211 | 235.46 | 40500 | 0.3173 | 0.3031 | | 0.7761 | 238.37 | 41000 | 0.3176 | 0.3025 | | 0.7761 | 241.28 | 41500 | 0.3179 | 0.3027 | | 0.7903 | 244.19 | 42000 | 0.3181 | 0.3016 | | 0.7807 | 247.09 | 42500 | 0.3170 | 0.3027 | | 0.8406 | 250.0 | 43000 | 0.3174 | 0.3022 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2 - Datasets 1.18.3.dev0 - Tokenizers 0.10.3
{"language": ["fa"], "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "common8", "results": []}]}
ghofrani/xls-r-1b-fa-cv8
null
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "fa", "dataset:common_voice", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghofrani/distilbert-base-uncased-finetuned-ner
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
automatic-speech-recognition
transformers
{}
ghofrani/wghts
null
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
null
null
{}
ghofrani/xls-r-1b-fa
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
automatic-speech-recognition
transformers
{}
ghofrani/xls-r-300m-fa
null
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00