Search is not available for this dataset
pipeline_tag
stringclasses 48
values | library_name
stringclasses 205
values | text
stringlengths 0
18.3M
| metadata
stringlengths 2
1.07B
| id
stringlengths 5
122
| last_modified
null | tags
listlengths 1
1.84k
| sha
null | created_at
stringlengths 25
25
|
---|---|---|---|---|---|---|---|---|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical-imbalanced-PubMedBERT-base-uncased-abstract_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical-imbalanced-biobert-v1.1_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical-imbalanced-biobert
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical-imbalanced-bluebert_pubmed_uncased_L-12_H-768_A-12_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical-imbalanced-pubmedbert
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical-imbalanced-scibert_scivocab_uncased_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghadeermobasher/BC5CDR-Chemical_Imbalanced-biobert-large-cased-v1.1
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical_Imbalanced-scibert_scivocab_cased
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical_ImbalancedBioM-ELECTRA-Base-Discriminator
| null |
[
"transformers",
"pytorch",
"electra",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical_Modified_BioM-ELECTRA-Base-Discriminator
| null |
[
"transformers",
"pytorch",
"electra",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical_Modified_PubMedBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Chemical_Modified_scibert_scivocab_cased
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease-Modified_BiomedNLP-PubMedBERT-base-uncased-abstract
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/WELT-BioBERT-BC5CDRDisease
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease-Modified_bluebert_pubmed_uncased_L-12_H-768_A-12_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease-Modified_scibert_scivocab_uncased
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease-balancedPubMedBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease-imbalanced-BiomedNLP-PubMedBERT-base-uncased-abstract
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Original-BioBERT-BC5CDRDisease
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease-imbalanced-bluebert_pubmed_uncased_L-12_H-768_A-12_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease-imbalanced-scibert_scivocab_uncased
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease_Modified_BioM-ELECTRA-Base-Discriminator
| null |
[
"transformers",
"pytorch",
"electra",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease_Modified_PubMedBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghadeermobasher/BC5CDR-Disease_Modified_biobert-large-cased-v1.1
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Disease_Modified_scibert_scivocab_cased
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Imbalanced-PubMedBERT
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Imbalanced-SapBERT-from-PubMedBERT-fulltext
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Imbalanced-biobert-base-cased-v1.2
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghadeermobasher/BC5CDR-Imbalanced-biobert-large-cased-v1.1
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BC5CDR-Imbalanced-scibert_scivocab_cased
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BCHEM4-Modified-BioBERT-v1
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified-Bioformers
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified-Bioformers_2
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified-PubMedBert-abstract-3
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified-PubMedBert-full-3
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified-pubmedabstract_latest
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified_BioM-ELECTRA-Base-Discriminator
| null |
[
"transformers",
"pytorch",
"electra",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified_PubMedBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified_SciBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem-Modified_biobert-large
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem_Imbalanced-biobert
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem_Imbalanced-scibert_scivocab_cased
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem_ImbalancedBioM-ELECTRA-Base-Discriminator
| null |
[
"transformers",
"pytorch",
"electra",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Chem_ImbalancedPubMedBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Modified-biobert-v1.1_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Modified-bluebert_pubmed_uncased_L-12_H-768_A-12_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Modified-pubmedabstract_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Modified-scibert-uncased_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Original-PubmedBert-abstract-latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Original-biobert-v1.1_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Original-bluebert_pubmed_uncased_L-12_H-768_A-12_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/BioNLP13CG-Original-scibert_latest
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified-Bioformers
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified-BiomedNLP-PubMedBERT-base-uncased-abstract
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified-biobert-v1.1
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified-bluebert_pubmed_uncased_L-12_H-768_A-12
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified-scibert_scivocab_uncased
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified_BioM-ELECTRA-Base-Discriminator
| null |
[
"transformers",
"pytorch",
"electra",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified_PubMedBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified_SciBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem-Modified_biobert-large-cased
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem_Imbalanced-SciBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghadeermobasher/CRAFT-Chem_Imbalanced-biobert-large-cased-v1.1
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem_ImbalancedBioM-ELECTRA-Base-Discriminator
| null |
[
"transformers",
"pytorch",
"electra",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem_ImbalancedPubMedBERT
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem_Original-BiomedNLP-PubMedBERT-base-uncased-abstract
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem_Original-biobert-v1.1
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem_Original-bluebert_pubmed_uncased_L-12_H-768_A-12
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/CRAFT-Chem_Original-scibert_scivocab_uncased
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghadeermobasher/Model1
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghadeermobasher/Model2
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Model_co_imb
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Model_col-mod
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Model_imb
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Model_imb_1
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Model_imb_2
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghadeermobasher/Model_imb_3
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Model_org
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Model_org_1
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/Model_org_2
| null |
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghadeermobasher/Model_org_3
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
token-classification
|
transformers
|
{}
|
ghadeermobasher/bc4chemd-imbalanced-biobert-base-casesd-v1.1
| null |
[
"transformers",
"pytorch",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
text-classification
|
transformers
|
A fake news detector using RoBERTa.
Dataset: https://www.kaggle.com/clmentbisaillon/fake-and-real-news-dataset
Training involved using hyperparameter search with 10 trials.
|
{}
|
ghanashyamvtatti/roberta-fake-news
| null |
[
"transformers",
"pytorch",
"tf",
"jax",
"roberta",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
text-classification
|
transformers
|
{}
|
ghazikhanihamed/A-TCDB-BERT-C
| null |
[
"transformers",
"pytorch",
"bert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
text-classification
|
transformers
|
{"license": "afl-3.0"}
|
ghazikhanihamed/MembraneBERT
| null |
[
"transformers",
"pytorch",
"bert",
"text-classification",
"license:afl-3.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
text-classification
|
transformers
|
{}
|
ghazikhanihamed/TCDB-BERT-C
| null |
[
"transformers",
"pytorch",
"bert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
fill-mask
|
transformers
|
{}
|
ghazikhanihamed/TCDB-BERT
| null |
[
"transformers",
"pytorch",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
text-classification
|
transformers
|
{}
|
ghazikhanihamed/IonchannelBERT
| null |
[
"transformers",
"pytorch",
"bert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
text-classification
|
transformers
|
{}
|
ghazikhanihamed/TooT-BERT-M
| null |
[
"transformers",
"pytorch",
"bert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
text-classification
|
transformers
|
{}
|
ghazikhanihamed/TransporterBERT
| null |
[
"transformers",
"pytorch",
"bert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null |
transformers
|
This repository belongs to TransportersBERT from ActTrans publication.
Taju, Semmy Wellem, Syed Muazzam Ali Shah, and Yu-Yen Ou. “ActTRANS: Functional Classification in Active Transport Proteins Based on Transfer Learning and Contextual Representations.” Computational Biology and Chemistry 93 (August 1, 2021): 107537. https://doi.org/10.1016/j.compbiolchem.2021.107537.
|
{}
|
ghazikhanihamed/TransportersBERT
| null |
[
"transformers",
"pytorch",
"bert",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
text-generation
|
transformers
|
# Connor
|
{"tags": ["conversational"]}
|
ghhostboy/DialoGPT-medium-connorDBH3-1
| null |
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
text-generation
|
transformers
|
# Connor
|
{"tags": ["conversational"]}
|
ghhostboy/DialoGPT-medium-connorDBH3-21
| null |
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
automatic-speech-recognition
|
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# common6
This model is a fine-tuned version of [common6/checkpoint-3500](https://huggingface.co/common6/checkpoint-3500) on the COMMON_VOICE - FA dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3706
- Wer: 0.3421
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-05
- train_batch_size: 32
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 256
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- num_epochs: 200.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:-----:|:---------------:|:------:|
| 1.0344 | 10.0 | 500 | 0.4043 | 0.4511 |
| 0.9651 | 20.0 | 1000 | 0.3793 | 0.4159 |
| 0.9125 | 30.0 | 1500 | 0.3756 | 0.4046 |
| 0.8831 | 40.0 | 2000 | 0.3650 | 0.3876 |
| 0.8399 | 50.0 | 2500 | 0.3605 | 0.3772 |
| 0.819 | 60.0 | 3000 | 0.3622 | 0.3714 |
| 0.8029 | 70.0 | 3500 | 0.3561 | 0.3664 |
| 0.8104 | 80.0 | 4000 | 0.3595 | 0.3660 |
| 0.8118 | 90.0 | 4500 | 0.3460 | 0.3592 |
| 0.7831 | 100.0 | 5000 | 0.3566 | 0.3593 |
| 0.744 | 110.0 | 5500 | 0.3578 | 0.3535 |
| 0.7388 | 120.0 | 6000 | 0.3538 | 0.3520 |
| 0.714 | 130.0 | 6500 | 0.3682 | 0.3506 |
| 0.7291 | 140.0 | 7000 | 0.3625 | 0.3505 |
| 0.697 | 150.0 | 7500 | 0.3619 | 0.3479 |
| 0.6811 | 160.0 | 8000 | 0.3631 | 0.3440 |
| 0.6841 | 170.0 | 8500 | 0.3672 | 0.3460 |
| 0.6616 | 180.0 | 9000 | 0.3677 | 0.3410 |
| 0.6471 | 190.0 | 9500 | 0.3707 | 0.3420 |
| 0.6759 | 200.0 | 10000 | 0.3706 | 0.3421 |
### Framework versions
- Transformers 4.17.0.dev0
- Pytorch 1.10.2
- Datasets 1.18.3.dev0
- Tokenizers 0.10.3
|
{"language": ["fa"], "tags": ["automatic-speech-recognition", "common_voice", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "common6", "results": []}]}
|
ghofrani/common6
| null |
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"common_voice",
"generated_from_trainer",
"fa",
"dataset:common_voice",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
automatic-speech-recognition
|
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# common7
This model is a fine-tuned version of [common7/checkpoint-18500](https://huggingface.co/common7/checkpoint-18500) on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - FA dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3448
- Wer: 0.3478
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-05
- train_batch_size: 32
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- num_epochs: 150.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:------:|:-----:|:---------------:|:------:|
| 2.957 | 3.29 | 500 | 2.9503 | 1.0 |
| 1.7225 | 6.58 | 1000 | 0.8860 | 0.7703 |
| 1.4907 | 9.86 | 1500 | 0.6555 | 0.6673 |
| 1.4177 | 13.16 | 2000 | 0.5784 | 0.6076 |
| 1.3425 | 16.45 | 2500 | 0.5379 | 0.5718 |
| 1.33 | 19.73 | 3000 | 0.4962 | 0.5245 |
| 1.4378 | 23.03 | 3500 | 0.4699 | 0.5098 |
| 1.1894 | 26.31 | 4000 | 0.4527 | 0.4848 |
| 1.1844 | 29.6 | 4500 | 0.4309 | 0.4651 |
| 1.1795 | 32.89 | 5000 | 0.4131 | 0.4524 |
| 1.1471 | 36.18 | 5500 | 0.4052 | 0.4435 |
| 1.1337 | 39.47 | 6000 | 0.3927 | 0.4363 |
| 1.1896 | 42.76 | 6500 | 0.3811 | 0.4254 |
| 1.1847 | 46.05 | 7000 | 0.3855 | 0.4129 |
| 0.9954 | 49.34 | 7500 | 0.3729 | 0.3981 |
| 1.0293 | 52.63 | 8000 | 0.3637 | 0.4014 |
| 1.0224 | 55.92 | 8500 | 0.3578 | 0.3885 |
| 1.012 | 59.21 | 9000 | 0.3629 | 0.3930 |
| 1.0772 | 62.5 | 9500 | 0.3635 | 0.3906 |
| 1.0344 | 65.79 | 10000 | 0.3469 | 0.3771 |
| 0.9457 | 69.08 | 10500 | 0.3435 | 0.3735 |
| 0.9307 | 72.37 | 11000 | 0.3519 | 0.3762 |
| 0.9523 | 75.65 | 11500 | 0.3443 | 0.3666 |
| 0.9523 | 78.94 | 12000 | 0.3502 | 0.3757 |
| 0.9475 | 82.24 | 12500 | 0.3509 | 0.3643 |
| 0.9971 | 85.52 | 13000 | 0.3502 | 0.3626 |
| 0.9058 | 88.81 | 13500 | 0.3472 | 0.3605 |
| 0.8922 | 92.1 | 14000 | 0.3530 | 0.3618 |
| 0.9 | 95.39 | 14500 | 0.3500 | 0.3574 |
| 0.9051 | 98.68 | 15000 | 0.3456 | 0.3535 |
| 0.9304 | 101.97 | 15500 | 0.3438 | 0.3578 |
| 0.9433 | 105.26 | 16000 | 0.3396 | 0.3530 |
| 0.8988 | 108.55 | 16500 | 0.3436 | 0.3539 |
| 0.8789 | 111.84 | 17000 | 0.3426 | 0.3516 |
| 0.8667 | 115.13 | 17500 | 0.3438 | 0.3506 |
| 0.8895 | 118.42 | 18000 | 0.3434 | 0.3503 |
| 0.8888 | 121.71 | 18500 | 0.3425 | 0.3494 |
| 0.9453 | 125.0 | 19000 | 0.3415 | 0.3480 |
| 0.9267 | 128.29 | 19500 | 0.3477 | 0.3503 |
| 0.8315 | 131.58 | 20000 | 0.3476 | 0.3505 |
| 0.8542 | 134.86 | 20500 | 0.3475 | 0.3506 |
| 0.8478 | 138.16 | 21000 | 0.3430 | 0.3481 |
| 0.8643 | 141.45 | 21500 | 0.3451 | 0.3485 |
| 0.8705 | 144.73 | 22000 | 0.3444 | 0.3474 |
| 0.9869 | 148.03 | 22500 | 0.3441 | 0.3493 |
### Framework versions
- Transformers 4.17.0.dev0
- Pytorch 1.10.2
- Datasets 1.18.3.dev0
- Tokenizers 0.10.3
|
{"language": ["fa"], "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_7_0", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "common7", "results": []}]}
|
ghofrani/common7
| null |
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"mozilla-foundation/common_voice_7_0",
"generated_from_trainer",
"fa",
"dataset:common_voice",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
automatic-speech-recognition
|
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# common8
This model is a fine-tuned version of [wghts/checkpoint-20000](https://huggingface.co/wghts/checkpoint-20000) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - FA dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3174
- Wer: 0.3022
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-06
- train_batch_size: 32
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 6
- total_train_batch_size: 192
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- num_epochs: 250.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:------:|:-----:|:---------------:|:------:|
| 3.5847 | 1.93 | 500 | 3.5104 | 1.0 |
| 2.7858 | 3.86 | 1000 | 2.9601 | 1.0001 |
| 1.6827 | 5.79 | 1500 | 0.7853 | 0.7030 |
| 1.4656 | 7.72 | 2000 | 0.6076 | 0.6014 |
| 1.3693 | 9.65 | 2500 | 0.5114 | 0.5307 |
| 1.379 | 11.58 | 3000 | 0.4666 | 0.4940 |
| 1.2832 | 13.51 | 3500 | 0.4257 | 0.4593 |
| 1.1931 | 15.44 | 4000 | 0.4039 | 0.4427 |
| 1.2911 | 17.37 | 4500 | 0.3956 | 0.4295 |
| 1.1577 | 19.3 | 5000 | 0.3705 | 0.4114 |
| 1.1135 | 21.24 | 5500 | 0.3740 | 0.4010 |
| 1.19 | 23.17 | 6000 | 0.3611 | 0.3935 |
| 1.1008 | 25.1 | 6500 | 0.3503 | 0.3880 |
| 1.0805 | 27.03 | 7000 | 0.3427 | 0.3781 |
| 1.1556 | 28.96 | 7500 | 0.3442 | 0.3727 |
| 1.0596 | 30.89 | 8000 | 0.3398 | 0.3646 |
| 1.0219 | 32.82 | 8500 | 0.3312 | 0.3660 |
| 1.1042 | 34.75 | 9000 | 0.3287 | 0.3612 |
| 1.0273 | 36.68 | 9500 | 0.3236 | 0.3556 |
| 1.0383 | 38.61 | 10000 | 0.3217 | 0.3558 |
| 1.0498 | 40.54 | 10500 | 0.3205 | 0.3520 |
| 0.9969 | 42.47 | 11000 | 0.3125 | 0.3504 |
| 1.0658 | 44.4 | 11500 | 0.3120 | 0.3493 |
| 0.992 | 46.33 | 12000 | 0.3137 | 0.3476 |
| 0.9737 | 48.26 | 12500 | 0.3085 | 0.3413 |
| 1.0817 | 50.19 | 13000 | 0.3091 | 0.3418 |
| 0.9414 | 52.12 | 13500 | 0.3072 | 0.3344 |
| 0.9295 | 54.05 | 14000 | 0.3039 | 0.3322 |
| 1.0248 | 55.98 | 14500 | 0.2991 | 0.3325 |
| 0.9474 | 57.91 | 15000 | 0.3032 | 0.3348 |
| 0.928 | 59.85 | 15500 | 0.2999 | 0.3285 |
| 1.0321 | 61.78 | 16000 | 0.2982 | 0.3253 |
| 0.9255 | 63.71 | 16500 | 0.2970 | 0.3231 |
| 0.8928 | 65.64 | 17000 | 0.2993 | 0.3250 |
| 1.008 | 67.57 | 17500 | 0.2985 | 0.3222 |
| 0.9371 | 69.5 | 18000 | 0.2968 | 0.3216 |
| 0.9077 | 71.43 | 18500 | 0.3011 | 0.3299 |
| 1.0044 | 73.36 | 19000 | 0.3053 | 0.3306 |
| 0.9625 | 75.29 | 19500 | 0.3159 | 0.3295 |
| 0.9816 | 77.22 | 20000 | 0.3080 | 0.3304 |
| 0.9587 | 119.19 | 20500 | 0.3088 | 0.3284 |
| 0.9178 | 122.09 | 21000 | 0.3132 | 0.3320 |
| 1.0282 | 125.0 | 21500 | 0.3099 | 0.3266 |
| 0.9337 | 127.9 | 22000 | 0.3110 | 0.3317 |
| 0.8822 | 130.81 | 22500 | 0.3037 | 0.3247 |
| 0.9644 | 133.72 | 23000 | 0.3037 | 0.3238 |
| 0.9214 | 136.62 | 23500 | 0.3040 | 0.3234 |
| 0.9167 | 139.53 | 24000 | 0.3079 | 0.3203 |
| 0.9047 | 142.44 | 24500 | 0.3018 | 0.3177 |
| 0.8909 | 145.35 | 25000 | 0.3053 | 0.3181 |
| 0.9646 | 148.25 | 25500 | 0.3095 | 0.3229 |
| 0.8802 | 151.16 | 26000 | 0.3111 | 0.3192 |
| 0.8411 | 154.07 | 26500 | 0.3068 | 0.3123 |
| 0.9235 | 156.97 | 27000 | 0.3090 | 0.3177 |
| 0.8943 | 159.88 | 27500 | 0.3115 | 0.3179 |
| 0.8854 | 162.79 | 28000 | 0.3052 | 0.3157 |
| 0.8734 | 165.69 | 28500 | 0.3077 | 0.3124 |
| 0.8515 | 168.6 | 29000 | 0.3117 | 0.3128 |
| 0.912 | 171.51 | 29500 | 0.3039 | 0.3121 |
| 0.8669 | 174.42 | 30000 | 0.3120 | 0.3123 |
| 0.823 | 177.32 | 30500 | 0.3148 | 0.3118 |
| 0.9129 | 180.23 | 31000 | 0.3179 | 0.3101 |
| 0.8255 | 183.14 | 31500 | 0.3164 | 0.3114 |
| 0.8948 | 186.05 | 32000 | 0.3128 | 0.3101 |
| 0.8397 | 188.95 | 32500 | 0.3143 | 0.3068 |
| 0.8341 | 191.86 | 33000 | 0.3127 | 0.3136 |
| 0.873 | 194.76 | 33500 | 0.3149 | 0.3124 |
| 0.8232 | 197.67 | 34000 | 0.3166 | 0.3086 |
| 0.8002 | 200.58 | 34500 | 0.3149 | 0.3061 |
| 0.8621 | 203.49 | 35000 | 0.3160 | 0.3093 |
| 0.8123 | 206.39 | 35500 | 0.3141 | 0.3063 |
| 0.7995 | 209.3 | 36000 | 0.3174 | 0.3075 |
| 0.8271 | 212.21 | 36500 | 0.3173 | 0.3043 |
| 0.8059 | 215.12 | 37000 | 0.3176 | 0.3079 |
| 0.8835 | 218.02 | 37500 | 0.3169 | 0.3062 |
| 0.8027 | 220.93 | 38000 | 0.3203 | 0.3098 |
| 0.775 | 223.83 | 38500 | 0.3159 | 0.3068 |
| 0.8487 | 226.74 | 39000 | 0.3161 | 0.3072 |
| 0.7929 | 229.65 | 39500 | 0.3143 | 0.3037 |
| 0.7653 | 232.56 | 40000 | 0.3160 | 0.3048 |
| 0.8211 | 235.46 | 40500 | 0.3173 | 0.3031 |
| 0.7761 | 238.37 | 41000 | 0.3176 | 0.3025 |
| 0.7761 | 241.28 | 41500 | 0.3179 | 0.3027 |
| 0.7903 | 244.19 | 42000 | 0.3181 | 0.3016 |
| 0.7807 | 247.09 | 42500 | 0.3170 | 0.3027 |
| 0.8406 | 250.0 | 43000 | 0.3174 | 0.3022 |
### Framework versions
- Transformers 4.17.0.dev0
- Pytorch 1.10.2
- Datasets 1.18.3.dev0
- Tokenizers 0.10.3
|
{"language": ["fa"], "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "common8", "results": []}]}
|
ghofrani/xls-r-1b-fa-cv8
| null |
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"mozilla-foundation/common_voice_8_0",
"generated_from_trainer",
"fa",
"dataset:common_voice",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
null | null |
{}
|
ghofrani/distilbert-base-uncased-finetuned-ner
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
automatic-speech-recognition
|
transformers
|
{}
|
ghofrani/wghts
| null |
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
null | null |
{}
|
ghofrani/xls-r-1b-fa
| null |
[
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
|
automatic-speech-recognition
|
transformers
|
{}
|
ghofrani/xls-r-300m-fa
| null |
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"endpoints_compatible",
"region:us"
] | null |
2022-03-02T23:29:05+00:00
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.