njprogrammer commited on
Commit
fbd264a
·
1 Parent(s): 2387882

Add SetFit model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - setfit
5
+ - sentence-transformers
6
+ - text-classification
7
+ pipeline_tag: text-classification
8
+ ---
9
+
10
+ # njprogrammer/goemotion-BERT
11
+
12
+ This is a [SetFit model](https://github.com/huggingface/setfit) that can be used for text classification. The model has been trained using an efficient few-shot learning technique that involves:
13
+
14
+ 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
15
+ 2. Training a classification head with features from the fine-tuned Sentence Transformer.
16
+
17
+ ## Usage
18
+
19
+ To use this model for inference, first install the SetFit library:
20
+
21
+ ```bash
22
+ python -m pip install setfit
23
+ ```
24
+
25
+ You can then run inference as follows:
26
+
27
+ ```python
28
+ from setfit import SetFitModel
29
+
30
+ # Download from Hub and run inference
31
+ model = SetFitModel.from_pretrained("njprogrammer/goemotion-BERT")
32
+ # Run inference
33
+ preds = model(["i loved the spiderman movie!", "pineapple on pizza is the worst 🤮"])
34
+ ```
35
+
36
+ ## BibTeX entry and citation info
37
+
38
+ ```bibtex
39
+ @article{https://doi.org/10.48550/arxiv.2209.11055,
40
+ doi = {10.48550/ARXIV.2209.11055},
41
+ url = {https://arxiv.org/abs/2209.11055},
42
+ author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
43
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
44
+ title = {Efficient Few-Shot Learning Without Prompts},
45
+ publisher = {arXiv},
46
+ year = {2022},
47
+ copyright = {Creative Commons Attribution 4.0 International}
48
+ }
49
+ ```
config.json ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "C:\\Users\\njpark/.cache\\torch\\sentence_transformers\\monologg_bert-base-cased-goemotions-original",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "finetuning_task": "goemotions",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "admiration",
14
+ "1": "amusement",
15
+ "2": "anger",
16
+ "3": "annoyance",
17
+ "4": "approval",
18
+ "5": "caring",
19
+ "6": "confusion",
20
+ "7": "curiosity",
21
+ "8": "desire",
22
+ "9": "disappointment",
23
+ "10": "disapproval",
24
+ "11": "disgust",
25
+ "12": "embarrassment",
26
+ "13": "excitement",
27
+ "14": "fear",
28
+ "15": "gratitude",
29
+ "16": "grief",
30
+ "17": "joy",
31
+ "18": "love",
32
+ "19": "nervousness",
33
+ "20": "optimism",
34
+ "21": "pride",
35
+ "22": "realization",
36
+ "23": "relief",
37
+ "24": "remorse",
38
+ "25": "sadness",
39
+ "26": "surprise",
40
+ "27": "neutral"
41
+ },
42
+ "initializer_range": 0.02,
43
+ "intermediate_size": 3072,
44
+ "label2id": {
45
+ "admiration": 0,
46
+ "amusement": 1,
47
+ "anger": 2,
48
+ "annoyance": 3,
49
+ "approval": 4,
50
+ "caring": 5,
51
+ "confusion": 6,
52
+ "curiosity": 7,
53
+ "desire": 8,
54
+ "disappointment": 9,
55
+ "disapproval": 10,
56
+ "disgust": 11,
57
+ "embarrassment": 12,
58
+ "excitement": 13,
59
+ "fear": 14,
60
+ "gratitude": 15,
61
+ "grief": 16,
62
+ "joy": 17,
63
+ "love": 18,
64
+ "nervousness": 19,
65
+ "neutral": 27,
66
+ "optimism": 20,
67
+ "pride": 21,
68
+ "realization": 22,
69
+ "relief": 23,
70
+ "remorse": 24,
71
+ "sadness": 25,
72
+ "surprise": 26
73
+ },
74
+ "layer_norm_eps": 1e-12,
75
+ "max_position_embeddings": 512,
76
+ "model_type": "bert",
77
+ "num_attention_heads": 12,
78
+ "num_hidden_layers": 12,
79
+ "pad_token_id": 0,
80
+ "position_embedding_type": "absolute",
81
+ "torch_dtype": "float32",
82
+ "transformers_version": "4.30.2",
83
+ "type_vocab_size": 2,
84
+ "use_cache": true,
85
+ "vocab_size": 28996
86
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.2",
4
+ "transformers": "4.30.2",
5
+ "pytorch": "1.8.1+cu111"
6
+ }
7
+ }
model_head.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33fd01870205b60c87d2af80a1c75f041268c4fa7045f5b9ff266fd4a26eec9b
3
+ size 87446
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0027df0550466df7abb3961d32a1c11191a7cdb257f4acbe1c323ba548793d8
3
+ size 433315311
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "[NAME]",
4
+ "[RELIGION]"
5
+ ],
6
+ "cls_token": "[CLS]",
7
+ "mask_token": "[MASK]",
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "unk_token": "[UNK]"
11
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "[NAME]",
4
+ "[RELIGION]"
5
+ ],
6
+ "clean_up_tokenization_spaces": true,
7
+ "cls_token": "[CLS]",
8
+ "do_basic_tokenize": true,
9
+ "do_lower_case": false,
10
+ "mask_token": "[MASK]",
11
+ "model_max_length": 512,
12
+ "never_split": null,
13
+ "pad_token": "[PAD]",
14
+ "sep_token": "[SEP]",
15
+ "strip_accents": null,
16
+ "tokenize_chinese_chars": true,
17
+ "tokenizer_class": "BertTokenizer",
18
+ "unk_token": "[UNK]"
19
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff