Charul1223
commited on
Commit
•
1b12507
1
Parent(s):
3015c2b
Upload folder using huggingface_hub
Browse files- README.md +1 -1
- config.json +1 -2
- config_sentence_transformers.json +3 -3
- desktop.ini +2 -0
- eval/similarity_evaluation_sts-dev_results.csv +10 -10
- pytorch_model.bin +1 -1
- sentence_bert_config.json +1 -1
- similarity_evaluation_sts-test_results.csv +1 -1
- tokenizer.json +1 -1
- tokenizer_config.json +8 -57
README.md
CHANGED
@@ -116,7 +116,7 @@ Parameters of the fit()-Method:
|
|
116 |
## Full Model Architecture
|
117 |
```
|
118 |
SentenceTransformer(
|
119 |
-
(0): Transformer({'max_seq_length':
|
120 |
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
|
121 |
)
|
122 |
```
|
|
|
116 |
## Full Model Architecture
|
117 |
```
|
118 |
SentenceTransformer(
|
119 |
+
(0): Transformer({'max_seq_length': 514, 'do_lower_case': False}) with Transformer model: RobertaModel
|
120 |
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
|
121 |
)
|
122 |
```
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "C:\\Users\\aniru/.cache\\torch\\sentence_transformers\\
|
3 |
"architectures": [
|
4 |
"RobertaModel"
|
5 |
],
|
@@ -7,7 +7,6 @@
|
|
7 |
"bos_token_id": 0,
|
8 |
"classifier_dropout": null,
|
9 |
"eos_token_id": 2,
|
10 |
-
"gradient_checkpointing": false,
|
11 |
"hidden_act": "gelu",
|
12 |
"hidden_dropout_prob": 0.1,
|
13 |
"hidden_size": 768,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "C:\\Users\\aniru/.cache\\torch\\sentence_transformers\\distilroberta-base",
|
3 |
"architectures": [
|
4 |
"RobertaModel"
|
5 |
],
|
|
|
7 |
"bos_token_id": 0,
|
8 |
"classifier_dropout": null,
|
9 |
"eos_token_id": 2,
|
|
|
10 |
"hidden_act": "gelu",
|
11 |
"hidden_dropout_prob": 0.1,
|
12 |
"hidden_size": 768,
|
config_sentence_transformers.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
"__version__": {
|
3 |
-
"sentence_transformers": "2.
|
4 |
-
"transformers": "4.
|
5 |
-
"pytorch": "
|
6 |
}
|
7 |
}
|
|
|
1 |
{
|
2 |
"__version__": {
|
3 |
+
"sentence_transformers": "2.2.2",
|
4 |
+
"transformers": "4.28.1",
|
5 |
+
"pytorch": "2.0.0+cpu"
|
6 |
}
|
7 |
}
|
desktop.ini
ADDED
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
1 |
+
[.ShellClassInfo]
|
2 |
+
LocalizedResourceName=distilroberta-base-on-sts
|
eval/similarity_evaluation_sts-dev_results.csv
CHANGED
@@ -1,11 +1,11 @@
|
|
1 |
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
-
0,-1,0.
|
3 |
-
1,-1,0.
|
4 |
-
2,-1,0.
|
5 |
-
3,-1,0.
|
6 |
-
4,-1,0.
|
7 |
-
5,-1,0.
|
8 |
-
6,-1,0.
|
9 |
-
7,-1,0.
|
10 |
-
8,-1,0.
|
11 |
-
9,-1,0.
|
|
|
1 |
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
+
0,-1,0.8615914886740045,0.8626620081436468,0.8235818250476207,0.82746470093546,0.8229972055905649,0.8265173550859571,0.761862607356981,0.7749299849653893
|
3 |
+
1,-1,0.864894129665433,0.8669473331650523,0.8347267078962851,0.8361935733609941,0.8338377428969682,0.8351395834561777,0.7630282084913088,0.7710472189406553
|
4 |
+
2,-1,0.8737236005772511,0.8738676439906548,0.8408239571230334,0.8419128835041697,0.8395596941329996,0.8404314963608522,0.7893179975686605,0.7959694656771303
|
5 |
+
3,-1,0.8706590059185022,0.8710030477077165,0.8441278543599693,0.8463152388554493,0.8428579626641867,0.8445717607839294,0.7979674342522111,0.8016647311733303
|
6 |
+
4,-1,0.87314625289717,0.874126323884475,0.8495568739799131,0.8522216984475744,0.8479432342994989,0.850154565348426,0.8136978200296894,0.8156756785648545
|
7 |
+
5,-1,0.8747856133586903,0.8763425371576796,0.8591660087870885,0.8624199136705251,0.8573639740695901,0.8601465099663936,0.8190495703815498,0.8197047631959027
|
8 |
+
6,-1,0.8732198013807473,0.8749570326428093,0.8556933870699567,0.8589934565534257,0.8537552911680004,0.856568569774799,0.8170046779656095,0.8175549307514262
|
9 |
+
7,-1,0.875091091100427,0.8761276829168797,0.8582136631996886,0.8614063880600481,0.8562007352737929,0.8588640108606868,0.8230568435995905,0.8237395773682713
|
10 |
+
8,-1,0.8755760805123022,0.8771158152713984,0.8619479319714531,0.865068068522956,0.8600903727609397,0.8628122075355573,0.8255042434523162,0.8256043277654755
|
11 |
+
9,-1,0.8752847426769929,0.8769273204793943,0.8622451430139331,0.8655589140574234,0.8604159773277158,0.8633625210825944,0.8256593408675647,0.8258070734569941
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 328510893
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:18972e73f4dd135ec663a47ecc2e8936a7497cadcc2a6e430152458941432a35
|
3 |
size 328510893
|
sentence_bert_config.json
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
{
|
2 |
-
"max_seq_length":
|
3 |
"do_lower_case": false
|
4 |
}
|
|
|
1 |
{
|
2 |
+
"max_seq_length": 514,
|
3 |
"do_lower_case": false
|
4 |
}
|
similarity_evaluation_sts-test_results.csv
CHANGED
@@ -1,2 +1,2 @@
|
|
1 |
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
-
-1,-1,0.
|
|
|
1 |
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
+
-1,-1,0.847696034868068,0.8468411019593707,0.8268577771952719,0.8241921384376979,0.8254449090181314,0.8229385375277086,0.797009890800946,0.7872730025810322
|
tokenizer.json
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
"version": "1.0",
|
3 |
"truncation": {
|
4 |
"direction": "Right",
|
5 |
-
"max_length":
|
6 |
"strategy": "LongestFirst",
|
7 |
"stride": 0
|
8 |
},
|
|
|
2 |
"version": "1.0",
|
3 |
"truncation": {
|
4 |
"direction": "Right",
|
5 |
+
"max_length": 514,
|
6 |
"strategy": "LongestFirst",
|
7 |
"stride": 0
|
8 |
},
|
tokenizer_config.json
CHANGED
@@ -1,64 +1,15 @@
|
|
1 |
{
|
2 |
"add_prefix_space": false,
|
3 |
-
"bos_token":
|
4 |
-
"__type": "AddedToken",
|
5 |
-
"content": "<s>",
|
6 |
-
"lstrip": false,
|
7 |
-
"normalized": true,
|
8 |
-
"rstrip": false,
|
9 |
-
"single_word": false
|
10 |
-
},
|
11 |
"clean_up_tokenization_spaces": true,
|
12 |
-
"cls_token":
|
13 |
-
|
14 |
-
"content": "<s>",
|
15 |
-
"lstrip": false,
|
16 |
-
"normalized": true,
|
17 |
-
"rstrip": false,
|
18 |
-
"single_word": false
|
19 |
-
},
|
20 |
-
"eos_token": {
|
21 |
-
"__type": "AddedToken",
|
22 |
-
"content": "</s>",
|
23 |
-
"lstrip": false,
|
24 |
-
"normalized": true,
|
25 |
-
"rstrip": false,
|
26 |
-
"single_word": false
|
27 |
-
},
|
28 |
"errors": "replace",
|
29 |
-
"mask_token":
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
"normalized": true,
|
34 |
-
"rstrip": false,
|
35 |
-
"single_word": false
|
36 |
-
},
|
37 |
-
"model_max_length": 512,
|
38 |
-
"pad_token": {
|
39 |
-
"__type": "AddedToken",
|
40 |
-
"content": "<pad>",
|
41 |
-
"lstrip": false,
|
42 |
-
"normalized": true,
|
43 |
-
"rstrip": false,
|
44 |
-
"single_word": false
|
45 |
-
},
|
46 |
-
"sep_token": {
|
47 |
-
"__type": "AddedToken",
|
48 |
-
"content": "</s>",
|
49 |
-
"lstrip": false,
|
50 |
-
"normalized": true,
|
51 |
-
"rstrip": false,
|
52 |
-
"single_word": false
|
53 |
-
},
|
54 |
"tokenizer_class": "RobertaTokenizer",
|
55 |
"trim_offsets": true,
|
56 |
-
"unk_token":
|
57 |
-
"__type": "AddedToken",
|
58 |
-
"content": "<unk>",
|
59 |
-
"lstrip": false,
|
60 |
-
"normalized": true,
|
61 |
-
"rstrip": false,
|
62 |
-
"single_word": false
|
63 |
-
}
|
64 |
}
|
|
|
1 |
{
|
2 |
"add_prefix_space": false,
|
3 |
+
"bos_token": "<s>",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
"clean_up_tokenization_spaces": true,
|
5 |
+
"cls_token": "<s>",
|
6 |
+
"eos_token": "</s>",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
"errors": "replace",
|
8 |
+
"mask_token": "<mask>",
|
9 |
+
"model_max_length": 1000000000000000019884624838656,
|
10 |
+
"pad_token": "<pad>",
|
11 |
+
"sep_token": "</s>",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
"tokenizer_class": "RobertaTokenizer",
|
13 |
"trim_offsets": true,
|
14 |
+
"unk_token": "<unk>"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
}
|