Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- 1_Pooling/config.json +7 -0
- README.md +125 -1
- config.json +28 -0
- config_sentence_transformers.json +7 -0
- eval/mse_evaluation_Tatoeba-eng-rus-dev.tsv.gz_results.csv +185 -0
- eval/mse_evaluation_Tatoeba-eng-ukr-dev.tsv.gz_results.csv +185 -0
- eval/translation_evaluation_Tatoeba-eng-rus-dev.tsv.gz_results.csv +185 -0
- eval/translation_evaluation_Tatoeba-eng-ukr-dev.tsv.gz_results.csv +185 -0
- model.safetensors +3 -0
- modules.json +14 -0
- sentence_bert_config.json +4 -0
- sentencepiece.bpe.model +3 -0
- special_tokens_map.json +15 -0
- tokenizer.json +3 -0
- tokenizer_config.json +54 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
1_Pooling/config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"word_embedding_dimension": 768,
|
3 |
+
"pooling_mode_cls_token": false,
|
4 |
+
"pooling_mode_mean_tokens": true,
|
5 |
+
"pooling_mode_max_tokens": false,
|
6 |
+
"pooling_mode_mean_sqrt_len_tokens": false
|
7 |
+
}
|
README.md
CHANGED
@@ -1,3 +1,127 @@
|
|
1 |
---
|
2 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
pipeline_tag: sentence-similarity
|
3 |
+
tags:
|
4 |
+
- sentence-transformers
|
5 |
+
- feature-extraction
|
6 |
+
- sentence-similarity
|
7 |
+
- transformers
|
8 |
+
|
9 |
---
|
10 |
+
|
11 |
+
# {MODEL_NAME}
|
12 |
+
|
13 |
+
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
|
14 |
+
|
15 |
+
<!--- Describe your model here -->
|
16 |
+
|
17 |
+
## Usage (Sentence-Transformers)
|
18 |
+
|
19 |
+
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
|
20 |
+
|
21 |
+
```
|
22 |
+
pip install -U sentence-transformers
|
23 |
+
```
|
24 |
+
|
25 |
+
Then you can use the model like this:
|
26 |
+
|
27 |
+
```python
|
28 |
+
from sentence_transformers import SentenceTransformer
|
29 |
+
sentences = ["This is an example sentence", "Each sentence is converted"]
|
30 |
+
|
31 |
+
model = SentenceTransformer('{MODEL_NAME}')
|
32 |
+
embeddings = model.encode(sentences)
|
33 |
+
print(embeddings)
|
34 |
+
```
|
35 |
+
|
36 |
+
|
37 |
+
|
38 |
+
## Usage (HuggingFace Transformers)
|
39 |
+
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
|
40 |
+
|
41 |
+
```python
|
42 |
+
from transformers import AutoTokenizer, AutoModel
|
43 |
+
import torch
|
44 |
+
|
45 |
+
|
46 |
+
#Mean Pooling - Take attention mask into account for correct averaging
|
47 |
+
def mean_pooling(model_output, attention_mask):
|
48 |
+
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
|
49 |
+
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
|
50 |
+
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
|
51 |
+
|
52 |
+
|
53 |
+
# Sentences we want sentence embeddings for
|
54 |
+
sentences = ['This is an example sentence', 'Each sentence is converted']
|
55 |
+
|
56 |
+
# Load model from HuggingFace Hub
|
57 |
+
tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
|
58 |
+
model = AutoModel.from_pretrained('{MODEL_NAME}')
|
59 |
+
|
60 |
+
# Tokenize sentences
|
61 |
+
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
|
62 |
+
|
63 |
+
# Compute token embeddings
|
64 |
+
with torch.no_grad():
|
65 |
+
model_output = model(**encoded_input)
|
66 |
+
|
67 |
+
# Perform pooling. In this case, mean pooling.
|
68 |
+
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
|
69 |
+
|
70 |
+
print("Sentence embeddings:")
|
71 |
+
print(sentence_embeddings)
|
72 |
+
```
|
73 |
+
|
74 |
+
|
75 |
+
|
76 |
+
## Evaluation Results
|
77 |
+
|
78 |
+
<!--- Describe how your model was evaluated -->
|
79 |
+
|
80 |
+
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
|
81 |
+
|
82 |
+
|
83 |
+
## Training
|
84 |
+
The model was trained with the parameters:
|
85 |
+
|
86 |
+
**DataLoader**:
|
87 |
+
|
88 |
+
`torch.utils.data.dataloader.DataLoader` of length 22369 with parameters:
|
89 |
+
```
|
90 |
+
{'batch_size': 64, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
|
91 |
+
```
|
92 |
+
|
93 |
+
**Loss**:
|
94 |
+
|
95 |
+
`sentence_transformers.losses.MSELoss.MSELoss`
|
96 |
+
|
97 |
+
Parameters of the fit()-Method:
|
98 |
+
```
|
99 |
+
{
|
100 |
+
"epochs": 8,
|
101 |
+
"evaluation_steps": 1000,
|
102 |
+
"evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
|
103 |
+
"max_grad_norm": 1,
|
104 |
+
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
|
105 |
+
"optimizer_params": {
|
106 |
+
"eps": 1e-06,
|
107 |
+
"lr": 2e-05
|
108 |
+
},
|
109 |
+
"scheduler": "WarmupLinear",
|
110 |
+
"steps_per_epoch": null,
|
111 |
+
"warmup_steps": 1000,
|
112 |
+
"weight_decay": 0.01
|
113 |
+
}
|
114 |
+
```
|
115 |
+
|
116 |
+
|
117 |
+
## Full Model Architecture
|
118 |
+
```
|
119 |
+
SentenceTransformer(
|
120 |
+
(0): Transformer({'max_seq_length': 384, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
|
121 |
+
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
|
122 |
+
)
|
123 |
+
```
|
124 |
+
|
125 |
+
## Citing & Authors
|
126 |
+
|
127 |
+
<!--- Describe where people can find more information -->
|
config.json
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "xlm-roberta-base",
|
3 |
+
"architectures": [
|
4 |
+
"XLMRobertaModel"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"classifier_dropout": null,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "gelu",
|
11 |
+
"hidden_dropout_prob": 0.1,
|
12 |
+
"hidden_size": 768,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 514,
|
17 |
+
"model_type": "xlm-roberta",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 12,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 1,
|
22 |
+
"position_embedding_type": "absolute",
|
23 |
+
"torch_dtype": "float32",
|
24 |
+
"transformers_version": "4.35.2",
|
25 |
+
"type_vocab_size": 1,
|
26 |
+
"use_cache": true,
|
27 |
+
"vocab_size": 250002
|
28 |
+
}
|
config_sentence_transformers.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"__version__": {
|
3 |
+
"sentence_transformers": "2.2.2",
|
4 |
+
"transformers": "4.35.2",
|
5 |
+
"pytorch": "2.1.0+cu121"
|
6 |
+
}
|
7 |
+
}
|
eval/mse_evaluation_Tatoeba-eng-rus-dev.tsv.gz_results.csv
ADDED
@@ -0,0 +1,185 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,MSE
|
2 |
+
0,1000,0.848949421197176
|
3 |
+
0,2000,0.4606836475431919
|
4 |
+
0,3000,0.8414469659328461
|
5 |
+
0,4000,42.11623966693878
|
6 |
+
0,5000,37.60883808135986
|
7 |
+
0,6000,28.293192386627197
|
8 |
+
0,7000,15.338416397571564
|
9 |
+
0,8000,18.78209561109543
|
10 |
+
0,9000,14.22184556722641
|
11 |
+
0,10000,21.08801007270813
|
12 |
+
0,11000,16.04025810956955
|
13 |
+
0,12000,0.15358192613348365
|
14 |
+
0,13000,0.15338454395532608
|
15 |
+
0,14000,0.13217859668657184
|
16 |
+
0,15000,8.262499421834946
|
17 |
+
0,16000,26.38794183731079
|
18 |
+
0,17000,19.209598004817963
|
19 |
+
0,18000,8.74292254447937
|
20 |
+
0,19000,5.29480054974556
|
21 |
+
0,20000,4.526643082499504
|
22 |
+
0,21000,5.248592421412468
|
23 |
+
0,22000,3.183121234178543
|
24 |
+
0,-1,11.969202011823654
|
25 |
+
1,1000,6.24447837471962
|
26 |
+
1,2000,6.333371251821518
|
27 |
+
1,3000,12.141624838113785
|
28 |
+
1,4000,3.921690955758095
|
29 |
+
1,5000,14.211288094520569
|
30 |
+
1,6000,11.02134883403778
|
31 |
+
1,7000,13.251310586929321
|
32 |
+
1,8000,3.8379158824682236
|
33 |
+
1,9000,3.659619018435478
|
34 |
+
1,10000,3.491947054862976
|
35 |
+
1,11000,0.12413397198542953
|
36 |
+
1,12000,0.12565606739372015
|
37 |
+
1,13000,0.12469175271689892
|
38 |
+
1,14000,0.12597381137311459
|
39 |
+
1,15000,0.1254286034964025
|
40 |
+
1,16000,0.12833582004532218
|
41 |
+
1,17000,0.1254844362847507
|
42 |
+
1,18000,0.12527049984782934
|
43 |
+
1,19000,0.12602504575625062
|
44 |
+
1,20000,0.12823607539758086
|
45 |
+
1,21000,0.12702911626547575
|
46 |
+
1,22000,0.12795787770301104
|
47 |
+
1,-1,0.1278354087844491
|
48 |
+
2,1000,0.13138328213244677
|
49 |
+
2,2000,0.12836138484999537
|
50 |
+
2,3000,0.1274282345548272
|
51 |
+
2,4000,0.1273218891583383
|
52 |
+
2,5000,0.12945140479132533
|
53 |
+
2,6000,0.12839996488764882
|
54 |
+
2,7000,0.12647401308640838
|
55 |
+
2,8000,0.12723514810204506
|
56 |
+
2,9000,0.13070841087028384
|
57 |
+
2,10000,0.12705009430646896
|
58 |
+
2,11000,0.12638139305636287
|
59 |
+
2,12000,0.12608106480911374
|
60 |
+
2,13000,0.12752475449815392
|
61 |
+
2,14000,0.1281437580473721
|
62 |
+
2,15000,0.12951804092153907
|
63 |
+
2,16000,0.12963052140548825
|
64 |
+
2,17000,0.12746391585096717
|
65 |
+
2,18000,0.12532207183539867
|
66 |
+
2,19000,0.1257385010831058
|
67 |
+
2,20000,0.12523052282631397
|
68 |
+
2,21000,0.12620625784620643
|
69 |
+
2,22000,0.12513790279626846
|
70 |
+
2,-1,0.12589959660544991
|
71 |
+
3,1000,0.12622427893802524
|
72 |
+
3,2000,0.12535502901300788
|
73 |
+
3,3000,0.1261907978914678
|
74 |
+
3,4000,0.12519396841526031
|
75 |
+
3,5000,0.12532700784504414
|
76 |
+
3,6000,0.12554720742627978
|
77 |
+
3,7000,0.12509647058323026
|
78 |
+
3,8000,0.12553180567920208
|
79 |
+
3,9000,0.12640993809327483
|
80 |
+
3,10000,0.12675226898863912
|
81 |
+
3,11000,0.12640354689210653
|
82 |
+
3,12000,0.12812383938580751
|
83 |
+
3,13000,0.12684185057878494
|
84 |
+
3,14000,0.12931845849379897
|
85 |
+
3,15000,0.12905217008665204
|
86 |
+
3,16000,0.12640327913686633
|
87 |
+
3,17000,0.12504806509241462
|
88 |
+
3,18000,0.12564064236357808
|
89 |
+
3,19000,0.12659903150051832
|
90 |
+
3,20000,0.12494247639551759
|
91 |
+
3,21000,0.12556747533380985
|
92 |
+
3,22000,0.12491807574406266
|
93 |
+
3,-1,0.12564186472445726
|
94 |
+
4,1000,0.12547316728159785
|
95 |
+
4,2000,0.1254525501281023
|
96 |
+
4,3000,0.12541613541543484
|
97 |
+
4,4000,0.12554129352793097
|
98 |
+
4,5000,0.12682921951636672
|
99 |
+
4,6000,0.1258071744814515
|
100 |
+
4,7000,0.1253918744623661
|
101 |
+
4,8000,0.12559785973280668
|
102 |
+
4,9000,0.12527299113571644
|
103 |
+
4,10000,0.12599037727341056
|
104 |
+
4,11000,0.12512532994151115
|
105 |
+
4,12000,0.125212618149817
|
106 |
+
4,13000,0.12555598514154553
|
107 |
+
4,14000,0.1251509878784418
|
108 |
+
4,15000,0.12564953649416566
|
109 |
+
4,16000,0.12506804196164012
|
110 |
+
4,17000,0.12520997552201152
|
111 |
+
4,18000,0.12657485203817487
|
112 |
+
4,19000,0.12641792418435216
|
113 |
+
4,20000,0.12666251277551055
|
114 |
+
4,21000,0.12724685948342085
|
115 |
+
4,22000,0.12806769227609038
|
116 |
+
4,-1,0.12863221345469356
|
117 |
+
5,1000,0.12642602669075131
|
118 |
+
5,2000,0.12800676049664617
|
119 |
+
5,3000,0.12629072880372405
|
120 |
+
5,4000,0.1290170243009925
|
121 |
+
5,5000,0.12826438760384917
|
122 |
+
5,6000,0.12519805459305644
|
123 |
+
5,7000,0.125317950733006
|
124 |
+
5,8000,0.12615734012797475
|
125 |
+
5,9000,0.12642822694033384
|
126 |
+
5,10000,0.12593038845807314
|
127 |
+
5,11000,0.12542834738269448
|
128 |
+
5,12000,0.12669269926846027
|
129 |
+
5,13000,0.12542062904685736
|
130 |
+
5,14000,0.12645676033571362
|
131 |
+
5,15000,0.12655231403186917
|
132 |
+
5,16000,0.126134033780545
|
133 |
+
5,17000,0.12598291505128145
|
134 |
+
5,18000,0.12577302986755967
|
135 |
+
5,19000,0.125892402138561
|
136 |
+
5,20000,0.12707466958090663
|
137 |
+
5,21000,0.12666366528719664
|
138 |
+
5,22000,0.12758017983287573
|
139 |
+
5,-1,0.12715521734207869
|
140 |
+
6,1000,0.12620933121070266
|
141 |
+
6,2000,0.12544727651402354
|
142 |
+
6,3000,0.12546692742034793
|
143 |
+
6,4000,0.12546839425340295
|
144 |
+
6,5000,0.1252964953891933
|
145 |
+
6,6000,0.12533211847767234
|
146 |
+
6,7000,0.1251009409315884
|
147 |
+
6,8000,0.12540931347757578
|
148 |
+
6,9000,0.12469106586650014
|
149 |
+
6,10000,0.1251082867383957
|
150 |
+
6,11000,0.12571561383083463
|
151 |
+
6,12000,0.1252957619726658
|
152 |
+
6,13000,0.12519304873421788
|
153 |
+
6,14000,0.1253728405572474
|
154 |
+
6,15000,0.12530144304037094
|
155 |
+
6,16000,0.12525240890681744
|
156 |
+
6,17000,0.12537494767457247
|
157 |
+
6,18000,0.12520120944827795
|
158 |
+
6,19000,0.12507732026278973
|
159 |
+
6,20000,0.12561904732137918
|
160 |
+
6,21000,0.12518406147137284
|
161 |
+
6,22000,0.1257007708773017
|
162 |
+
6,-1,0.12541512260213494
|
163 |
+
7,1000,0.12544342316687107
|
164 |
+
7,2000,0.12556433212012053
|
165 |
+
7,3000,0.12590863043442369
|
166 |
+
7,4000,0.12514503905549645
|
167 |
+
7,5000,0.1253378577530384
|
168 |
+
7,6000,0.12566092191264033
|
169 |
+
7,7000,0.1251959940418601
|
170 |
+
7,8000,0.12585095828399062
|
171 |
+
7,9000,0.12555383145809174
|
172 |
+
7,10000,0.12559937313199043
|
173 |
+
7,11000,0.12625348754227161
|
174 |
+
7,12000,0.12589046964421868
|
175 |
+
7,13000,0.125878828112036
|
176 |
+
7,14000,0.12598383473232388
|
177 |
+
7,15000,0.12609409168362617
|
178 |
+
7,16000,0.12586734956130385
|
179 |
+
7,17000,0.12592717539519072
|
180 |
+
7,18000,0.12580857146531343
|
181 |
+
7,19000,0.12572628911584616
|
182 |
+
7,20000,0.12575872242450714
|
183 |
+
7,21000,0.12580197071656585
|
184 |
+
7,22000,0.12579445028677583
|
185 |
+
7,-1,0.12579462490975857
|
eval/mse_evaluation_Tatoeba-eng-ukr-dev.tsv.gz_results.csv
ADDED
@@ -0,0 +1,185 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,MSE
|
2 |
+
0,1000,0.8054990321397781
|
3 |
+
0,2000,0.47793304547667503
|
4 |
+
0,3000,0.6615580059587955
|
5 |
+
0,4000,42.05380976200104
|
6 |
+
0,5000,37.66351044178009
|
7 |
+
0,6000,28.103342652320862
|
8 |
+
0,7000,14.893744885921478
|
9 |
+
0,8000,18.412677943706512
|
10 |
+
0,9000,14.194142818450928
|
11 |
+
0,10000,20.698603987693787
|
12 |
+
0,11000,14.665719866752625
|
13 |
+
0,12000,0.1389687997289002
|
14 |
+
0,13000,0.15049357898533344
|
15 |
+
0,14000,0.13093536254018545
|
16 |
+
0,15000,8.95860269665718
|
17 |
+
0,16000,26.40240788459778
|
18 |
+
0,17000,19.193385541439056
|
19 |
+
0,18000,8.742277324199677
|
20 |
+
0,19000,5.295905098319054
|
21 |
+
0,20000,4.525310918688774
|
22 |
+
0,21000,5.251230299472809
|
23 |
+
0,22000,3.1860072165727615
|
24 |
+
0,-1,11.993832886219025
|
25 |
+
1,1000,6.250041723251343
|
26 |
+
1,2000,6.328821927309036
|
27 |
+
1,3000,12.175609171390533
|
28 |
+
1,4000,3.9220526814460754
|
29 |
+
1,5000,14.225760102272034
|
30 |
+
1,6000,11.053801327943802
|
31 |
+
1,7000,13.287901878356934
|
32 |
+
1,8000,3.8226719945669174
|
33 |
+
1,9000,3.6519233137369156
|
34 |
+
1,10000,3.4523718059062958
|
35 |
+
1,11000,0.12269968865439296
|
36 |
+
1,12000,0.12397165410220623
|
37 |
+
1,13000,0.12290896847844124
|
38 |
+
1,14000,0.12419394915923476
|
39 |
+
1,15000,0.12384355068206787
|
40 |
+
1,16000,0.1259002834558487
|
41 |
+
1,17000,0.12366366572678089
|
42 |
+
1,18000,0.12348245363682508
|
43 |
+
1,19000,0.12393235228955746
|
44 |
+
1,20000,0.12636043829843402
|
45 |
+
1,21000,0.12466326588764787
|
46 |
+
1,22000,0.1255976501852274
|
47 |
+
1,-1,0.1256137154996395
|
48 |
+
2,1000,0.1287964405491948
|
49 |
+
2,2000,0.12630567653104663
|
50 |
+
2,3000,0.1253655762411654
|
51 |
+
2,4000,0.12530474923551083
|
52 |
+
2,5000,0.12748391600325704
|
53 |
+
2,6000,0.1263804384507239
|
54 |
+
2,7000,0.12450109934434295
|
55 |
+
2,8000,0.12522073229774833
|
56 |
+
2,9000,0.1290328335016966
|
57 |
+
2,10000,0.1253199647180736
|
58 |
+
2,11000,0.12431163340806961
|
59 |
+
2,12000,0.12405728921294212
|
60 |
+
2,13000,0.12577766319736838
|
61 |
+
2,14000,0.12671578442677855
|
62 |
+
2,15000,0.12822209391742945
|
63 |
+
2,16000,0.12833329383283854
|
64 |
+
2,17000,0.12560287723317742
|
65 |
+
2,18000,0.12336722575128078
|
66 |
+
2,19000,0.1235613483004272
|
67 |
+
2,20000,0.12324543204158545
|
68 |
+
2,21000,0.12404901208356023
|
69 |
+
2,22000,0.12313963379710913
|
70 |
+
2,-1,0.1236567273736
|
71 |
+
3,1000,0.12402331922203302
|
72 |
+
3,2000,0.12336981017142534
|
73 |
+
3,3000,0.12433367082849145
|
74 |
+
3,4000,0.12314142659306526
|
75 |
+
3,5000,0.12318437220528722
|
76 |
+
3,6000,0.12359702959656715
|
77 |
+
3,7000,0.1231974340043962
|
78 |
+
3,8000,0.1236186595633626
|
79 |
+
3,9000,0.12440326390787959
|
80 |
+
3,10000,0.12481571175158024
|
81 |
+
3,11000,0.12431573122739792
|
82 |
+
3,12000,0.12561732437461615
|
83 |
+
3,13000,0.1248013461008668
|
84 |
+
3,14000,0.12722813989967108
|
85 |
+
3,15000,0.1275797374546528
|
86 |
+
3,16000,0.12423582375049591
|
87 |
+
3,17000,0.12318630469962955
|
88 |
+
3,18000,0.12369363103061914
|
89 |
+
3,19000,0.12428282061591744
|
90 |
+
3,20000,0.12305304408073425
|
91 |
+
3,21000,0.12351544573903084
|
92 |
+
3,22000,0.12311561731621623
|
93 |
+
3,-1,0.12360598193481565
|
94 |
+
4,1000,0.12369530741125345
|
95 |
+
4,2000,0.12366650626063347
|
96 |
+
4,3000,0.12371109332889318
|
97 |
+
4,4000,0.12373398058116436
|
98 |
+
4,5000,0.12460004072636366
|
99 |
+
4,6000,0.12386356247588992
|
100 |
+
4,7000,0.12358080130070448
|
101 |
+
4,8000,0.12383370194584131
|
102 |
+
4,9000,0.12345609720796347
|
103 |
+
4,10000,0.12429772177711129
|
104 |
+
4,11000,0.12324347626417875
|
105 |
+
4,12000,0.1233651302754879
|
106 |
+
4,13000,0.12377903331071138
|
107 |
+
4,14000,0.12338090455159545
|
108 |
+
4,15000,0.12387686874717474
|
109 |
+
4,16000,0.123142811935395
|
110 |
+
4,17000,0.12337157968431711
|
111 |
+
4,18000,0.12481666635721922
|
112 |
+
4,19000,0.12445036554709077
|
113 |
+
4,20000,0.12473324313759804
|
114 |
+
4,21000,0.12555629946291447
|
115 |
+
4,22000,0.12646267423406243
|
116 |
+
4,-1,0.1270242384634912
|
117 |
+
5,1000,0.12452059891074896
|
118 |
+
5,2000,0.12606095988303423
|
119 |
+
5,3000,0.12432234361767769
|
120 |
+
5,4000,0.12658301275223494
|
121 |
+
5,5000,0.12622650247067213
|
122 |
+
5,6000,0.12355041690170765
|
123 |
+
5,7000,0.12346578296273947
|
124 |
+
5,8000,0.12455566320568323
|
125 |
+
5,9000,0.12474027462303638
|
126 |
+
5,10000,0.12430166825652122
|
127 |
+
5,11000,0.1238511991687119
|
128 |
+
5,12000,0.12524921912699938
|
129 |
+
5,13000,0.12373232748359442
|
130 |
+
5,14000,0.12470505898818374
|
131 |
+
5,15000,0.12461120495572686
|
132 |
+
5,16000,0.12418084079399705
|
133 |
+
5,17000,0.12404794106259942
|
134 |
+
5,18000,0.12389132753014565
|
135 |
+
5,19000,0.12383211869746447
|
136 |
+
5,20000,0.12506392085924745
|
137 |
+
5,21000,0.12490013614296913
|
138 |
+
5,22000,0.12559752212837338
|
139 |
+
5,-1,0.1251919544301927
|
140 |
+
6,1000,0.12436301913112402
|
141 |
+
6,2000,0.12355090584605932
|
142 |
+
6,3000,0.12357770465314388
|
143 |
+
6,4000,0.12361875269562006
|
144 |
+
6,5000,0.12351239565759897
|
145 |
+
6,6000,0.12347641168162227
|
146 |
+
6,7000,0.12330394238233566
|
147 |
+
6,8000,0.12360374676063657
|
148 |
+
6,9000,0.12283488176763058
|
149 |
+
6,10000,0.12329661985859275
|
150 |
+
6,11000,0.12394378427416086
|
151 |
+
6,12000,0.12342417612671852
|
152 |
+
6,13000,0.12325469870120287
|
153 |
+
6,14000,0.12352274497970939
|
154 |
+
6,15000,0.12338808737695217
|
155 |
+
6,16000,0.123451033141464
|
156 |
+
6,17000,0.12355117360129952
|
157 |
+
6,18000,0.12339792447164655
|
158 |
+
6,19000,0.12331136967986822
|
159 |
+
6,20000,0.12373700737953186
|
160 |
+
6,21000,0.12345025315880775
|
161 |
+
6,22000,0.12411287752911448
|
162 |
+
6,-1,0.12393624056130648
|
163 |
+
7,1000,0.12376323575153947
|
164 |
+
7,2000,0.12389406329020858
|
165 |
+
7,3000,0.1241093035787344
|
166 |
+
7,4000,0.12334435014054179
|
167 |
+
7,5000,0.12357927625998855
|
168 |
+
7,6000,0.12391237542033195
|
169 |
+
7,7000,0.12356614461168647
|
170 |
+
7,8000,0.12426013126969337
|
171 |
+
7,9000,0.12394718360155821
|
172 |
+
7,10000,0.12392947683110833
|
173 |
+
7,11000,0.12458892306312919
|
174 |
+
7,12000,0.12418971164152026
|
175 |
+
7,13000,0.12411647476255894
|
176 |
+
7,14000,0.12425838503986597
|
177 |
+
7,15000,0.12436830438673496
|
178 |
+
7,16000,0.12419078266248107
|
179 |
+
7,17000,0.12426362372934818
|
180 |
+
7,18000,0.12410922208800912
|
181 |
+
7,19000,0.12405223678797483
|
182 |
+
7,20000,0.12411032803356647
|
183 |
+
7,21000,0.12414938537403941
|
184 |
+
7,22000,0.12413804652169347
|
185 |
+
7,-1,0.12413694057613611
|
eval/translation_evaluation_Tatoeba-eng-rus-dev.tsv.gz_results.csv
ADDED
@@ -0,0 +1,185 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,src2trg,trg2src
|
2 |
+
0,1000,0.00195,0.00335
|
3 |
+
0,2000,0.00135,0.0013
|
4 |
+
0,3000,0.00025,0.00055
|
5 |
+
0,4000,0.00025,0.00025
|
6 |
+
0,5000,5e-05,0.00025
|
7 |
+
0,6000,0.0,0.0001
|
8 |
+
0,7000,0.00015,0.00015
|
9 |
+
0,8000,0.0002,0.00025
|
10 |
+
0,9000,0.0001,0.0002
|
11 |
+
0,10000,0.0002,0.00045
|
12 |
+
0,11000,0.00035,0.0004
|
13 |
+
0,12000,0.00015,0.00025
|
14 |
+
0,13000,0.0001,0.0002
|
15 |
+
0,14000,0.0003,0.0001
|
16 |
+
0,15000,0.00015,0.00025
|
17 |
+
0,16000,5e-05,0.0001
|
18 |
+
0,17000,5e-05,5e-05
|
19 |
+
0,18000,0.00015,5e-05
|
20 |
+
0,19000,5e-05,5e-05
|
21 |
+
0,20000,0.0,5e-05
|
22 |
+
0,21000,0.0001,0.00015
|
23 |
+
0,22000,5e-05,0.00025
|
24 |
+
0,-1,0.0001,0.0002
|
25 |
+
1,1000,5e-05,0.0002
|
26 |
+
1,2000,0.00015,0.0001
|
27 |
+
1,3000,0.0,0.0
|
28 |
+
1,4000,0.0001,0.00015
|
29 |
+
1,5000,5e-05,5e-05
|
30 |
+
1,6000,0.0001,0.0002
|
31 |
+
1,7000,0.0,0.0002
|
32 |
+
1,8000,5e-05,0.0003
|
33 |
+
1,9000,5e-05,5e-05
|
34 |
+
1,10000,0.0002,0.0002
|
35 |
+
1,11000,0.0,0.00015
|
36 |
+
1,12000,0.0003,0.0004
|
37 |
+
1,13000,0.0001,0.00025
|
38 |
+
1,14000,0.0002,0.0004
|
39 |
+
1,15000,0.0002,5e-05
|
40 |
+
1,16000,0.0001,0.0002
|
41 |
+
1,17000,0.00025,0.0002
|
42 |
+
1,18000,0.0003,0.00035
|
43 |
+
1,19000,5e-05,0.0001
|
44 |
+
1,20000,0.00025,0.0002
|
45 |
+
1,21000,0.0003,0.0002
|
46 |
+
1,22000,0.00015,0.00025
|
47 |
+
1,-1,0.0001,0.0002
|
48 |
+
2,1000,0.0002,0.0004
|
49 |
+
2,2000,0.00065,0.00045
|
50 |
+
2,3000,0.00025,0.00055
|
51 |
+
2,4000,0.0002,0.0002
|
52 |
+
2,5000,0.00025,0.00035
|
53 |
+
2,6000,0.00025,0.00035
|
54 |
+
2,7000,0.00035,0.00035
|
55 |
+
2,8000,0.0003,0.00045
|
56 |
+
2,9000,0.00035,0.00065
|
57 |
+
2,10000,0.0003,0.00025
|
58 |
+
2,11000,0.0003,0.00035
|
59 |
+
2,12000,0.0003,0.0005
|
60 |
+
2,13000,5e-05,0.00025
|
61 |
+
2,14000,0.00025,0.00015
|
62 |
+
2,15000,0.0003,0.00035
|
63 |
+
2,16000,0.00015,0.00035
|
64 |
+
2,17000,5e-05,0.00015
|
65 |
+
2,18000,0.00015,0.0003
|
66 |
+
2,19000,0.0001,0.0001
|
67 |
+
2,20000,0.00015,0.0002
|
68 |
+
2,21000,0.0001,0.0
|
69 |
+
2,22000,5e-05,0.00025
|
70 |
+
2,-1,5e-05,0.00015
|
71 |
+
3,1000,0.0001,0.00025
|
72 |
+
3,2000,5e-05,0.0
|
73 |
+
3,3000,0.00015,0.0002
|
74 |
+
3,4000,0.0001,5e-05
|
75 |
+
3,5000,0.00015,0.0001
|
76 |
+
3,6000,0.00015,0.0003
|
77 |
+
3,7000,0.00015,0.00045
|
78 |
+
3,8000,0.00025,0.00025
|
79 |
+
3,9000,0.0003,0.0003
|
80 |
+
3,10000,0.0003,0.0003
|
81 |
+
3,11000,0.00045,0.0004
|
82 |
+
3,12000,0.0001,0.0005
|
83 |
+
3,13000,0.0004,0.0004
|
84 |
+
3,14000,0.0003,0.0002
|
85 |
+
3,15000,0.0002,0.0003
|
86 |
+
3,16000,0.00035,0.0005
|
87 |
+
3,17000,0.00035,0.0003
|
88 |
+
3,18000,0.00015,0.00045
|
89 |
+
3,19000,0.0003,0.00055
|
90 |
+
3,20000,0.00025,0.0004
|
91 |
+
3,21000,0.00035,0.00045
|
92 |
+
3,22000,0.0002,0.00025
|
93 |
+
3,-1,0.0002,0.00025
|
94 |
+
4,1000,0.0002,0.00035
|
95 |
+
4,2000,0.0003,0.0003
|
96 |
+
4,3000,0.00025,0.0005
|
97 |
+
4,4000,0.0003,0.00025
|
98 |
+
4,5000,0.0002,0.00035
|
99 |
+
4,6000,0.00035,0.00035
|
100 |
+
4,7000,0.00015,0.0003
|
101 |
+
4,8000,0.00025,0.00035
|
102 |
+
4,9000,0.00025,0.0005
|
103 |
+
4,10000,0.00015,0.0003
|
104 |
+
4,11000,0.0003,0.00055
|
105 |
+
4,12000,0.00025,0.00025
|
106 |
+
4,13000,0.0003,0.0004
|
107 |
+
4,14000,0.00035,0.00035
|
108 |
+
4,15000,0.00015,0.0004
|
109 |
+
4,16000,0.00045,0.00035
|
110 |
+
4,17000,0.00015,0.0003
|
111 |
+
4,18000,0.00015,0.00015
|
112 |
+
4,19000,0.00035,0.0002
|
113 |
+
4,20000,0.0001,0.00015
|
114 |
+
4,21000,0.00025,0.00025
|
115 |
+
4,22000,0.00025,0.00035
|
116 |
+
4,-1,0.00015,0.00015
|
117 |
+
5,1000,0.0003,0.0002
|
118 |
+
5,2000,0.00025,0.00025
|
119 |
+
5,3000,0.00045,0.0004
|
120 |
+
5,4000,0.0001,0.0004
|
121 |
+
5,5000,0.00025,0.00015
|
122 |
+
5,6000,0.00015,0.00045
|
123 |
+
5,7000,0.00035,0.00035
|
124 |
+
5,8000,0.0004,0.00025
|
125 |
+
5,9000,0.00025,0.0003
|
126 |
+
5,10000,0.0002,0.0004
|
127 |
+
5,11000,0.0003,0.00035
|
128 |
+
5,12000,0.00025,0.00045
|
129 |
+
5,13000,0.00035,0.0004
|
130 |
+
5,14000,0.00015,0.0002
|
131 |
+
5,15000,0.0002,0.0002
|
132 |
+
5,16000,0.0001,0.0002
|
133 |
+
5,17000,0.00015,0.0003
|
134 |
+
5,18000,0.0001,0.00015
|
135 |
+
5,19000,0.00025,0.0003
|
136 |
+
5,20000,0.0002,0.0005
|
137 |
+
5,21000,0.0002,0.00035
|
138 |
+
5,22000,5e-05,0.00015
|
139 |
+
5,-1,0.0001,0.0003
|
140 |
+
6,1000,0.0001,0.00025
|
141 |
+
6,2000,0.0003,0.0002
|
142 |
+
6,3000,0.00025,0.0001
|
143 |
+
6,4000,5e-05,0.0002
|
144 |
+
6,5000,0.0001,0.0002
|
145 |
+
6,6000,0.0004,0.00025
|
146 |
+
6,7000,0.0001,0.0003
|
147 |
+
6,8000,0.0002,0.00025
|
148 |
+
6,9000,0.00015,0.0002
|
149 |
+
6,10000,0.0001,0.0002
|
150 |
+
6,11000,0.0003,0.0003
|
151 |
+
6,12000,0.0002,0.00025
|
152 |
+
6,13000,0.0001,0.0002
|
153 |
+
6,14000,0.0003,0.0004
|
154 |
+
6,15000,0.00015,0.00025
|
155 |
+
6,16000,0.0001,0.00015
|
156 |
+
6,17000,0.00015,0.0
|
157 |
+
6,18000,5e-05,0.00025
|
158 |
+
6,19000,0.0001,0.0003
|
159 |
+
6,20000,5e-05,0.00035
|
160 |
+
6,21000,0.0003,0.0005
|
161 |
+
6,22000,0.0003,0.00035
|
162 |
+
6,-1,0.00035,0.00045
|
163 |
+
7,1000,0.00015,0.0002
|
164 |
+
7,2000,0.0002,0.00025
|
165 |
+
7,3000,0.0003,0.00025
|
166 |
+
7,4000,0.00025,0.00015
|
167 |
+
7,5000,0.0003,0.0003
|
168 |
+
7,6000,0.00015,0.0002
|
169 |
+
7,7000,0.0003,0.00025
|
170 |
+
7,8000,0.0003,0.00025
|
171 |
+
7,9000,5e-05,0.00015
|
172 |
+
7,10000,0.0001,0.00025
|
173 |
+
7,11000,0.0,0.00035
|
174 |
+
7,12000,5e-05,0.0003
|
175 |
+
7,13000,5e-05,0.0003
|
176 |
+
7,14000,0.00015,0.0003
|
177 |
+
7,15000,5e-05,0.0003
|
178 |
+
7,16000,5e-05,0.00025
|
179 |
+
7,17000,0.0003,0.00025
|
180 |
+
7,18000,0.0001,0.00025
|
181 |
+
7,19000,0.00015,0.00055
|
182 |
+
7,20000,0.0001,0.0003
|
183 |
+
7,21000,0.0001,0.0002
|
184 |
+
7,22000,0.00015,0.0002
|
185 |
+
7,-1,0.0001,0.0003
|
eval/translation_evaluation_Tatoeba-eng-ukr-dev.tsv.gz_results.csv
ADDED
@@ -0,0 +1,185 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,src2trg,trg2src
|
2 |
+
0,1000,0.0012,0.00155
|
3 |
+
0,2000,0.00045,0.00085
|
4 |
+
0,3000,0.00035,0.0005
|
5 |
+
0,4000,0.00015,0.00035
|
6 |
+
0,5000,0.00015,0.0003
|
7 |
+
0,6000,0.0,0.0001
|
8 |
+
0,7000,0.00035,0.0002
|
9 |
+
0,8000,5e-05,0.00025
|
10 |
+
0,9000,0.00015,0.0002
|
11 |
+
0,10000,0.00015,0.0003
|
12 |
+
0,11000,0.00025,0.0003
|
13 |
+
0,12000,0.0003,0.00035
|
14 |
+
0,13000,0.00015,0.0
|
15 |
+
0,14000,5e-05,0.00025
|
16 |
+
0,15000,0.00015,0.00015
|
17 |
+
0,16000,0.0001,0.0
|
18 |
+
0,17000,0.0,0.0
|
19 |
+
0,18000,0.0001,0.0001
|
20 |
+
0,19000,5e-05,0.0001
|
21 |
+
0,20000,0.0,0.0
|
22 |
+
0,21000,5e-05,5e-05
|
23 |
+
0,22000,0.0002,0.00025
|
24 |
+
0,-1,5e-05,0.0001
|
25 |
+
1,1000,0.0001,5e-05
|
26 |
+
1,2000,5e-05,5e-05
|
27 |
+
1,3000,5e-05,0.0002
|
28 |
+
1,4000,5e-05,0.0001
|
29 |
+
1,5000,0.00015,0.0
|
30 |
+
1,6000,5e-05,0.00015
|
31 |
+
1,7000,5e-05,5e-05
|
32 |
+
1,8000,0.00025,0.00025
|
33 |
+
1,9000,5e-05,5e-05
|
34 |
+
1,10000,0.00015,0.0001
|
35 |
+
1,11000,5e-05,0.00025
|
36 |
+
1,12000,0.0002,0.00025
|
37 |
+
1,13000,0.0003,0.00035
|
38 |
+
1,14000,0.00015,0.0002
|
39 |
+
1,15000,0.0004,0.00025
|
40 |
+
1,16000,0.00035,0.00035
|
41 |
+
1,17000,0.00015,0.00035
|
42 |
+
1,18000,0.00015,0.00015
|
43 |
+
1,19000,0.0001,5e-05
|
44 |
+
1,20000,0.0001,0.0001
|
45 |
+
1,21000,0.00015,0.0003
|
46 |
+
1,22000,0.0002,0.00055
|
47 |
+
1,-1,0.0001,0.0001
|
48 |
+
2,1000,0.00015,0.00015
|
49 |
+
2,2000,0.0001,0.0002
|
50 |
+
2,3000,0.00035,0.00025
|
51 |
+
2,4000,0.00025,0.00015
|
52 |
+
2,5000,0.00025,0.00035
|
53 |
+
2,6000,0.00025,0.00025
|
54 |
+
2,7000,0.0002,0.00025
|
55 |
+
2,8000,0.00025,0.0003
|
56 |
+
2,9000,0.0004,0.00035
|
57 |
+
2,10000,0.0003,0.00015
|
58 |
+
2,11000,0.00025,0.00015
|
59 |
+
2,12000,0.00025,0.0004
|
60 |
+
2,13000,0.00015,0.0003
|
61 |
+
2,14000,0.00025,0.0003
|
62 |
+
2,15000,0.0002,0.00045
|
63 |
+
2,16000,0.0003,0.00055
|
64 |
+
2,17000,0.0001,0.0002
|
65 |
+
2,18000,0.0002,0.00015
|
66 |
+
2,19000,0.0002,0.0001
|
67 |
+
2,20000,0.00035,0.0001
|
68 |
+
2,21000,0.0001,0.0001
|
69 |
+
2,22000,0.0002,0.0003
|
70 |
+
2,-1,0.0001,0.00015
|
71 |
+
3,1000,0.0003,0.0002
|
72 |
+
3,2000,5e-05,0.00015
|
73 |
+
3,3000,0.00015,0.0003
|
74 |
+
3,4000,0.0001,0.0002
|
75 |
+
3,5000,0.00035,0.00025
|
76 |
+
3,6000,5e-05,0.0002
|
77 |
+
3,7000,0.0001,0.00025
|
78 |
+
3,8000,0.00025,0.0002
|
79 |
+
3,9000,0.00025,0.00035
|
80 |
+
3,10000,0.0003,0.00035
|
81 |
+
3,11000,5e-05,0.0002
|
82 |
+
3,12000,0.00045,0.0003
|
83 |
+
3,13000,0.00025,0.00035
|
84 |
+
3,14000,0.0001,0.00015
|
85 |
+
3,15000,0.0003,0.0002
|
86 |
+
3,16000,0.0003,0.00025
|
87 |
+
3,17000,5e-05,5e-05
|
88 |
+
3,18000,0.0002,0.0002
|
89 |
+
3,19000,0.00015,0.00035
|
90 |
+
3,20000,0.0003,0.00025
|
91 |
+
3,21000,0.00035,0.00025
|
92 |
+
3,22000,0.0003,0.00035
|
93 |
+
3,-1,0.0001,5e-05
|
94 |
+
4,1000,0.0002,0.0002
|
95 |
+
4,2000,0.00025,0.00015
|
96 |
+
4,3000,0.0004,0.00055
|
97 |
+
4,4000,0.0002,0.0005
|
98 |
+
4,5000,0.00025,0.00035
|
99 |
+
4,6000,0.0002,0.00025
|
100 |
+
4,7000,0.0003,0.0004
|
101 |
+
4,8000,0.0003,0.00045
|
102 |
+
4,9000,0.00025,0.0005
|
103 |
+
4,10000,0.00025,0.00025
|
104 |
+
4,11000,0.00015,0.00055
|
105 |
+
4,12000,0.00025,0.0004
|
106 |
+
4,13000,0.0003,0.00025
|
107 |
+
4,14000,0.0002,0.00025
|
108 |
+
4,15000,0.0002,0.00035
|
109 |
+
4,16000,0.00015,0.00015
|
110 |
+
4,17000,0.00035,0.0003
|
111 |
+
4,18000,0.0003,0.0003
|
112 |
+
4,19000,0.00025,0.00015
|
113 |
+
4,20000,0.00015,0.0001
|
114 |
+
4,21000,0.00015,0.0003
|
115 |
+
4,22000,0.0002,0.00015
|
116 |
+
4,-1,0.0002,0.00015
|
117 |
+
5,1000,0.0001,0.00045
|
118 |
+
5,2000,0.0002,0.00035
|
119 |
+
5,3000,0.00025,0.0003
|
120 |
+
5,4000,0.0003,0.0003
|
121 |
+
5,5000,0.0002,0.00015
|
122 |
+
5,6000,0.0003,0.0002
|
123 |
+
5,7000,5e-05,0.00045
|
124 |
+
5,8000,0.0002,0.0003
|
125 |
+
5,9000,0.00025,0.0001
|
126 |
+
5,10000,0.0001,0.0001
|
127 |
+
5,11000,0.0003,0.0003
|
128 |
+
5,12000,0.00015,0.00025
|
129 |
+
5,13000,0.0003,0.00035
|
130 |
+
5,14000,5e-05,0.0002
|
131 |
+
5,15000,0.0001,0.0003
|
132 |
+
5,16000,0.0002,0.0004
|
133 |
+
5,17000,0.0002,0.0002
|
134 |
+
5,18000,0.00035,0.00025
|
135 |
+
5,19000,0.00015,0.0004
|
136 |
+
5,20000,0.00025,0.0004
|
137 |
+
5,21000,0.0003,0.00025
|
138 |
+
5,22000,0.0002,0.0004
|
139 |
+
5,-1,0.0001,0.00035
|
140 |
+
6,1000,0.0001,0.00025
|
141 |
+
6,2000,0.0001,0.0002
|
142 |
+
6,3000,0.00015,0.00015
|
143 |
+
6,4000,0.0003,0.0003
|
144 |
+
6,5000,5e-05,0.00025
|
145 |
+
6,6000,0.00025,0.0003
|
146 |
+
6,7000,5e-05,0.00015
|
147 |
+
6,8000,0.0002,0.0004
|
148 |
+
6,9000,0.0001,0.0002
|
149 |
+
6,10000,0.0002,0.0003
|
150 |
+
6,11000,0.0002,0.0002
|
151 |
+
6,12000,0.00025,0.00015
|
152 |
+
6,13000,0.0,0.00015
|
153 |
+
6,14000,0.0001,0.00035
|
154 |
+
6,15000,0.0003,0.00035
|
155 |
+
6,16000,0.0004,0.00025
|
156 |
+
6,17000,0.00015,0.0002
|
157 |
+
6,18000,0.0002,0.00025
|
158 |
+
6,19000,0.00015,0.0002
|
159 |
+
6,20000,0.00015,5e-05
|
160 |
+
6,21000,0.00015,0.00045
|
161 |
+
6,22000,0.00015,5e-05
|
162 |
+
6,-1,0.00015,0.0001
|
163 |
+
7,1000,0.0001,0.0003
|
164 |
+
7,2000,5e-05,0.0004
|
165 |
+
7,3000,0.00015,0.00025
|
166 |
+
7,4000,0.00015,0.00025
|
167 |
+
7,5000,0.0001,0.0003
|
168 |
+
7,6000,0.0002,0.00035
|
169 |
+
7,7000,0.00015,0.00015
|
170 |
+
7,8000,0.0002,0.0002
|
171 |
+
7,9000,5e-05,0.00025
|
172 |
+
7,10000,0.00015,0.00025
|
173 |
+
7,11000,0.0001,0.00025
|
174 |
+
7,12000,0.00025,0.00025
|
175 |
+
7,13000,0.0002,0.0003
|
176 |
+
7,14000,0.0002,0.0002
|
177 |
+
7,15000,0.0001,0.00035
|
178 |
+
7,16000,0.0001,0.00035
|
179 |
+
7,17000,0.0003,0.0002
|
180 |
+
7,18000,0.00025,0.0002
|
181 |
+
7,19000,0.0002,0.00035
|
182 |
+
7,20000,0.00025,0.00035
|
183 |
+
7,21000,0.00025,0.00015
|
184 |
+
7,22000,0.00015,0.0003
|
185 |
+
7,-1,0.0001,0.00025
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cccf452832436f9f6f9292aff96a079f188f30429764e987c60ce947db0dfe09
|
3 |
+
size 1112197096
|
modules.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"idx": 0,
|
4 |
+
"name": "0",
|
5 |
+
"path": "",
|
6 |
+
"type": "sentence_transformers.models.Transformer"
|
7 |
+
},
|
8 |
+
{
|
9 |
+
"idx": 1,
|
10 |
+
"name": "1",
|
11 |
+
"path": "1_Pooling",
|
12 |
+
"type": "sentence_transformers.models.Pooling"
|
13 |
+
}
|
14 |
+
]
|
sentence_bert_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"max_seq_length": 384,
|
3 |
+
"do_lower_case": false
|
4 |
+
}
|
sentencepiece.bpe.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
|
3 |
+
size 5069051
|
special_tokens_map.json
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<s>",
|
3 |
+
"cls_token": "<s>",
|
4 |
+
"eos_token": "</s>",
|
5 |
+
"mask_token": {
|
6 |
+
"content": "<mask>",
|
7 |
+
"lstrip": true,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false
|
11 |
+
},
|
12 |
+
"pad_token": "<pad>",
|
13 |
+
"sep_token": "</s>",
|
14 |
+
"unk_token": "<unk>"
|
15 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8ab208220fd15ae86d71654fbffe08ba919e926330fa27dc19fe32874e3e0492
|
3 |
+
size 17083009
|
tokenizer_config.json
ADDED
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<s>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<pad>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "</s>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"3": {
|
28 |
+
"content": "<unk>",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"250001": {
|
36 |
+
"content": "<mask>",
|
37 |
+
"lstrip": true,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"bos_token": "<s>",
|
45 |
+
"clean_up_tokenization_spaces": true,
|
46 |
+
"cls_token": "<s>",
|
47 |
+
"eos_token": "</s>",
|
48 |
+
"mask_token": "<mask>",
|
49 |
+
"model_max_length": 512,
|
50 |
+
"pad_token": "<pad>",
|
51 |
+
"sep_token": "</s>",
|
52 |
+
"tokenizer_class": "XLMRobertaTokenizer",
|
53 |
+
"unk_token": "<unk>"
|
54 |
+
}
|