Goshective commited on
Commit
9432920
·
verified ·
1 Parent(s): 63672e3

kazparc_en_ru_marian_1

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ target.spm filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-ru](https://huggingface.co/Helsinki-NLP/opus-mt-en-ru) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.0291
20
 
21
  ## Model description
22
 
@@ -39,8 +39,6 @@ The following hyperparameters were used during training:
39
  - train_batch_size: 8
40
  - eval_batch_size: 16
41
  - seed: 42
42
- - gradient_accumulation_steps: 4
43
- - total_train_batch_size: 32
44
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
  - lr_scheduler_type: linear
46
  - num_epochs: 2
@@ -48,15 +46,15 @@ The following hyperparameters were used during training:
48
 
49
  ### Training results
50
 
51
- | Training Loss | Epoch | Step | Validation Loss |
52
- |:-------------:|:------:|:----:|:---------------:|
53
- | 0.0337 | 0.9997 | 2825 | 0.0295 |
54
- | 0.0298 | 1.9995 | 5650 | 0.0291 |
55
 
56
 
57
  ### Framework versions
58
 
59
- - Transformers 4.46.2
60
  - Pytorch 2.5.1+cu121
61
  - Datasets 3.1.0
62
  - Tokenizers 0.20.3
 
16
 
17
  This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-ru](https://huggingface.co/Helsinki-NLP/opus-mt-en-ru) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.0186
20
 
21
  ## Model description
22
 
 
39
  - train_batch_size: 8
40
  - eval_batch_size: 16
41
  - seed: 42
 
 
42
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 2
 
46
 
47
  ### Training results
48
 
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:-----:|:-----:|:---------------:|
51
+ | 0.0225 | 1.0 | 9492 | 0.0195 |
52
+ | 0.0168 | 2.0 | 18984 | 0.0186 |
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.46.3
58
  - Pytorch 2.5.1+cu121
59
  - Datasets 3.1.0
60
  - Tokenizers 0.20.3
config.json CHANGED
@@ -50,7 +50,7 @@
50
  "share_encoder_decoder_embeddings": true,
51
  "static_position_embeddings": true,
52
  "torch_dtype": "float32",
53
- "transformers_version": "4.46.2",
54
  "use_cache": true,
55
  "vocab_size": 62518
56
  }
 
50
  "share_encoder_decoder_embeddings": true,
51
  "static_position_embeddings": true,
52
  "torch_dtype": "float32",
53
+ "transformers_version": "4.46.3",
54
  "use_cache": true,
55
  "vocab_size": 62518
56
  }
generation_config.json CHANGED
@@ -12,5 +12,5 @@
12
  "num_beams": 4,
13
  "pad_token_id": 62517,
14
  "renormalize_logits": true,
15
- "transformers_version": "4.46.2"
16
  }
 
12
  "num_beams": 4,
13
  "pad_token_id": 62517,
14
  "renormalize_logits": true,
15
+ "transformers_version": "4.46.3"
16
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2cfaba5e03c22043d16c99710ab693e7ae4ff4c4519214f3ad07c9aab08519b4
3
  size 304869976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c926993dbf175ecbed64967eeb8ebdda0ab4b0b5a00f3517aa5162c5d732d74
3
  size 304869976
runs/Dec08_16-57-00_7a753c7f4b43/events.out.tfevents.1733677025.7a753c7f4b43.1232.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cf8000441574f931f93149222f6be81e311c7c5aa17f606e96115168ed4ab86
3
+ size 6642
runs/Dec08_17-53-57_7a753c7f4b43/events.out.tfevents.1733680438.7a753c7f4b43.15243.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:422568d38e251c6d6242850be1ee17726fe1d22a9b5593146aeb96790e61e18d
3
+ size 6009
runs/Dec08_18-13-25_7a753c7f4b43/events.out.tfevents.1733681605.7a753c7f4b43.15243.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a3725736e3d409627591ede5eb772cece47324f80b04c0c529e62abc0bd3fe3
3
+ size 14474
source.spm ADDED
Binary file (803 kB). View file
 
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "eos_token": "</s>",
3
+ "pad_token": "<pad>",
4
+ "unk_token": "<unk>"
5
+ }
target.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745998e51ba5b058e38b7ac7765c25c43ed5c1c39cc92b27163b9b2e323c9d7c
3
+ size 1080169
tokenizer_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "</s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<unk>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "62517": {
20
+ "content": "<pad>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "clean_up_tokenization_spaces": false,
29
+ "eos_token": "</s>",
30
+ "model_max_length": 512,
31
+ "pad_token": "<pad>",
32
+ "separate_vocabs": false,
33
+ "source_lang": "en",
34
+ "sp_model_kwargs": {},
35
+ "target_lang": "ru",
36
+ "tokenizer_class": "MarianTokenizer",
37
+ "unk_token": "<unk>"
38
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:058e005e81412ac409705b6301157a8a782eca6ebb95da81764b6794e0d1e0e1
3
- size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cd18280c85830717bb00f953cec4950d613df18648140b2943263fc287ff3b9
3
+ size 5304
vocab.json ADDED
The diff for this file is too large to render. See raw diff