davelotito commited on
Commit
82fad32
1 Parent(s): f37ca8b

End of training

Browse files
.ipynb_checkpoints/hyperparameters-checkpoint.yaml ADDED
@@ -0,0 +1,160 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ __cached__setup_devices: !!python/object/apply:torch.device
2
+ - cuda
3
+ - 0
4
+ _n_gpu: 1
5
+ accelerator_config: !!python/object:transformers.trainer_pt_utils.AcceleratorConfig
6
+ dispatch_batches: null
7
+ even_batches: true
8
+ gradient_accumulation_kwargs: null
9
+ split_batches: false
10
+ use_seedable_sampler: true
11
+ adafactor: false
12
+ adam_beta1: 0.9
13
+ adam_beta2: 0.999
14
+ adam_epsilon: 1.0e-08
15
+ auto_find_batch_size: false
16
+ bf16: false
17
+ bf16_full_eval: false
18
+ data_seed: null
19
+ dataloader_drop_last: false
20
+ dataloader_num_workers: 0
21
+ dataloader_persistent_workers: false
22
+ dataloader_pin_memory: true
23
+ dataloader_prefetch_factor: null
24
+ ddp_backend: null
25
+ ddp_broadcast_buffers: null
26
+ ddp_bucket_cap_mb: null
27
+ ddp_find_unused_parameters: null
28
+ ddp_timeout: 1800
29
+ debug: []
30
+ deepspeed: null
31
+ deepspeed_plugin: null
32
+ disable_tqdm: false
33
+ dispatch_batches: null
34
+ distributed_state: !!python/object:accelerate.state.PartialState
35
+ _cpu: false
36
+ backend: null
37
+ debug: false
38
+ device: !!python/object/apply:torch.device
39
+ - cuda
40
+ distributed_type: !!python/object/apply:accelerate.utils.dataclasses.DistributedType
41
+ - 'NO'
42
+ fork_launched: false
43
+ local_process_index: 0
44
+ num_processes: 1
45
+ process_index: 0
46
+ do_eval: true
47
+ do_predict: false
48
+ do_train: false
49
+ eval_accumulation_steps: null
50
+ eval_delay: 0
51
+ eval_do_concat_batches: true
52
+ eval_steps: null
53
+ evaluation_strategy: !!python/object/apply:transformers.trainer_utils.IntervalStrategy
54
+ - epoch
55
+ fp16: true
56
+ fp16_backend: auto
57
+ fp16_full_eval: false
58
+ fp16_opt_level: O1
59
+ fsdp: []
60
+ fsdp_config:
61
+ min_num_params: 0
62
+ xla: false
63
+ xla_fsdp_grad_ckpt: false
64
+ xla_fsdp_v2: false
65
+ fsdp_min_num_params: 0
66
+ fsdp_transformer_layer_cls_to_wrap: null
67
+ full_determinism: false
68
+ generation_config: null
69
+ generation_max_length: null
70
+ generation_num_beams: null
71
+ gradient_accumulation_steps: 2
72
+ gradient_checkpointing: false
73
+ gradient_checkpointing_kwargs: null
74
+ greater_is_better: false
75
+ group_by_length: false
76
+ half_precision_backend: auto
77
+ hub_always_push: false
78
+ hub_model_id: donut_experiment_5
79
+ hub_private_repo: false
80
+ hub_strategy: !!python/object/apply:transformers.trainer_utils.HubStrategy
81
+ - every_save
82
+ hub_token: null
83
+ ignore_data_skip: false
84
+ include_inputs_for_metrics: false
85
+ include_num_input_tokens_seen: false
86
+ include_tokens_per_second: false
87
+ jit_mode_eval: false
88
+ label_names: null
89
+ label_smoothing_factor: 0.0
90
+ learning_rate: 2.0e-05
91
+ length_column_name: length
92
+ load_best_model_at_end: true
93
+ local_rank: 0
94
+ log_level: passive
95
+ log_level_replica: warning
96
+ log_on_each_node: true
97
+ logging_dir: model_runs/donut_experiment_5/runs/May14_14-13-52_ip-172-16-23-1.ec2.internal
98
+ logging_first_step: false
99
+ logging_nan_inf_filter: true
100
+ logging_steps: 100
101
+ logging_strategy: !!python/object/apply:transformers.trainer_utils.IntervalStrategy
102
+ - steps
103
+ lr_scheduler_kwargs: {}
104
+ lr_scheduler_type: !!python/object/apply:transformers.trainer_utils.SchedulerType
105
+ - linear
106
+ max_grad_norm: 1.0
107
+ max_steps: -1
108
+ metric_for_best_model: loss
109
+ mp_parameters: ''
110
+ neftune_noise_alpha: null
111
+ no_cuda: false
112
+ num_train_epochs: 4
113
+ optim: !!python/object/apply:transformers.training_args.OptimizerNames
114
+ - adamw_torch
115
+ optim_args: null
116
+ optim_target_modules: null
117
+ output_dir: model_runs/donut_experiment_5
118
+ overwrite_output_dir: false
119
+ past_index: -1
120
+ per_device_eval_batch_size: 1
121
+ per_device_train_batch_size: 1
122
+ per_gpu_eval_batch_size: null
123
+ per_gpu_train_batch_size: null
124
+ predict_with_generate: true
125
+ prediction_loss_only: false
126
+ push_to_hub: true
127
+ push_to_hub_model_id: null
128
+ push_to_hub_organization: null
129
+ push_to_hub_token: null
130
+ ray_scope: last
131
+ remove_unused_columns: true
132
+ report_to:
133
+ - tensorboard
134
+ resume_from_checkpoint: null
135
+ run_name: model_runs/donut_experiment_5
136
+ save_on_each_node: false
137
+ save_only_model: false
138
+ save_safetensors: true
139
+ save_steps: 500
140
+ save_strategy: !!python/object/apply:transformers.trainer_utils.IntervalStrategy
141
+ - epoch
142
+ save_total_limit: 2
143
+ seed: 42
144
+ skip_memory_metrics: true
145
+ sortish_sampler: false
146
+ split_batches: null
147
+ tf32: null
148
+ torch_compile: false
149
+ torch_compile_backend: null
150
+ torch_compile_mode: null
151
+ torchdynamo: null
152
+ tpu_metrics_debug: false
153
+ tpu_num_cores: null
154
+ use_cpu: false
155
+ use_ipex: false
156
+ use_legacy_prediction_loop: false
157
+ use_mps_device: false
158
+ warmup_ratio: 0.0
159
+ warmup_steps: 0
160
+ weight_decay: 0.01
README.md ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: naver-clova-ix/donut-base
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - bleu
8
+ - wer
9
+ model-index:
10
+ - name: donut_experiment_5
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # donut_experiment_5
18
+
19
+ This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the None dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.3987
22
+ - Bleu: 0.0661
23
+ - Precisions: [0.8020833333333334, 0.7375886524822695, 0.6994535519125683, 0.6601941747572816]
24
+ - Brevity Penalty: 0.0915
25
+ - Length Ratio: 0.2948
26
+ - Translation Length: 480
27
+ - Reference Length: 1628
28
+ - Cer: 0.7576
29
+ - Wer: 0.8280
30
+
31
+ ## Model description
32
+
33
+ More information needed
34
+
35
+ ## Intended uses & limitations
36
+
37
+ More information needed
38
+
39
+ ## Training and evaluation data
40
+
41
+ More information needed
42
+
43
+ ## Training procedure
44
+
45
+ ### Training hyperparameters
46
+
47
+ The following hyperparameters were used during training:
48
+ - learning_rate: 2e-05
49
+ - train_batch_size: 1
50
+ - eval_batch_size: 1
51
+ - seed: 42
52
+ - gradient_accumulation_steps: 2
53
+ - total_train_batch_size: 2
54
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
+ - lr_scheduler_type: linear
56
+ - num_epochs: 4
57
+ - mixed_precision_training: Native AMP
58
+
59
+ ### Training results
60
+
61
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Cer | Wer |
62
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:--------------------------------------------------------------------------------:|:---------------:|:------------:|:------------------:|:----------------:|:------:|:------:|
63
+ | 0.3274 | 1.0 | 253 | 0.4698 | 0.0586 | [0.7707006369426752, 0.6956521739130435, 0.6582633053221288, 0.62] | 0.0857 | 0.2893 | 471 | 1628 | 0.7660 | 0.8432 |
64
+ | 0.2539 | 2.0 | 506 | 0.4198 | 0.0643 | [0.799163179916318, 0.7315914489311164, 0.6868131868131868, 0.6416938110749185] | 0.0902 | 0.2936 | 478 | 1628 | 0.7605 | 0.8313 |
65
+ | 0.224 | 3.0 | 759 | 0.3941 | 0.0658 | [0.8075313807531381, 0.7387173396674585, 0.7060439560439561, 0.6710097719869706] | 0.0902 | 0.2936 | 478 | 1628 | 0.7573 | 0.8283 |
66
+ | 0.1566 | 4.0 | 1012 | 0.3987 | 0.0661 | [0.8020833333333334, 0.7375886524822695, 0.6994535519125683, 0.6601941747572816] | 0.0915 | 0.2948 | 480 | 1628 | 0.7576 | 0.8280 |
67
+
68
+
69
+ ### Framework versions
70
+
71
+ - Transformers 4.40.0
72
+ - Pytorch 2.1.0
73
+ - Datasets 2.18.0
74
+ - Tokenizers 0.19.1
added_tokens.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</s_address>": 57532,
3
+ "</s_company>": 57530,
4
+ "</s_date>": 57528,
5
+ "</s_total>": 57526,
6
+ "<s_address>": 57531,
7
+ "<s_company>": 57529,
8
+ "<s_date>": 57527,
9
+ "<s_iitcdip>": 57523,
10
+ "<s_synthdog>": 57524,
11
+ "<s_total>": 57525,
12
+ "<sep/>": 57522
13
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 0,
3
+ "eos_token_id": 2,
4
+ "forced_eos_token_id": 2,
5
+ "pad_token_id": 1,
6
+ "transformers_version": "4.40.0"
7
+ }
metrics.jsonl CHANGED
@@ -1,3 +1,4 @@
1
  {"eval_loss": 0.46984946727752686, "eval_bleu": 0.058638588671502775, "eval_precisions": [0.7707006369426752, 0.6956521739130435, 0.6582633053221288, 0.62], "eval_brevity_penalty": 0.08573659052650866, "eval_length_ratio": 0.2893120393120393, "eval_translation_length": 471, "eval_reference_length": 1628, "eval_cer": 0.7660104054123782, "eval_wer": 0.8431764116699929, "eval_runtime": 75.9302, "eval_samples_per_second": 0.751, "eval_steps_per_second": 0.751, "epoch": 1.0}
2
  {"eval_loss": 0.4197552800178528, "eval_bleu": 0.06425647536525644, "eval_precisions": [0.799163179916318, 0.7315914489311164, 0.6868131868131868, 0.6416938110749185], "eval_brevity_penalty": 0.0901881044262292, "eval_length_ratio": 0.2936117936117936, "eval_translation_length": 478, "eval_reference_length": 1628, "eval_cer": 0.7604681276252823, "eval_wer": 0.8312649598507473, "eval_runtime": 75.4245, "eval_samples_per_second": 0.756, "eval_steps_per_second": 0.756, "epoch": 2.0}
3
  {"eval_loss": 0.39413073658943176, "eval_bleu": 0.06575803552373187, "eval_precisions": [0.8075313807531381, 0.7387173396674585, 0.7060439560439561, 0.6710097719869706], "eval_brevity_penalty": 0.0901881044262292, "eval_length_ratio": 0.2936117936117936, "eval_translation_length": 478, "eval_reference_length": 1628, "eval_cer": 0.757338139839836, "eval_wer": 0.8282692162386603, "eval_runtime": 75.0128, "eval_samples_per_second": 0.76, "eval_steps_per_second": 0.76, "epoch": 3.0}
 
 
1
  {"eval_loss": 0.46984946727752686, "eval_bleu": 0.058638588671502775, "eval_precisions": [0.7707006369426752, 0.6956521739130435, 0.6582633053221288, 0.62], "eval_brevity_penalty": 0.08573659052650866, "eval_length_ratio": 0.2893120393120393, "eval_translation_length": 471, "eval_reference_length": 1628, "eval_cer": 0.7660104054123782, "eval_wer": 0.8431764116699929, "eval_runtime": 75.9302, "eval_samples_per_second": 0.751, "eval_steps_per_second": 0.751, "epoch": 1.0}
2
  {"eval_loss": 0.4197552800178528, "eval_bleu": 0.06425647536525644, "eval_precisions": [0.799163179916318, 0.7315914489311164, 0.6868131868131868, 0.6416938110749185], "eval_brevity_penalty": 0.0901881044262292, "eval_length_ratio": 0.2936117936117936, "eval_translation_length": 478, "eval_reference_length": 1628, "eval_cer": 0.7604681276252823, "eval_wer": 0.8312649598507473, "eval_runtime": 75.4245, "eval_samples_per_second": 0.756, "eval_steps_per_second": 0.756, "epoch": 2.0}
3
  {"eval_loss": 0.39413073658943176, "eval_bleu": 0.06575803552373187, "eval_precisions": [0.8075313807531381, 0.7387173396674585, 0.7060439560439561, 0.6710097719869706], "eval_brevity_penalty": 0.0901881044262292, "eval_length_ratio": 0.2936117936117936, "eval_translation_length": 478, "eval_reference_length": 1628, "eval_cer": 0.757338139839836, "eval_wer": 0.8282692162386603, "eval_runtime": 75.0128, "eval_samples_per_second": 0.76, "eval_steps_per_second": 0.76, "epoch": 3.0}
4
+ {"eval_loss": 0.39874938130378723, "eval_bleu": 0.06613455502080588, "eval_precisions": [0.8020833333333334, 0.7375886524822695, 0.6994535519125683, 0.6601941747572816], "eval_brevity_penalty": 0.09147709493048638, "eval_length_ratio": 0.29484029484029484, "eval_translation_length": 480, "eval_reference_length": 1628, "eval_cer": 0.7576375295447421, "eval_wer": 0.8280149631962583, "eval_runtime": 75.4844, "eval_samples_per_second": 0.755, "eval_steps_per_second": 0.755, "epoch": 4.0}
preprocessor_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_thumbnail",
8
+ "do_align_long_axis",
9
+ "do_pad",
10
+ "random_padding",
11
+ "do_rescale",
12
+ "rescale_factor",
13
+ "do_normalize",
14
+ "image_mean",
15
+ "image_std",
16
+ "return_tensors",
17
+ "data_format",
18
+ "input_data_format"
19
+ ],
20
+ "do_align_long_axis": false,
21
+ "do_normalize": true,
22
+ "do_pad": true,
23
+ "do_rescale": true,
24
+ "do_resize": true,
25
+ "do_thumbnail": true,
26
+ "image_mean": [
27
+ 0.5,
28
+ 0.5,
29
+ 0.5
30
+ ],
31
+ "image_processor_type": "DonutImageProcessor",
32
+ "image_std": [
33
+ 0.5,
34
+ 0.5,
35
+ 0.5
36
+ ],
37
+ "processor_class": "DonutProcessor",
38
+ "resample": 2,
39
+ "rescale_factor": 0.00392156862745098,
40
+ "size": [
41
+ 720,
42
+ 960
43
+ ]
44
+ }
runs/May14_14-13-52_ip-172-16-23-1.ec2.internal/events.out.tfevents.1715696032.ip-172-16-23-1.ec2.internal.9483.1 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0b5643fea30751143a3ae7226cecbc575c888f251d32abc4b02d6208cbffeb4
3
- size 12558
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9061dc95c2de5cbbb5a34d446b653ba2f5061f26bb18564a576b8c654204120
3
+ size 14195
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb9e3dce4c326195d08fc3dd0f7e2eee1da8595c847bf4c1a9c78b7a82d47e2d
3
+ size 1296245
special_tokens_map.json ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<s_total>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "</s_total>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<s_date>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "</s_date>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "<s_company>",
33
+ "lstrip": false,
34
+ "normalized": false,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "</s_company>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ },
45
+ {
46
+ "content": "<s_address>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false
51
+ },
52
+ {
53
+ "content": "</s_address>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false
58
+ },
59
+ {
60
+ "content": "<s>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false
65
+ },
66
+ {
67
+ "content": "</s>",
68
+ "lstrip": false,
69
+ "normalized": false,
70
+ "rstrip": false,
71
+ "single_word": false
72
+ }
73
+ ],
74
+ "bos_token": {
75
+ "content": "<s>",
76
+ "lstrip": false,
77
+ "normalized": false,
78
+ "rstrip": false,
79
+ "single_word": false
80
+ },
81
+ "cls_token": {
82
+ "content": "<s>",
83
+ "lstrip": false,
84
+ "normalized": false,
85
+ "rstrip": false,
86
+ "single_word": false
87
+ },
88
+ "eos_token": {
89
+ "content": "</s>",
90
+ "lstrip": false,
91
+ "normalized": false,
92
+ "rstrip": false,
93
+ "single_word": false
94
+ },
95
+ "mask_token": {
96
+ "content": "<mask>",
97
+ "lstrip": true,
98
+ "normalized": true,
99
+ "rstrip": false,
100
+ "single_word": false
101
+ },
102
+ "pad_token": {
103
+ "content": "<pad>",
104
+ "lstrip": false,
105
+ "normalized": false,
106
+ "rstrip": false,
107
+ "single_word": false
108
+ },
109
+ "sep_token": {
110
+ "content": "</s>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false
115
+ },
116
+ "unk_token": {
117
+ "content": "<unk>",
118
+ "lstrip": false,
119
+ "normalized": false,
120
+ "rstrip": false,
121
+ "single_word": false
122
+ }
123
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,156 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "57521": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": true,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "57522": {
44
+ "content": "<sep/>",
45
+ "lstrip": false,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": false
50
+ },
51
+ "57523": {
52
+ "content": "<s_iitcdip>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "57524": {
60
+ "content": "<s_synthdog>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "57525": {
68
+ "content": "<s_total>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "57526": {
76
+ "content": "</s_total>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "57527": {
84
+ "content": "<s_date>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "57528": {
92
+ "content": "</s_date>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "57529": {
100
+ "content": "<s_company>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "57530": {
108
+ "content": "</s_company>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "57531": {
116
+ "content": "<s_address>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "57532": {
124
+ "content": "</s_address>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ }
131
+ },
132
+ "additional_special_tokens": [
133
+ "<s_total>",
134
+ "</s_total>",
135
+ "<s_date>",
136
+ "</s_date>",
137
+ "<s_company>",
138
+ "</s_company>",
139
+ "<s_address>",
140
+ "</s_address>",
141
+ "<s>",
142
+ "</s>"
143
+ ],
144
+ "bos_token": "<s>",
145
+ "clean_up_tokenization_spaces": true,
146
+ "cls_token": "<s>",
147
+ "eos_token": "</s>",
148
+ "mask_token": "<mask>",
149
+ "model_max_length": 1000000000000000019884624838656,
150
+ "pad_token": "<pad>",
151
+ "processor_class": "DonutProcessor",
152
+ "sep_token": "</s>",
153
+ "sp_model_kwargs": {},
154
+ "tokenizer_class": "XLMRobertaTokenizer",
155
+ "unk_token": "<unk>"
156
+ }