File size: 11,605 Bytes
7718d64
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f25a924
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
2022-01-30 19:19:54,899 INFO    MainThread:26084 [wandb_setup.py:_flush():71] setting env: {}
2022-01-30 19:19:54,899 INFO    MainThread:26084 [wandb_setup.py:_flush():71] setting login settings: {}
2022-01-30 19:19:54,900 INFO    MainThread:26084 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220130_191954-1mtmnz5y/logs/debug.log
2022-01-30 19:19:54,900 INFO    MainThread:26084 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220130_191954-1mtmnz5y/logs/debug-internal.log
2022-01-30 19:19:54,900 INFO    MainThread:26084 [wandb_init.py:init():404] calling init triggers
2022-01-30 19:19:54,900 INFO    MainThread:26084 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-01-30 19:19:54,900 INFO    MainThread:26084 [wandb_init.py:init():460] starting backend
2022-01-30 19:19:54,901 INFO    MainThread:26084 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-01-30 19:19:54,987 INFO    MainThread:26084 [backend.py:ensure_launched():216] starting backend process...
2022-01-30 19:19:55,070 INFO    MainThread:26084 [backend.py:ensure_launched():221] started backend process with pid: 55777
2022-01-30 19:19:55,073 INFO    MainThread:26084 [wandb_init.py:init():469] backend started and connected
2022-01-30 19:19:55,082 INFO    MainThread:26084 [wandb_init.py:init():533] updated telemetry
2022-01-30 19:19:55,252 INFO    MainThread:26084 [wandb_init.py:init():563] communicating current version
2022-01-30 19:19:55,947 INFO    MainThread:26084 [wandb_init.py:init():568] got version response 
2022-01-30 19:19:55,948 INFO    MainThread:26084 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-01-30 19:19:56,146 INFO    MainThread:26084 [wandb_init.py:init():606] starting run threads in backend
2022-01-30 19:19:56,763 INFO    MainThread:26084 [wandb_run.py:_console_start():1810] atexit reg
2022-01-30 19:19:56,764 INFO    MainThread:26084 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-01-30 19:19:56,764 INFO    MainThread:26084 [wandb_run.py:_redirect():1689] Redirecting console.
2022-01-30 19:19:56,771 INFO    MainThread:26084 [wandb_run.py:_redirect():1745] Redirects installed.
2022-01-30 19:19:56,771 INFO    MainThread:26084 [wandb_init.py:init():633] run started, returning control to user process
2022-01-30 19:19:56,774 INFO    MainThread:26084 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 218, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 219, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.6, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2700, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jan30_18-03-27_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
2022-01-30 19:19:56,778 INFO    MainThread:26084 [wandb_watch.py:watch():43] Watching
2022-02-01 02:46:57,430 INFO    MainThread:26084 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 1
2022-02-01 02:46:57,433 INFO    MainThread:26084 [wandb_run.py:_restore():1752] restore
2022-02-01 02:46:59,849 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2180
  total_bytes: 2180
}

2022-02-01 02:47:00,014 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2180
  total_bytes: 2180
}

2022-02-01 02:47:00,820 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2180
  total_bytes: 2180
}

2022-02-01 02:47:01,511 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 2
}
pusher_stats {
  uploaded_bytes: 2180
  total_bytes: 2925431
}

2022-02-01 02:47:01,614 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 2180
  total_bytes: 3863613
}

2022-02-01 02:47:01,725 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 2180
  total_bytes: 3863613
}

2022-02-01 02:47:01,828 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:01,932 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:02,035 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:02,137 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:02,241 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:02,344 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:02,446 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:02,548 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:03,643 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}

2022-02-01 02:47:04,050 INFO    MainThread:26084 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 3863613
  total_bytes: 3863613
}
local_info {
}

2022-02-01 02:47:05,238 INFO    MainThread:26084 [wandb_run.py:_append_history():2130] rendering history
2022-02-01 02:47:05,240 INFO    MainThread:26084 [wandb_run.py:_append_summary():2085] rendering summary
2022-02-01 02:47:05,242 INFO    MainThread:26084 [wandb_run.py:_append_files():2180] logging synced files