pere commited on
Commit
3e8bdbf
•
1 Parent(s): ae91c97

Training in progress, step 250

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1d08c9740a6b886a8f3440244ee7d894c555c8ee7c64898e14fb64ab76ed829
3
  size 1262104049
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c56d8171c7e99a5c2a9e54c0c4428bdc8901de1a9fc1f2248238850b2cae028f
3
  size 1262104049
run.sh CHANGED
@@ -8,17 +8,17 @@ python run_speech_recognition_ctc_bnb.py \
8
  --output_dir="./" \
9
  --overwrite_output_dir \
10
  --hub_model_id="NbAiLab/wav2vec2-xlsr-300m-norwegian2" \
11
- --num_train_epochs="50" \
12
- --per_device_train_batch_size="16" \
13
- --per_device_eval_batch_size="16" \
14
  --gradient_accumulation_steps="4" \
15
  --learning_rate="7.5e-5" \
16
  --warmup_steps="2000" \
17
  --evaluation_strategy="steps" \
18
  --length_column_name="input_length" \
19
  --chars_to_ignore , ? . ! \- \; \: \" " % ‘ " � — ’ … – \
20
- --save_steps="100" \
21
- --eval_steps="100" \
22
  --logging_steps="100" \
23
  --layerdrop="0.0" \
24
  --activation_dropout="0.1" \
 
8
  --output_dir="./" \
9
  --overwrite_output_dir \
10
  --hub_model_id="NbAiLab/wav2vec2-xlsr-300m-norwegian2" \
11
+ --num_train_epochs="100" \
12
+ --per_device_train_batch_size="8" \
13
+ --per_device_eval_batch_size="8" \
14
  --gradient_accumulation_steps="4" \
15
  --learning_rate="7.5e-5" \
16
  --warmup_steps="2000" \
17
  --evaluation_strategy="steps" \
18
  --length_column_name="input_length" \
19
  --chars_to_ignore , ? . ! \- \; \: \" " % ‘ " � — ’ … – \
20
+ --save_steps="250" \
21
+ --eval_steps="250" \
22
  --logging_steps="100" \
23
  --layerdrop="0.0" \
24
  --activation_dropout="0.1" \
run_speech_recognition_ctc_bnb.py CHANGED
@@ -686,25 +686,25 @@ def main():
686
  # Instantiate custom data collator
687
  data_collator = DataCollatorCTCWithPadding(processor=processor)
688
 
689
- decay_parameters = get_parameter_names(model, [torch.nn.LayerNorm])
690
- decay_parameters = [name for name in decay_parameters if "bias" not in name]
691
- optimizer_grouped_parameters = [
692
- {
693
- "params": [p for n, p in model.named_parameters() if n in decay_parameters],
694
- "weight_decay": training_args.weight_decay,
695
- },
696
- {
697
- "params": [p for n, p in model.named_parameters() if n not in decay_parameters],
698
- "weight_decay": 0.0,
699
- },
700
- ]
701
- optimizer = bnb.optim.Adam8bit(
702
- params=optimizer_grouped_parameters,
703
- betas=(training_args.adam_beta1, training_args.adam_beta2),
704
- eps=training_args.adam_epsilon,
705
- )
706
-
707
- optimizers = (optimizer, None)
708
 
709
  # Initialize Trainer
710
  trainer = Trainer(
@@ -715,8 +715,8 @@ def main():
715
  train_dataset=vectorized_datasets["train"] if training_args.do_train else None,
716
  eval_dataset=vectorized_datasets["eval"] if training_args.do_eval else None,
717
  tokenizer=feature_extractor,
718
- optimizers=optimizers,
719
- )
720
 
721
  # 8. Finally, we can start training
722
 
 
686
  # Instantiate custom data collator
687
  data_collator = DataCollatorCTCWithPadding(processor=processor)
688
 
689
+ #decay_parameters = get_parameter_names(model, [torch.nn.LayerNorm])
690
+ #decay_parameters = [name for name in decay_parameters if "bias" not in name]
691
+ #optimizer_grouped_parameters = [
692
+ # {
693
+ # "params": [p for n, p in model.named_parameters() if n in decay_parameters],
694
+ # "weight_decay": training_args.weight_decay,
695
+ # },
696
+ # {
697
+ # "params": [p for n, p in model.named_parameters() if n not in decay_parameters],
698
+ # "weight_decay": 0.0,
699
+ # },
700
+ #]
701
+ #optimizer = bnb.optim.Adam8bit(
702
+ # params=optimizer_grouped_parameters,
703
+ # betas=(training_args.adam_beta1, training_args.adam_beta2),
704
+ # eps=training_args.adam_epsilon,
705
+ #)
706
+
707
+ #optimizers = (optimizer, None)
708
 
709
  # Initialize Trainer
710
  trainer = Trainer(
 
715
  train_dataset=vectorized_datasets["train"] if training_args.do_train else None,
716
  eval_dataset=vectorized_datasets["eval"] if training_args.do_eval else None,
717
  tokenizer=feature_extractor,
718
+ #optimizers=optimizers,
719
+ )
720
 
721
  # 8. Finally, we can start training
722
 
runs/Jan28_13-50-07_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/events.out.tfevents.1643378396.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.124745.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:969c3f5031c2e78d325e0766501a0f980394cd5e2824976618fcda868f808e28
3
- size 16585
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2abb1d86b00c8a06e558229507420c951af09afb063ee92e8eb4f72c8b2a4d9c
3
+ size 18010
runs/Jan28_19-16-02_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/1643397461.831888/events.out.tfevents.1643397461.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.261029.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22295f46dbc159d371cb3f6f7811ecb0a43041a21bd080154de37696780d1096
3
+ size 4786
runs/Jan28_19-16-02_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/events.out.tfevents.1643397461.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.261029.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9193521f5ce2e3c7e6b144d8d2e145fe762c0966a69c1e47edb42c648ad67bc5
3
+ size 5347
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b0526ca3463f8ec74dc67da7620f73cd644133f9228ca70552a449cd4c1f1ee
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:249302cc973fa88a60488bebc95f0dde19627b173724b2e0910323ff0a12fe01
3
  size 3055