nlparabic commited on
Commit
d0a617a
·
verified ·
1 Parent(s): 55774b2

Training in progress, epoch 11

Browse files
egy_training_log.txt CHANGED
@@ -461,3 +461,148 @@ INFO:root:Epoch 10.0: Train Loss = 0.2719, Eval Loss = 0.6550981998443604
461
  INFO:absl:Using default tokenizer.
462
  INFO:__main__:*** Evaluate ***
463
  INFO:absl:Using default tokenizer.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
461
  INFO:absl:Using default tokenizer.
462
  INFO:__main__:*** Evaluate ***
463
  INFO:absl:Using default tokenizer.
464
+ WARNING:__main__:Process rank: 0, device: cuda:0, n_gpu: 1, distributed training: False, 16-bits training: False
465
+ INFO:__main__:Training/evaluation parameters TrainingArguments(
466
+ _n_gpu=1,
467
+ accelerator_config={'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None, 'use_configured_state': False},
468
+ adafactor=False,
469
+ adam_beta1=0.9,
470
+ adam_beta2=0.999,
471
+ adam_epsilon=1e-08,
472
+ auto_find_batch_size=False,
473
+ batch_eval_metrics=False,
474
+ bf16=False,
475
+ bf16_full_eval=False,
476
+ data_seed=None,
477
+ dataloader_drop_last=False,
478
+ dataloader_num_workers=0,
479
+ dataloader_persistent_workers=False,
480
+ dataloader_pin_memory=True,
481
+ dataloader_prefetch_factor=None,
482
+ ddp_backend=None,
483
+ ddp_broadcast_buffers=None,
484
+ ddp_bucket_cap_mb=None,
485
+ ddp_find_unused_parameters=None,
486
+ ddp_timeout=1800,
487
+ debug=[],
488
+ deepspeed=None,
489
+ disable_tqdm=False,
490
+ dispatch_batches=None,
491
+ do_eval=True,
492
+ do_predict=False,
493
+ do_train=True,
494
+ eval_accumulation_steps=None,
495
+ eval_delay=0,
496
+ eval_do_concat_batches=True,
497
+ eval_on_start=False,
498
+ eval_steps=None,
499
+ eval_strategy=IntervalStrategy.EPOCH,
500
+ eval_use_gather_object=False,
501
+ evaluation_strategy=epoch,
502
+ fp16=False,
503
+ fp16_backend=auto,
504
+ fp16_full_eval=False,
505
+ fp16_opt_level=O1,
506
+ fsdp=[],
507
+ fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False},
508
+ fsdp_min_num_params=0,
509
+ fsdp_transformer_layer_cls_to_wrap=None,
510
+ full_determinism=False,
511
+ gradient_accumulation_steps=1,
512
+ gradient_checkpointing=False,
513
+ gradient_checkpointing_kwargs=None,
514
+ greater_is_better=False,
515
+ group_by_length=False,
516
+ half_precision_backend=auto,
517
+ hub_always_push=False,
518
+ hub_model_id=None,
519
+ hub_private_repo=False,
520
+ hub_strategy=HubStrategy.EVERY_SAVE,
521
+ hub_token=<HUB_TOKEN>,
522
+ ignore_data_skip=False,
523
+ include_inputs_for_metrics=False,
524
+ include_num_input_tokens_seen=False,
525
+ include_tokens_per_second=False,
526
+ jit_mode_eval=False,
527
+ label_names=None,
528
+ label_smoothing_factor=0.0,
529
+ learning_rate=5e-05,
530
+ length_column_name=length,
531
+ load_best_model_at_end=True,
532
+ local_rank=0,
533
+ log_level=passive,
534
+ log_level_replica=warning,
535
+ log_on_each_node=True,
536
+ logging_dir=/home/iais_marenpielka/Bouthaina/res_nw_dj/runs/Sep01_08-31-45_lmgpu-node-09,
537
+ logging_first_step=False,
538
+ logging_nan_inf_filter=True,
539
+ logging_steps=500,
540
+ logging_strategy=IntervalStrategy.EPOCH,
541
+ lr_scheduler_kwargs={},
542
+ lr_scheduler_type=SchedulerType.LINEAR,
543
+ max_grad_norm=1.0,
544
+ max_steps=-1,
545
+ metric_for_best_model=loss,
546
+ mp_parameters=,
547
+ neftune_noise_alpha=None,
548
+ no_cuda=False,
549
+ num_train_epochs=20.0,
550
+ optim=OptimizerNames.ADAMW_TORCH,
551
+ optim_args=None,
552
+ optim_target_modules=None,
553
+ output_dir=/home/iais_marenpielka/Bouthaina/res_nw_dj,
554
+ overwrite_output_dir=False,
555
+ past_index=-1,
556
+ per_device_eval_batch_size=8,
557
+ per_device_train_batch_size=8,
558
+ prediction_loss_only=False,
559
+ push_to_hub=True,
560
+ push_to_hub_model_id=None,
561
+ push_to_hub_organization=None,
562
+ push_to_hub_token=<PUSH_TO_HUB_TOKEN>,
563
+ ray_scope=last,
564
+ remove_unused_columns=True,
565
+ report_to=[],
566
+ restore_callback_states_from_checkpoint=False,
567
+ resume_from_checkpoint=None,
568
+ run_name=/home/iais_marenpielka/Bouthaina/res_nw_dj,
569
+ save_on_each_node=False,
570
+ save_only_model=False,
571
+ save_safetensors=True,
572
+ save_steps=500,
573
+ save_strategy=IntervalStrategy.EPOCH,
574
+ save_total_limit=None,
575
+ seed=42,
576
+ skip_memory_metrics=True,
577
+ split_batches=None,
578
+ tf32=None,
579
+ torch_compile=False,
580
+ torch_compile_backend=None,
581
+ torch_compile_mode=None,
582
+ torch_empty_cache_steps=None,
583
+ torchdynamo=None,
584
+ tpu_metrics_debug=False,
585
+ tpu_num_cores=None,
586
+ use_cpu=False,
587
+ use_ipex=False,
588
+ use_legacy_prediction_loop=False,
589
+ use_mps_device=False,
590
+ warmup_ratio=0.0,
591
+ warmup_steps=500,
592
+ weight_decay=0.0,
593
+ )
594
+ INFO:__main__:Checkpoint detected, resuming training at /home/iais_marenpielka/Bouthaina/res_nw_dj/checkpoint-27030. To avoid this behavior, change the `--output_dir` or add `--overwrite_output_dir` to train from scratch.
595
+ INFO:datasets.builder:Using custom data configuration default-98487e126fdb56c7
596
+ INFO:datasets.info:Loading Dataset Infos from /home/iais_marenpielka/Bouthaina/miniconda3/lib/python3.12/site-packages/datasets/packaged_modules/text
597
+ INFO:datasets.builder:Overwrite dataset info from restored data version if exists.
598
+ INFO:datasets.info:Loading Dataset info from /home/iais_marenpielka/.cache/huggingface/datasets/text/default-98487e126fdb56c7/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101
599
+ INFO:datasets.builder:Found cached dataset text (/home/iais_marenpielka/.cache/huggingface/datasets/text/default-98487e126fdb56c7/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101)
600
+ INFO:datasets.info:Loading Dataset info from /home/iais_marenpielka/.cache/huggingface/datasets/text/default-98487e126fdb56c7/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101
601
+ INFO:datasets.arrow_dataset:Loading cached processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-98487e126fdb56c7/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-22cb64ac8b531f65.arrow
602
+ INFO:datasets.arrow_dataset:Loading cached processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-98487e126fdb56c7/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-7840ce8488671e31.arrow
603
+ WARNING:__main__:The tokenizer picked seems to have a very large `model_max_length` (1000000000000000019884624838656). Using block_size=768 instead. You can change that default value by passing --block_size xxx.
604
+ INFO:datasets.arrow_dataset:Loading cached processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-98487e126fdb56c7/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-1f7e65ff1cef4012.arrow
605
+ INFO:datasets.arrow_dataset:Loading cached processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-98487e126fdb56c7/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-cb43c016a75212ff.arrow
606
+ WARNING:accelerate.utils.other:Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
607
+ INFO:root:Epoch 11.0: Train Loss = 0.2396, Eval Loss = 0.665830671787262
608
+ INFO:absl:Using default tokenizer.
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bfe3c622bc3f36c295706ec7e3fd153d28f79268fcf6e11f09f6a77f69068610
3
  size 539221632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1089e2fa684b1e1cde4ceef16bd0dde6db830812d17cd299b429a8d602f3a51
3
  size 539221632
special_tokens_map.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "additional_special_tokens": [
3
  {
4
- "content": "[s]",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
 
1
  {
2
  "additional_special_tokens": [
3
  {
4
+ "content": "<s>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
tokenizer.json CHANGED
@@ -46,7 +46,7 @@
46
  },
47
  {
48
  "id": 64002,
49
- "content": "[s]",
50
  "single_word": false,
51
  "lstrip": false,
52
  "rstrip": false,
 
46
  },
47
  {
48
  "id": 64002,
49
+ "content": "<s>",
50
  "single_word": false,
51
  "lstrip": false,
52
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -25,7 +25,7 @@
25
  "special": true
26
  },
27
  "64002": {
28
- "content": "[s]",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
@@ -50,7 +50,7 @@
50
  }
51
  },
52
  "additional_special_tokens": [
53
- "[s]",
54
  "</s>",
55
  "[sep]"
56
  ],
 
25
  "special": true
26
  },
27
  "64002": {
28
+ "content": "<s>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
 
50
  }
51
  },
52
  "additional_special_tokens": [
53
+ "<s>",
54
  "</s>",
55
  "[sep]"
56
  ],
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b0800034678b076a1ebdb11a754f981ea0244f0189ff323f9e1e43722aa3e4c
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f941b19a4bdd74821701732fc2412174ac26cea61752479531dfffb1551ec828
3
  size 5240