mikeam commited on
Commit
403343c
·
verified ·
1 Parent(s): 0231348

Upload configs.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. configs.yaml +2 -2
configs.yaml CHANGED
@@ -14,12 +14,12 @@ global_batch_size: 512
14
  gradient_accumulation_steps: 8
15
  gradient_checkpointing: true
16
  hub_model_id: mlfoundations-dev/open-o1-sft-original
17
- include_hp: dcft/train/hp_settings/hritik.yaml
18
  learning_rate: 5.0e-06
19
  logging_steps: 10
20
  lr_scheduler_type: constant
21
  max_grad_norm: 1
22
- model_name_or_path: meta-llama/Meta-Llama-3.1-8B
23
  neat_packing: true
24
  num_train_epochs: 3.0
25
  output_dir: ./experiments/train/checkpoints/open_o1_sft_original
 
14
  gradient_accumulation_steps: 8
15
  gradient_checkpointing: true
16
  hub_model_id: mlfoundations-dev/open-o1-sft-original
17
+ include_hp: dcft/train/hp_settings/hritik_no_model.yaml
18
  learning_rate: 5.0e-06
19
  logging_steps: 10
20
  lr_scheduler_type: constant
21
  max_grad_norm: 1
22
+ model_name_or_path: meta-llama/Llama-3.1-8B
23
  neat_packing: true
24
  num_train_epochs: 3.0
25
  output_dir: ./experiments/train/checkpoints/open_o1_sft_original