Upload configs.yaml with huggingface_hub
Browse files- configs.yaml +2 -2
configs.yaml
CHANGED
@@ -14,12 +14,12 @@ global_batch_size: 512
|
|
14 |
gradient_accumulation_steps: 8
|
15 |
gradient_checkpointing: true
|
16 |
hub_model_id: mlfoundations-dev/open-o1-sft-original
|
17 |
-
include_hp: dcft/train/hp_settings/
|
18 |
learning_rate: 5.0e-06
|
19 |
logging_steps: 10
|
20 |
lr_scheduler_type: constant
|
21 |
max_grad_norm: 1
|
22 |
-
model_name_or_path: meta-llama/
|
23 |
neat_packing: true
|
24 |
num_train_epochs: 3.0
|
25 |
output_dir: ./experiments/train/checkpoints/open_o1_sft_original
|
|
|
14 |
gradient_accumulation_steps: 8
|
15 |
gradient_checkpointing: true
|
16 |
hub_model_id: mlfoundations-dev/open-o1-sft-original
|
17 |
+
include_hp: dcft/train/hp_settings/hritik_no_model.yaml
|
18 |
learning_rate: 5.0e-06
|
19 |
logging_steps: 10
|
20 |
lr_scheduler_type: constant
|
21 |
max_grad_norm: 1
|
22 |
+
model_name_or_path: meta-llama/Llama-3.1-8B
|
23 |
neat_packing: true
|
24 |
num_train_epochs: 3.0
|
25 |
output_dir: ./experiments/train/checkpoints/open_o1_sft_original
|