Delta-Vector commited on
Commit
47eda06
·
verified ·
1 Parent(s): 027f814

Upload picaro.yml

Browse files
Files changed (1) hide show
  1. picaro.yml +99 -0
picaro.yml ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: meta-llama/Llama-3.3-70B-Instruct
2
+ base_model_ignore_patterns: "*/*"
3
+ model_type: AutoModelForCausalLM
4
+ tokenizer_type: AutoTokenizer
5
+
6
+ hub_model_id: NewEden/BigPicaro
7
+ hub_strategy: "all_checkpoints"
8
+ push_dataset_to_hub:
9
+ hf_use_auth_token: true
10
+
11
+ load_in_8bit: false
12
+ load_in_4bit: false
13
+ strict: false
14
+
15
+ datasets:
16
+ - path: NewEden/Picaro
17
+ type: chat_template
18
+ roles_to_train: ["gpt"]
19
+ field_messages: conversations
20
+ message_field_role: from
21
+ message_field_content: value
22
+ train_on_eos: turn
23
+
24
+ shuffle_merged_datasets: true
25
+ dataset_prepared_path: ./prepared_data
26
+ val_set_size: 0.0
27
+ output_dir: ./output
28
+
29
+ plugins:
30
+ - axolotl.integrations.liger.LigerPlugin
31
+ - axolotl.integrations.cut_cross_entropy.CutCrossEntropyPlugin
32
+ liger_rope: true
33
+ liger_rms_norm: true
34
+ liger_layer_norm: true
35
+ liger_glu_activation: true
36
+ liger_fused_linear_cross_entropy: false
37
+ cut_cross_entropy: true
38
+
39
+ torch_compile: auto
40
+ torch_compile_backend:
41
+
42
+ sequence_len: 32768
43
+ sample_packing: true
44
+ eval_sample_packing: false
45
+ pad_to_sequence_len: true
46
+
47
+ adapter: lora
48
+ lora_model_dir:
49
+ lora_r: 128
50
+ lora_alpha: 16
51
+ lora_dropout: 0.05
52
+ lora_target_linear: true
53
+ lora_fan_in_fan_out:
54
+ peft_use_rslora: true
55
+ lora_modules_to_save:
56
+ - embed_tokens
57
+ - lm_head
58
+
59
+ wandb_project: Pgymalion
60
+ wandb_entity:
61
+ wandb_watch:
62
+ wandb_name: Big-Picaro-attempt-1
63
+ wandb_log_model:
64
+
65
+ gradient_accumulation_steps: 1
66
+ micro_batch_size: 2
67
+ num_epochs: 2
68
+ optimizer: paged_adamw_8bit
69
+ lr_scheduler: cosine
70
+ learning_rate: 4.0e-5
71
+ max_grad_norm: 0.2
72
+
73
+ train_on_inputs: false
74
+ group_by_length: false
75
+ bf16: auto
76
+ fp16:
77
+ tf32: false
78
+
79
+ gradient_checkpointing: unsloth
80
+ early_stopping_patience:
81
+ resume_from_checkpoint:
82
+ local_rank:
83
+ logging_steps: 1
84
+ xformers_attention:
85
+ flash_attention: true
86
+ s2_attention:
87
+
88
+ warmup_steps: 40
89
+ evals_per_epoch:
90
+ eval_table_size:
91
+ eval_max_new_tokens:
92
+ saves_per_epoch: 2
93
+ debug:
94
+ deepspeed: ./deepspeed_configs/zero3_bf16.json
95
+ weight_decay: 0.01
96
+ fsdp:
97
+ fsdp_config:
98
+ special_tokens:
99
+ pad_token: <|finetune_right_pad_id|>