Tonic commited on
Commit
c7cffbb
·
1 Parent(s): fa7de39

adds quantization configuration correctly

Browse files
Files changed (1) hide show
  1. scripts/training/train_gpt_oss.py +11 -4
scripts/training/train_gpt_oss.py CHANGED
@@ -70,9 +70,9 @@ def setup_lora_for_gpt_oss(model, config):
70
 
71
  # LoRA configuration as per tutorial
72
  lora_config = LoraConfig(
73
- r=config.lora_config.get("r", 8),
74
- lora_alpha=config.lora_config.get("lora_alpha", 16),
75
- target_modules=config.lora_config.get("target_modules", "all-linear"),
76
  target_parameters=config.lora_config.get("target_parameters", [
77
  "7.mlp.experts.gate_up_proj",
78
  "7.mlp.experts.down_proj",
@@ -80,7 +80,14 @@ def setup_lora_for_gpt_oss(model, config):
80
  "15.mlp.experts.down_proj",
81
  "23.mlp.experts.gate_up_proj",
82
  "23.mlp.experts.down_proj",
83
- ]),
 
 
 
 
 
 
 
84
  )
85
 
86
  peft_model = get_peft_model(model, lora_config)
 
70
 
71
  # LoRA configuration as per tutorial
72
  lora_config = LoraConfig(
73
+ r=config.lora_config.get("r", 8) if config.lora_config else 8,
74
+ lora_alpha=config.lora_config.get("lora_alpha", 16) if config.lora_config else 16,
75
+ target_modules=config.lora_config.get("target_modules", "all-linear") if config.lora_config else "all-linear",
76
  target_parameters=config.lora_config.get("target_parameters", [
77
  "7.mlp.experts.gate_up_proj",
78
  "7.mlp.experts.down_proj",
 
80
  "15.mlp.experts.down_proj",
81
  "23.mlp.experts.gate_up_proj",
82
  "23.mlp.experts.down_proj",
83
+ ]) if config.lora_config else [
84
+ "7.mlp.experts.gate_up_proj",
85
+ "7.mlp.experts.down_proj",
86
+ "15.mlp.experts.gate_up_proj",
87
+ "15.mlp.experts.down_proj",
88
+ "23.mlp.experts.gate_up_proj",
89
+ "23.mlp.experts.down_proj",
90
+ ],
91
  )
92
 
93
  peft_model = get_peft_model(model, lora_config)