Spaces:
Running
Running
adds quantization configuration correctly
Browse files
scripts/training/train_gpt_oss.py
CHANGED
@@ -70,9 +70,9 @@ def setup_lora_for_gpt_oss(model, config):
|
|
70 |
|
71 |
# LoRA configuration as per tutorial
|
72 |
lora_config = LoraConfig(
|
73 |
-
r=config.lora_config.get("r", 8),
|
74 |
-
lora_alpha=config.lora_config.get("lora_alpha", 16),
|
75 |
-
target_modules=config.lora_config.get("target_modules", "all-linear"),
|
76 |
target_parameters=config.lora_config.get("target_parameters", [
|
77 |
"7.mlp.experts.gate_up_proj",
|
78 |
"7.mlp.experts.down_proj",
|
@@ -80,7 +80,14 @@ def setup_lora_for_gpt_oss(model, config):
|
|
80 |
"15.mlp.experts.down_proj",
|
81 |
"23.mlp.experts.gate_up_proj",
|
82 |
"23.mlp.experts.down_proj",
|
83 |
-
])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
84 |
)
|
85 |
|
86 |
peft_model = get_peft_model(model, lora_config)
|
|
|
70 |
|
71 |
# LoRA configuration as per tutorial
|
72 |
lora_config = LoraConfig(
|
73 |
+
r=config.lora_config.get("r", 8) if config.lora_config else 8,
|
74 |
+
lora_alpha=config.lora_config.get("lora_alpha", 16) if config.lora_config else 16,
|
75 |
+
target_modules=config.lora_config.get("target_modules", "all-linear") if config.lora_config else "all-linear",
|
76 |
target_parameters=config.lora_config.get("target_parameters", [
|
77 |
"7.mlp.experts.gate_up_proj",
|
78 |
"7.mlp.experts.down_proj",
|
|
|
80 |
"15.mlp.experts.down_proj",
|
81 |
"23.mlp.experts.gate_up_proj",
|
82 |
"23.mlp.experts.down_proj",
|
83 |
+
]) if config.lora_config else [
|
84 |
+
"7.mlp.experts.gate_up_proj",
|
85 |
+
"7.mlp.experts.down_proj",
|
86 |
+
"15.mlp.experts.gate_up_proj",
|
87 |
+
"15.mlp.experts.down_proj",
|
88 |
+
"23.mlp.experts.gate_up_proj",
|
89 |
+
"23.mlp.experts.down_proj",
|
90 |
+
],
|
91 |
)
|
92 |
|
93 |
peft_model = get_peft_model(model, lora_config)
|