Commit
·
26b9a85
1
Parent(s):
6bb3f86
Update adapter_config.json
Browse files- adapter_config.json +1 -1
adapter_config.json
CHANGED
@@ -8,7 +8,7 @@
|
|
8 |
"init_lora_weights": true,
|
9 |
"layers_pattern": null,
|
10 |
"layers_to_transform": null,
|
11 |
-
|
12 |
"lora_alpha": 16,
|
13 |
"lora_dropout": 0.1,
|
14 |
"megatron_config": null,
|
|
|
8 |
"init_lora_weights": true,
|
9 |
"layers_pattern": null,
|
10 |
"layers_to_transform": null,
|
11 |
+
|
12 |
"lora_alpha": 16,
|
13 |
"lora_dropout": 0.1,
|
14 |
"megatron_config": null,
|