Davlan commited on
Commit
4b51a78
·
1 Parent(s): 9937f0c

Upload adapter_config.json

Browse files
Files changed (1) hide show
  1. adapter_config.json +60 -0
adapter_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "architecture": "union",
4
+ "configs": [
5
+ {
6
+ "alpha": 1,
7
+ "architecture": "lora",
8
+ "attn_matrices": [
9
+ "k",
10
+ "v"
11
+ ],
12
+ "composition_mode": "scale",
13
+ "dropout": 0.0,
14
+ "init_weights": "ia3",
15
+ "intermediate_lora": false,
16
+ "output_lora": false,
17
+ "r": 1,
18
+ "selfattn_lora": true
19
+ },
20
+ {
21
+ "adapter_residual_before_ln": false,
22
+ "cross_adapter": false,
23
+ "factorized_phm_W": true,
24
+ "factorized_phm_rule": false,
25
+ "hypercomplex_nonlinearity": "glorot-uniform",
26
+ "init_weights": "bert",
27
+ "inv_adapter": "nice",
28
+ "inv_adapter_reduction_factor": 2,
29
+ "is_parallel": false,
30
+ "learn_phm": true,
31
+ "leave_out": [],
32
+ "ln_after": false,
33
+ "ln_before": false,
34
+ "mh_adapter": false,
35
+ "non_linearity": "relu",
36
+ "original_ln_after": false,
37
+ "original_ln_before": false,
38
+ "output_adapter": false,
39
+ "phm_bias": true,
40
+ "phm_c_init": "normal",
41
+ "phm_dim": 4,
42
+ "phm_init_range": 0.0001,
43
+ "phm_layer": false,
44
+ "phm_rank": 1,
45
+ "reduction_factor": 16,
46
+ "residual_before_ln": true,
47
+ "scaling": 1.0,
48
+ "shared_W_phm": false,
49
+ "shared_phm_rule": true
50
+ }
51
+ ]
52
+ },
53
+ "config_id": "d2c8db3436274834",
54
+ "hidden_size": 1024,
55
+ "model_class": "BloomForCausalLM",
56
+ "model_name": "bigscience/bloom-560m",
57
+ "model_type": "bloom",
58
+ "name": "oscar_ia3+inv_am",
59
+ "version": "3.1.0a0"
60
+ }