RaymondAISG commited on
Commit
7cf4267
1 Parent(s): 484e86e

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +10 -5
config.json CHANGED
@@ -4,7 +4,15 @@
4
  ],
5
  "attn_config": {
6
  "alibi": false,
7
- "qk_ln": true
 
 
 
 
 
 
 
 
8
  },
9
  "auto_map": {
10
  "AutoConfig": "configuration_mpt.MPTConfig",
@@ -34,8 +42,6 @@
34
  "init_std": 0.02
35
  },
36
  "init_device": "cpu",
37
- "initializer_range": 0.02,
38
- "layer_norm_epsilon": 1e-05,
39
  "learned_pos_emb": true,
40
  "logit_scale": "inv_sqrt_d_model",
41
  "max_seq_len": 2048,
@@ -46,8 +52,7 @@
46
  "norm_type": "low_precision_layernorm",
47
  "resid_pdrop": 0.0,
48
  "torch_dtype": "bfloat16",
49
- "transformers_version": "4.34.1",
50
  "use_cache": false,
51
- "verbose": 0,
52
  "vocab_size": 256000
53
  }
 
4
  ],
5
  "attn_config": {
6
  "alibi": false,
7
+ "alibi_bias_max": 8,
8
+ "attn_impl": "torch",
9
+ "attn_pdrop": 0.0,
10
+ "attn_type": "multihead_attention",
11
+ "attn_uses_sequence_id": false,
12
+ "clip_qkv": null,
13
+ "prefix_lm": false,
14
+ "qk_ln": true,
15
+ "softmax_scale": null
16
  },
17
  "auto_map": {
18
  "AutoConfig": "configuration_mpt.MPTConfig",
 
42
  "init_std": 0.02
43
  },
44
  "init_device": "cpu",
 
 
45
  "learned_pos_emb": true,
46
  "logit_scale": "inv_sqrt_d_model",
47
  "max_seq_len": 2048,
 
52
  "norm_type": "low_precision_layernorm",
53
  "resid_pdrop": 0.0,
54
  "torch_dtype": "bfloat16",
55
+ "transformers_version": "4.31.0",
56
  "use_cache": false,
 
57
  "vocab_size": 256000
58
  }