autoprogrammer commited on
Commit
a0eb3e2
·
verified ·
1 Parent(s): 03b5df7

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +6 -1
config.json CHANGED
@@ -3,6 +3,11 @@
3
  "architectures": [
4
  "DenseBackwardOLMoEForCausalLM"
5
  ],
 
 
 
 
 
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "clip_qkv": null,
@@ -12,7 +17,7 @@
12
  "initializer_range": 0.02,
13
  "intermediate_size": 1024,
14
  "max_position_embeddings": 4096,
15
- "model_type": "DenseBackward_olmoe",
16
  "norm_topk_prob": false,
17
  "num_attention_heads": 16,
18
  "num_experts": 64,
 
3
  "architectures": [
4
  "DenseBackwardOLMoEForCausalLM"
5
  ],
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_densebackward_olmoe.DenseBackwardOLMoEConfig",
8
+ "AutoModel": "modeling_densebackward_olmoe.DenseBackwardOLMoEForCausalLM",
9
+ "AutoModelForCausalLM": "modeling_densebackward_olmoe.DenseBackwardOLMoEForCausalLM"
10
+ },
11
  "attention_bias": false,
12
  "attention_dropout": 0.0,
13
  "clip_qkv": null,
 
17
  "initializer_range": 0.02,
18
  "intermediate_size": 1024,
19
  "max_position_embeddings": 4096,
20
+ "model_type": "olmoe",
21
  "norm_topk_prob": false,
22
  "num_attention_heads": 16,
23
  "num_experts": 64,