teowu commited on
Commit
f595320
1 Parent(s): c0dddce

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -25,7 +25,7 @@
25
  "num_experts_per_tok": 6,
26
  "num_hidden_layers": 28,
27
  "num_key_value_heads": 20,
28
- "rope_theta": 5000000,
29
  "vocab_size": 100352
30
  },
31
  "torch_dtype": "bfloat16",
 
25
  "num_experts_per_tok": 6,
26
  "num_hidden_layers": 28,
27
  "num_key_value_heads": 20,
28
+ "rope_theta": 100000,
29
  "vocab_size": 100352
30
  },
31
  "torch_dtype": "bfloat16",