MathGLM / model_config.json
ZhenYang21's picture
Upload 2 files
840e994
raw
history blame contribute delete
903 Bytes
{
"model_class": "GLMModel",
"tokenizer_type": "glm_ChineseSPTokenizer",
"num_layers": 48,
"hidden_size": 4096,
"num_attention_heads": 64,
"vocab_size": 50048,
"hidden_dropout": 0.1,
"attention_dropout": 0.1,
"layernorm_order": "pre",
"model_parallel_size": 1,
"max_sequence_length": 1025,
"block_lm": "true",
"masked_lm": false,
"bert_prob": 0.5,
"gpt_infill_prob": 0.5,
"gpt_min_ratio": 0.5,
"gap_sentence_prob": 0.0,
"gap_sentence_ratio": 0.15,
"avg_block_length": 3,
"short_seq_prob": 0.0,
"single_span_prob": 0.0,
"task_mask": "true",
"no_shuffle_block": false,
"no_block_position": false,
"sentinel_token": false,
"block_mask_prob": 0.0,
"context_mask_ratio": 0.0,
"random_position": false,
"cloze_eval": "true",
"old_checkpoint": false,
"tokenizer_model_type": "glm-10b"
}