{ "model_class": "GLMModel", "tokenizer_type": "glm_ChineseSPTokenizer", "num_layers": 48, "hidden_size": 4096, "num_attention_heads": 64, "vocab_size": 50048, "hidden_dropout": 0.1, "attention_dropout": 0.1, "layernorm_order": "pre", "model_parallel_size": 1, "max_sequence_length": 1025, "block_lm": "true", "masked_lm": false, "bert_prob": 0.5, "gpt_infill_prob": 0.5, "gpt_min_ratio": 0.5, "gap_sentence_prob": 0.0, "gap_sentence_ratio": 0.15, "avg_block_length": 3, "short_seq_prob": 0.0, "single_span_prob": 0.0, "task_mask": "true", "no_shuffle_block": false, "no_block_position": false, "sentinel_token": false, "block_mask_prob": 0.0, "context_mask_ratio": 0.0, "random_position": false, "cloze_eval": "true", "old_checkpoint": false, "tokenizer_model_type": "glm-10b" }