{"block_size": 10, "vocab_size": 128, "n_layer": 4, "n_head": 12, "n_embd": 120, "dropout": 0.01, "bias": true, "attention": {"attention_type": "default_attention", "scaling_factor": 2}, "activation": "gelu", "epsilon": 1e-05, "model_type": "llm_gym_gpt2"}