{
	"dim": 4096,
	"multiple_of": 256,
	"n_heads": 32,
	"n_layers": 32,
	"norm_eps": 1e-05,
	"vocab_size": -1,
	"model_type": "llama"
}