{ "architectures": [ "PhiForCausalLM" ], "model_type": "phi", "torch_dtype": "float16", "transformers_version": "4.35.0", "max_position_embeddings": 2048, "quantization_config": { "load_in_4bit": true, "bnb_4bit_compute_dtype": "float16" }, "peft_config": { "peft_type": "LORA", "task_type": "CAUSAL_LM", "inference_mode": false, "r": 16, "lora_alpha": 16, "lora_dropout": 0 } }