{ | |
"adapter_path": "/Users/teun/Documents/output-cleaned/adapters", | |
"batch_size": 4, | |
"config": null, | |
"data": "/Users/teun/Documents/output-cleaned", | |
"fine_tune_type": "lora", | |
"grad_checkpoint": null, | |
"iters": 600, | |
"learning_rate": 1e-05, | |
"lora_parameters": { | |
"rank": 8, | |
"alpha": 16, | |
"dropout": 0.0, | |
"scale": 10.0 | |
}, | |
"lr_schedule": null, | |
"max_seq_length": 2048, | |
"model": "mlx-community/gemma-2-9b-it-4bit", | |
"num_layers": 16, | |
"resume_adapter_file": null, | |
"save_every": 100, | |
"seed": 0, | |
"steps_per_eval": 200, | |
"steps_per_report": 10, | |
"test": false, | |
"test_batches": 500, | |
"train": true, | |
"val_batches": 25 | |
} |