# Create output directory | |
OUTPUT_DIR="./solana_model_output" | |
mkdir -p $OUTPUT_DIR | |
# Run training script | |
python finetune_deepseekcoder.py \ | |
--model_name_or_path "deepseek-ai/deepseek-coder-6.7b-instruct" \ | |
--data_path "./solana_1000_yourgpt.jsonl" \ | |
--output_dir $OUTPUT_DIR \ | |
--num_train_epochs 3 \ | |
--per_device_train_batch_size 4 \ | |
--gradient_accumulation_steps 4 \ | |
--learning_rate 2e-5 \ | |
--warmup_ratio 0.03 \ | |
--logging_steps 10 \ | |
--save_steps 100 \ | |
--save_total_limit 3 \ | |
--lr_scheduler_type "cosine" \ | |
--fp16 true \ | |
--report_to "wandb" \ | |
--overwrite_output_dir \ | |
--ddp_find_unused_parameters false \ | |
--gradient_checkpointing true | |