Uploaded model

  • Developed by: qingy2019
  • License: apache-2.0
  • Finetuned from model : unsloth/qwen2.5-14b-bnb-4bit

Huge thanks to Unsloth and the Huggingface TRL library.

This model is Qwen 2.5 14B fine tuned for a full epoch on the high quality garage-bAInd/Open-Platypus dataset for STEM reasoning.

Training Detail Value
Epochs 1
Steps 2077
Loss 0.4218
Batch size 4
Gradient Acc. Steps 3
Learning Rate 2e-4
LR Scheduler cosine
Rank 32
Rank-Stabilized LoRA Yes
Warm up steps 5
Weight Decay 0.01
Seed 3407

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including qingy2024/Qwen2.5-14B-Math-LoRA-Iter2