|
--- |
|
datasets: |
|
- TIGER-Lab/MathInstruct |
|
base_model: |
|
- meta-llama/Llama-3.2-3B-Instruct |
|
tags: |
|
- lora |
|
- adapters |
|
- llama |
|
--- |
|
|
|
# **LLAMA-3.2-3B-MathInstruct_LORA_SFT** |
|
|
|
This model is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on the [MathInstruct](https://huggingface.co/datasets/TIGER-Lab/MathInstruct) dataset. The fine-tuning process was designed to enhance the model's performance for mathematical instruction-following tasks, ensuring improved accuracy and precision when solving math-related problems. |
|
|
|
It achieves the following results on the evaluation set: |
|
- **Loss**: 0.6895 |
|
|
|
--- |
|
|
|
## **Model Description** |
|
|
|
This model is specifically fine-tuned for mathematical reasoning, problem-solving, and instruction-following tasks. Leveraging the LLaMA-3.2-3B-Instruct base model, it has been optimized to handle mathematical queries and tasks with improved efficiency and context understanding. |
|
|
|
--- |
|
|
|
## **Training and Evaluation Data** |
|
|
|
The model was fine-tuned on the [MathInstruct](https://huggingface.co/datasets/TIGER-Lab/MathInstruct) dataset. |
|
- **Dataset Source**: TIGER-Lab. |
|
- **Dataset Focus**: Mathematical instruction-following and reasoning tasks. |
|
- **Scope**: A wide range of math topics, including arithmetic, algebra, calculus, and problem-solving. |
|
|
|
The dataset was carefully curated to align with instructional objectives for solving mathematical problems and understanding step-by-step reasoning. |
|
|
|
--- |
|
|
|
## **Training Procedure** |
|
|
|
### **Hyperparameters** |
|
- **Learning rate**: 0.0001 |
|
- **Train batch size**: 1 |
|
- **Eval batch size**: 1 |
|
- **Gradient accumulation steps**: 8 |
|
- **Total effective batch size**: 8 |
|
- **Optimizer**: AdamW (torch) |
|
- **Betas**: (0.9, 0.999) |
|
- **Epsilon**: 1e-08 |
|
- **Learning rate scheduler**: Cosine schedule with 10% warmup. |
|
- **Number of epochs**: 3.0 |
|
|
|
### **Framework Versions** |
|
- **PEFT**: 0.12.0 |
|
- **Transformers**: 4.46.1 |
|
- **PyTorch**: 2.5.1+cu124 |
|
- **Datasets**: 3.1.0 |
|
- **Tokenizers**: 0.20.3 |
|
|
|
--- |
|
|
|
## **Training Results** |
|
|
|
- **Loss**: 0.6895 |
|
- Evaluation indicates strong performance on math instruction-following tasks. Further testing on specific use cases is recommended to assess the model’s generalizability. |
|
|
|
--- |
|
|
|
## **Additional Information** |
|
|
|
- **Author**: Sri Santh M |
|
- **Purpose**: Fine-tuned for educational and development purposes, particularly for math-related tasks. |
|
- **Dataset Link**: [MathInstruct Dataset](https://huggingface.co/datasets/TIGER-Lab/MathInstruct) |
|
|
|
This model represents a focused effort to adapt the LLaMA-3.2-3B-Instruct model for specialized mathematical use cases. It can be further fine-tuned or extended for more specific mathematical domains or applications. |