--- base_model: - qingy2024/Qwarkstar-4B license: apache-2.0 datasets: - HuggingFaceTB/smoltalk language: - en pipeline_tag: text-generation --- ## Qwarkstar 4B Instruct (Preview) > [!NOTE] > Training complete! This model is fine-tuned using Supervised Fine-Tuning (SFT) on 100k samples from the `HuggingFaceTB/smoltalk` dataset. It follows the ChatML input-output formatting template. ### Training Details: - **Base Model**: `qingy2024/Qwarkstar-4B` - **Batch Size**: 32 (2 H100s x 8 per GPU) - **Max Gradient Norm**: 1.0 - **Final Loss**: ~0.59