--- library_name: transformers license: apache-2.0 base_model: Qwen/Qwen2.5-0.5B-Instruct tags: - generated_from_trainer model-index: - name: careerbot_PG6_Qwen_Qwen2.5-0.5B-Instruct_model results: [] --- # careerbot_PG6_Qwen_Qwen2.5-0.5B-Instruct_model This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.7927 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Use OptimizerNames.ADAFACTOR and the args are: No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1 - training_steps: 3372 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-------:|:----:|:---------------:| | No log | 0.9953 | 105 | 1.1560 | | No log | 2.0 | 211 | 1.1306 | | No log | 2.9953 | 316 | 1.1640 | | No log | 4.0 | 422 | 1.2417 | | No log | 4.9953 | 527 | 1.2741 | | No log | 6.0 | 633 | 1.3009 | | No log | 6.9953 | 738 | 1.3479 | | No log | 8.0 | 844 | 1.4053 | | No log | 8.9953 | 949 | 1.4463 | | No log | 10.0 | 1055 | 1.4523 | | No log | 10.9953 | 1160 | 1.4646 | | No log | 12.0 | 1266 | 1.5013 | | No log | 12.9953 | 1371 | 1.4868 | | No log | 14.0 | 1477 | 1.5466 | | No log | 14.9953 | 1582 | 1.5501 | | No log | 16.0 | 1688 | 1.5816 | | No log | 16.9953 | 1793 | 1.5601 | | No log | 18.0 | 1899 | 1.5774 | | No log | 18.9953 | 2004 | 1.6338 | | No log | 20.0 | 2110 | 1.6242 | | No log | 20.9953 | 2215 | 1.6408 | | No log | 22.0 | 2321 | 1.6462 | | No log | 22.9953 | 2426 | 1.6565 | | No log | 24.0 | 2532 | 1.6767 | | No log | 24.9953 | 2637 | 1.7010 | | No log | 26.0 | 2743 | 1.6947 | | No log | 26.9953 | 2848 | 1.7124 | | No log | 28.0 | 2954 | 1.7340 | | No log | 28.9953 | 3059 | 1.7522 | | No log | 30.0 | 3165 | 1.7693 | | No log | 30.9953 | 3270 | 1.7857 | | No log | 31.9621 | 3372 | 1.7927 | ### Framework versions - Transformers 4.46.1 - Pytorch 2.5.0+cu124 - Datasets 2.19.0 - Tokenizers 0.20.1