|
--- |
|
base_model: unsloth/Qwen2.5-Coder-7B-Instruct-bnb-4bit |
|
datasets: |
|
- Hypersniper/unity_api_2022_3 |
|
- ibranze/codellama_unity3d_v2 |
|
- neph1/Unity_Code_QnA |
|
language: |
|
- en |
|
license: apache-2.0 |
|
tags: |
|
- text-generation-inference |
|
- transformers |
|
- unsloth |
|
- qwen2 |
|
- trl |
|
- sft |
|
--- |
|
|
|
# Description |
|
|
|
Qwen2.5-Coder-7B-Instruct trained on a merged dataset of Unity3d q&a from these three datasets: |
|
|
|
[ibranze/codellama_unity3d_v2](https://huggingface.co/datasets/ibranze/codellama_unity3d_v2) (Full) |
|
|
|
[Hypersniper/unity_api_2022_3](https://huggingface.co/datasets/Hypersniper/unity_api_2022_3) (10%) |
|
|
|
[neph1/Unity_Code_QnA](https://huggingface.co/datasets/neph1/Unity_Code_QnA) (Full) |
|
|
|
|
|
preview 2: |
|
26210 rows, of which ca 1000 are from my own multi response dataset |
|
|
|
preview 1: |
|
15062 rows in total with a 10% validation split. |
|
|
|
Trained with native chat template (minus tools usage, see this issue: https://github.com/unslothai/unsloth/issues/1053). With a little superficial testing done, it seems to respond well to the mistral template. |
|
|
|
|
|
Consider this a preview while I develop a dataset of my own. |
|
|
|
If you have any feedback, please share. I've only done some basic testing so far. I'm especially interested if you're using it with Tabby or a similar coding tool. |
|
|
|
|
|
# Uploaded model |
|
|
|
- **Developed by:** neph1 |
|
- **License:** apache-2.0 |
|
- **Finetuned from model :** unsloth/Qwen2.5-Coder-7B-Instruct-bnb-4bit |
|
|
|
This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. |
|
|
|
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth) |
|
|
|
# Training details |
|
|
|
About 1.5 epochs. It's probably a bit overfitting and I should introduce some general coding questions to my validation set to ensure it doesn't lose too much general performance. |
|
|
|
Rank: 128 |
|
|
|
Alpha: 256 |
|
|
|
TrainingArguments( |
|
per_device_train_batch_size =2, |
|
gradient_accumulation_steps = 64, |
|
#max_steps=10, |
|
num_train_epochs=3, |
|
warmup_steps = 5, |
|
learning_rate = 1e-4, |
|
fp16 = not torch.cuda.is_bf16_supported(), |
|
bf16 = torch.cuda.is_bf16_supported(), |
|
logging_steps = 10, |
|
optim = "adamw_8bit", |
|
weight_decay = 0.01, |
|
lr_scheduler_type = "linear", |
|
seed = 3407, |
|
per_device_eval_batch_size = 2, |
|
eval_strategy="steps", |
|
eval_accumulation_steps = 64, |
|
eval_steps = 10, |
|
eval_delay = 0, |
|
save_strategy="steps", |
|
save_steps=25, |
|
report_to="none", |
|
), |
|
|
|
|
|
Step Training Loss Validation Loss |
|
|
|
20 2.043000 1.197104 |
|
|
|
40 1.087300 0.933553 |
|
|
|
60 0.942200 0.890801 |
|
|
|
80 0.865600 0.866198 |
|
|
|
100 0.851400 0.849733 |
|
|
|
120 0.812900 0.837039 |
|
|
|
140 0.812400 0.827064 |
|
|
|
160 0.817300 0.818410 |
|
|
|
180 0.802600 0.810163 |
|
|
|
200 0.788600 0.803399 |