|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- vicgalle/alpaca-gpt4 |
|
language: |
|
- en |
|
--- |
|
This is a finetuned version microsoft phi - 2. This is finetuned using the alpaca dataset. |
|
|
|
|
|
## Training arguments |
|
|
|
num_train_epochs=1 |
|
|
|
per_device_train_batch_size=1 |
|
|
|
gradient_accumulation_steps=8 |
|
|
|
optim="paged_adamw_32bit" |
|
|
|
logging_steps=25 |
|
|
|
learning_rate=2e-4 |
|
|
|
weight_decay=0.001 |
|
|
|
max_grad_norm=0.3 |
|
|
|
warmup_ratio=0.03 |
|
|
|
lr_scheduler_type="cosine" |
|
|
|
## Prompt format: |
|
|
|
Use the below format |
|
|
|
``` |
|
Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction:{prompt} ### Response:{generated_output} |
|
``` |
|
|