phi-2_alpaca_52k / README.md
KathirKs's picture
Update README.md
e8854f3
metadata
license: apache-2.0
datasets:
  - vicgalle/alpaca-gpt4
language:
  - en

This is a finetuned version microsoft phi - 2. This is finetuned using the alpaca dataset.

Training arguments

num_train_epochs=1

per_device_train_batch_size=1

gradient_accumulation_steps=8

optim="paged_adamw_32bit"

logging_steps=25

learning_rate=2e-4

weight_decay=0.001

max_grad_norm=0.3

warmup_ratio=0.03

lr_scheduler_type="cosine"

Prompt format:

Use the below format

Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction:{prompt} ### Response:{generated_output}