--- language: - en license: apache-2.0 tags: - code - unsloth - trl - sft --- This is a fine-tune of Llama3-8B base model using an Alpaca-like instruction-tuning dataset generated with GPT4. My aim with doing this was to compare the instruction-tuned performance of this model with the one Llama has available.