--- library_name: transformers license: apache-2.0 datasets: - TokenBender/code_instructions_122k_alpaca_style --- # Model Card for Model ID This is a LoRA adapter for Mistral-Instruct-v0.2 Base Model: [Mistral-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) Dataset: [TokenBender/code_instructions_122k_alpaca_style](https://huggingface.co/datasets/TokenBender/code_instructions_122k_alpaca_style) Test run, trained on subsample of 5k examples for 1 epoch using QLoRa technique. Final loss: 0.394200