--- base_model: meta-llama/Meta-Llama-3-8B-Instruct inference: false pipeline_tag: text-generation language: - en license: other license_name: llama3 license_link: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/LICENSE model_creator: meta-llama model_name: Meta-Llama-3-8B-Instruct model_type: llama tags: - facebook - meta - pytorch - llama - llama-3 quantized_by: brittlewis12 --- # Meta-Llama-3-8B-Instruct GGUF **Original model**: [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) **Model creator**: [Meta](https://huggingface.co/meta-llama) This repo contains GGUF format model files for Meta’s Llama-3-8B-Instruct, **updated as of 2024-04-20** to handle the `<|eot_id|>` special token as EOS token. Learn more on Meta’s [Llama 3 page](https://llama.meta.com/llama3). ### What is GGUF? GGUF is a file format for representing AI models. It is the third version of the format, introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp. Converted with llama.cpp build 2700 (revision [aed82f6](https://github.com/ggerganov/llama.cpp/commit/aed82f6837a3ea515f4d50201cfc77effc7d41b4)), using [autogguf](https://github.com/brittlewis12/autogguf). ### Prompt template ``` <|start_header_id|>system<|end_header_id|> {{system_prompt}}<|eot_id|><|start_header_id|>user<|end_header_id|> {{prompt}}<|eot_id|><|start_header_id|>assistant<|end_header_id|> ``` --- ## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac! ![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg) [cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device: - create & save **Characters** with custom system prompts & temperature settings - download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)! - make it your own with custom **Theme colors** - powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggerganov/llama.cpp), with **haptics** during response streaming! - **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)! - follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date --- ## Original Model Evaluation
Benchmark Llama 3 8B Llama 2 7B Llama 2 13B
MMLU (5-shot) 68.4 34.1 47.8
GPQA (0-shot) 34.2 21.7 22.3
HumanEval (0-shot) 62.2 7.9 14.0
GSM-8K (8-shot, CoT) 79.6 25.7 77.4
MATH (4-shot, CoT) 30.0 3.8 6.7