--- base_model: meta-llama/Meta-Llama-3-8B-Instruct inference: false pipeline_tag: text-generation language: - en license: other license_name: llama3 license_link: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/LICENSE model_creator: meta-llama model_name: Meta-Llama-3-8B-Instruct model_type: llama tags: - facebook - meta - pytorch - llama - llama-3 quantized_by: brittlewis12 --- # Meta-Llama-3-8B-Instruct GGUF **Original model**: [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) **Model creator**: [Meta](https://huggingface.co/meta-llama) This repo contains GGUF format model files for Meta’s Llama-3-8B-Instruct, **updated as of 2024-04-20** to handle the `<|eot_id|>` special token as EOS token. Learn more on Meta’s [Llama 3 page](https://llama.meta.com/llama3). ### What is GGUF? GGUF is a file format for representing AI models. It is the third version of the format, introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp. Converted with llama.cpp build 2700 (revision [aed82f6](https://github.com/ggerganov/llama.cpp/commit/aed82f6837a3ea515f4d50201cfc77effc7d41b4)), using [autogguf](https://github.com/brittlewis12/autogguf). ### Prompt template ``` <|start_header_id|>system<|end_header_id|> {{system_prompt}}<|eot_id|><|start_header_id|>user<|end_header_id|> {{prompt}}<|eot_id|><|start_header_id|>assistant<|end_header_id|> ``` --- ## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac! ![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg) [cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device: - create & save **Characters** with custom system prompts & temperature settings - download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)! - make it your own with custom **Theme colors** - powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggerganov/llama.cpp), with **haptics** during response streaming! - **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)! - follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date --- ## Original Model Evaluation
Benchmark | Llama 3 8B | Llama 2 7B | Llama 2 13B |
MMLU (5-shot) | 68.4 | 34.1 | 47.8 |
GPQA (0-shot) | 34.2 | 21.7 | 22.3 |
HumanEval (0-shot) | 62.2 | 7.9 | 14.0 |
GSM-8K (8-shot, CoT) | 79.6 | 25.7 | 77.4 |
MATH (4-shot, CoT) | 30.0 | 3.8 | 6.7 |