Suparious commited on
Commit
2fba67d
·
verified ·
1 Parent(s): f554ab5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -2
README.md CHANGED
@@ -1,13 +1,40 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  inference: false
 
3
  ---
4
  # mlabonne/ChimeraLlama-3-8B AWQ
5
 
6
- ** PROCESSING .... ETA 30mins **
7
-
8
  - Model creator: [mlabonne](https://huggingface.co/mlabonne)
9
  - Original model: [ChimeraLlama-3-8B](https://huggingface.co/mlabonne/ChimeraLlama-3-8B)
10
 
 
 
 
 
 
 
 
 
 
 
11
  ### About AWQ
12
 
13
  AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
 
1
  ---
2
+ license: other
3
+ library_name: transformers
4
+ tags:
5
+ - 4-bit
6
+ - AWQ
7
+ - text-generation
8
+ - autotrain_compatible
9
+ - endpoints_compatible
10
+ - merge
11
+ - mergekit
12
+ - lazymergekit
13
+ - llama
14
+ base_model:
15
+ - NousResearch/Meta-Llama-3-8B-Instruct
16
+ - mlabonne/OrpoLlama-3-8B
17
+ - Locutusque/Llama-3-Orca-1.0-8B
18
+ - abacusai/Llama-3-Smaug-8B
19
+ pipeline_tag: text-generation
20
  inference: false
21
+ quantized_by: Suparious
22
  ---
23
  # mlabonne/ChimeraLlama-3-8B AWQ
24
 
 
 
25
  - Model creator: [mlabonne](https://huggingface.co/mlabonne)
26
  - Original model: [ChimeraLlama-3-8B](https://huggingface.co/mlabonne/ChimeraLlama-3-8B)
27
 
28
+ ## Model Summary
29
+
30
+ ChimeraLlama-3-8B outperforms Llama 3 8B Instruct on Nous' benchmark suite.
31
+
32
+ ChimeraLlama-3-8B is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
33
+ * [NousResearch/Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct)
34
+ * [mlabonne/OrpoLlama-3-8B](https://huggingface.co/mlabonne/OrpoLlama-3-8B)
35
+ * [Locutusque/Llama-3-Orca-1.0-8B](https://huggingface.co/Locutusque/Llama-3-Orca-1.0-8B)
36
+ * [abacusai/Llama-3-Smaug-8B](https://huggingface.co/abacusai/Llama-3-Smaug-8B)
37
+
38
  ### About AWQ
39
 
40
  AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.