test
#1
by
van-qa
- opened
- .gitattributes +0 -11
- README.md +5 -6
- llama-3.1-8b-instruct-q2_k.gguf +0 -3
- llama-3.1-8b-instruct-q3_k_l.gguf +0 -3
- llama-3.1-8b-instruct-q3_k_m.gguf +0 -3
- llama-3.1-8b-instruct-q3_k_s.gguf +0 -3
- llama-3.1-8b-instruct-q4_k_m.gguf +0 -3
- llama-3.1-8b-instruct-q4_k_s.gguf +0 -3
- llama-3.1-8b-instruct-q5_k_m.gguf +0 -3
- llama-3.1-8b-instruct-q5_k_s.gguf +0 -3
- llama-3.1-8b-instruct-q6_k.gguf +0 -3
- llama-3.1-8b-instruct-q8_0.gguf +0 -3
- metadata.yml +0 -5
- model.yml +0 -23
.gitattributes
CHANGED
@@ -33,14 +33,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
-
model.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
-
llama-3.1-8b-instruct-q2_k.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
-
llama-3.1-8b-instruct-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
-
llama-3.1-8b-instruct-q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
-
llama-3.1-8b-instruct-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
-
llama-3.1-8b-instruct-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
-
llama-3.1-8b-instruct-q4_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
-
llama-3.1-8b-instruct-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
-
llama-3.1-8b-instruct-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
-
llama-3.1-8b-instruct-q6_k.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
-
llama-3.1-8b-instruct-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
README.md
CHANGED
@@ -1,8 +1,5 @@
|
|
1 |
---
|
2 |
license: llama3
|
3 |
-
pipeline_tag: text-generation
|
4 |
-
tags:
|
5 |
-
- cortex.cpp
|
6 |
---
|
7 |
|
8 |
## Overview
|
@@ -13,13 +10,15 @@ Meta developed and released the [Meta Llama 3](https://huggingface.co/meta-llama
|
|
13 |
|
14 |
| No | Variant | Cortex CLI command |
|
15 |
| --- | --- | --- |
|
16 |
-
| 1 | [
|
|
|
|
|
17 |
|
18 |
## Use it with Jan (UI)
|
19 |
|
20 |
1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
|
21 |
2. Use in Jan model Hub:
|
22 |
-
```
|
23 |
cortexso/llama3
|
24 |
```
|
25 |
|
@@ -27,7 +26,7 @@ Meta developed and released the [Meta Llama 3](https://huggingface.co/meta-llama
|
|
27 |
|
28 |
1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart)
|
29 |
2. Run the model with command:
|
30 |
-
```
|
31 |
cortex run llama3
|
32 |
```
|
33 |
|
|
|
1 |
---
|
2 |
license: llama3
|
|
|
|
|
|
|
3 |
---
|
4 |
|
5 |
## Overview
|
|
|
10 |
|
11 |
| No | Variant | Cortex CLI command |
|
12 |
| --- | --- | --- |
|
13 |
+
| 1 | [onnx](https://huggingface.co/cortexso/llama3/tree/onnx) | `cortex run llama3:onnx` |
|
14 |
+
| 2 | [gguf](https://huggingface.co/cortexso/llama3/tree/gguf) | `cortex run llama3:gguf` |
|
15 |
+
| 3 | [default](https://huggingface.co/cortexso/llama3/tree/default) | `cortex run llama3` |
|
16 |
|
17 |
## Use it with Jan (UI)
|
18 |
|
19 |
1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
|
20 |
2. Use in Jan model Hub:
|
21 |
+
```
|
22 |
cortexso/llama3
|
23 |
```
|
24 |
|
|
|
26 |
|
27 |
1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart)
|
28 |
2. Run the model with command:
|
29 |
+
```
|
30 |
cortex run llama3
|
31 |
```
|
32 |
|
llama-3.1-8b-instruct-q2_k.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:7ba07f0e6f20ed9a9a52c3a430dda0ab1c5eb71b5f61f194dbdb2691ad28bd94
|
3 |
-
size 3179136256
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q3_k_l.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:a0dec74f5fd617ad11172f763de849332254632248a430bd4b014971d12ae58a
|
3 |
-
size 4321961216
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q3_k_m.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:85aac3fae4c469e566bfe7371ff67f143f6b7a99f347147da83479c379daed8d
|
3 |
-
size 4018922752
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q3_k_s.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:622914d9363361a6a24225cc1ccc5f952fe664517f1d2054dc84ab8fd3838017
|
3 |
-
size 3664504064
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q4_k_m.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:c33f5a84eb2bbfbefcee9ec25cc8e639836eff6b1a54818774f254c18dbb4f5b
|
3 |
-
size 4920739072
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q4_k_s.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:24dd25091cde35e4c709dbcc4120430d0500fa2176dde27e67018967fb231f1f
|
3 |
-
size 4692673792
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q5_k_m.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:0bf7159d0d4313a818da7270399ddcbb2dbfdb25d502b82e1680e40b55a1a7a7
|
3 |
-
size 5732992256
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q5_k_s.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:6ba5070018af5e009c832fbc90f7a7ab224b578be78fa7ada80c9e61ebbd39ec
|
3 |
-
size 5599298816
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q6_k.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:e4537e7b02653bfa7fe109ac917ac374bedf69cb15b9c6fea678cda4cb8de3b2
|
3 |
-
size 6596011264
|
|
|
|
|
|
|
|
llama-3.1-8b-instruct-q8_0.gguf
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:8349b563753acc025716792c2cad15de4f7e7ad5ea72a91f65c5df2bc226726f
|
3 |
-
size 8540775680
|
|
|
|
|
|
|
|
metadata.yml
DELETED
@@ -1,5 +0,0 @@
|
|
1 |
-
# metadata.yml
|
2 |
-
version: 1
|
3 |
-
name: llama3
|
4 |
-
default: 8b
|
5 |
-
author: "meta-llama"
|
|
|
|
|
|
|
|
|
|
|
|
model.yml
DELETED
@@ -1,23 +0,0 @@
|
|
1 |
-
name: Llama 3
|
2 |
-
model: llama3:8B
|
3 |
-
version: 1
|
4 |
-
|
5 |
-
# Results Preferences
|
6 |
-
stop:
|
7 |
-
- <|end_of_text|>
|
8 |
-
- <|eot_id|>
|
9 |
-
top_p: 0.95
|
10 |
-
temperature: 0.7
|
11 |
-
frequency_penalty: 0
|
12 |
-
presence_penalty: 0
|
13 |
-
max_tokens: 8192 # Infer from base config.json -> max_position_embeddings
|
14 |
-
stream: true # true | false
|
15 |
-
|
16 |
-
# Engine / Model Settings
|
17 |
-
ngl: 33 # Infer from base config.json -> num_attention_heads
|
18 |
-
ctx_len: 8192 # Infer from base config.json -> max_position_embeddings
|
19 |
-
engine: llama-cpp
|
20 |
-
prompt_template: "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
21 |
-
# Prompt template: Can only be retrieved from instruct model
|
22 |
-
# - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/tokenizer_config.json#L2053
|
23 |
-
# - Requires jinja format parser
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|