MaziyarPanahi
commited on
Commit
•
545e8f9
1
Parent(s):
19cefd1
Upload folder using huggingface_hub
Browse files- README.md +75 -0
- config.json +41 -0
- generation_config.json +7 -0
- model-00001-of-00015.safetensors +3 -0
- model-00002-of-00015.safetensors +3 -0
- model-00003-of-00015.safetensors +3 -0
- model-00004-of-00015.safetensors +3 -0
- model-00005-of-00015.safetensors +3 -0
- model-00006-of-00015.safetensors +3 -0
- model-00007-of-00015.safetensors +3 -0
- model-00008-of-00015.safetensors +3 -0
- model-00009-of-00015.safetensors +3 -0
- model-00010-of-00015.safetensors +3 -0
- model-00011-of-00015.safetensors +3 -0
- model-00012-of-00015.safetensors +3 -0
- model-00013-of-00015.safetensors +3 -0
- model-00014-of-00015.safetensors +3 -0
- model-00015-of-00015.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +30 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +43 -0
README.md
ADDED
@@ -0,0 +1,75 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- finetuned
|
4 |
+
- quantized
|
5 |
+
- 4-bit
|
6 |
+
- AWQ
|
7 |
+
- transformers
|
8 |
+
- tensorboard
|
9 |
+
- safetensors
|
10 |
+
- mixtral
|
11 |
+
- text-generation
|
12 |
+
- trl
|
13 |
+
- orpo
|
14 |
+
- generated_from_trainer
|
15 |
+
- conversational
|
16 |
+
- dataset:argilla/distilabel-capybara-dpo-7k-binarized
|
17 |
+
- arxiv:2403.07691
|
18 |
+
- arxiv:2311.07911
|
19 |
+
- base_model:mistral-community/Mixtral-8x22B-v0.1
|
20 |
+
- license:apache-2.0
|
21 |
+
- autotrain_compatible
|
22 |
+
- endpoints_compatible
|
23 |
+
- has_space
|
24 |
+
- text-generation-inference
|
25 |
+
- region:us
|
26 |
+
model_name: zephyr-orpo-141b-A35b-v0.1-AWQ
|
27 |
+
base_model: HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1
|
28 |
+
inference: false
|
29 |
+
model_creator: HuggingFaceH4
|
30 |
+
pipeline_tag: text-generation
|
31 |
+
quantized_by: MaziyarPanahi
|
32 |
+
---
|
33 |
+
# Description
|
34 |
+
[MaziyarPanahi/zephyr-orpo-141b-A35b-v0.1-AWQ](https://huggingface.co/MaziyarPanahi/zephyr-orpo-141b-A35b-v0.1-AWQ) is a quantized (AWQ) version of [HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1](https://huggingface.co/HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1)
|
35 |
+
|
36 |
+
## How to use
|
37 |
+
### Install the necessary packages
|
38 |
+
|
39 |
+
```
|
40 |
+
pip install --upgrade accelerate autoawq transformers
|
41 |
+
```
|
42 |
+
|
43 |
+
### Example Python code
|
44 |
+
|
45 |
+
|
46 |
+
```python
|
47 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
48 |
+
|
49 |
+
model_id = "MaziyarPanahi/zephyr-orpo-141b-A35b-v0.1-AWQ"
|
50 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
51 |
+
model = AutoModelForCausalLM.from_pretrained(model_id).to(0)
|
52 |
+
|
53 |
+
text = "User:\nHello can you provide me with top-3 cool places to visit in Paris?\n\nAssistant:\n"
|
54 |
+
inputs = tokenizer(text, return_tensors="pt").to(0)
|
55 |
+
|
56 |
+
out = model.generate(**inputs, max_new_tokens=300)
|
57 |
+
print(tokenizer.decode(out[0], skip_special_tokens=True))
|
58 |
+
```
|
59 |
+
|
60 |
+
Results:
|
61 |
+
```
|
62 |
+
User:
|
63 |
+
Hello can you provide me with top-3 cool places to visit in Paris?
|
64 |
+
|
65 |
+
Assistant:
|
66 |
+
Absolutely, here are my top-3 recommendations for must-see places in Paris:
|
67 |
+
|
68 |
+
1. The Eiffel Tower: An icon of Paris, this wrought-iron lattice tower is a global cultural icon of France and is among the most recognizable structures in the world. Climbing up to the top offers breathtaking views of the city.
|
69 |
+
|
70 |
+
2. The Louvre Museum: Home to thousands of works of art, the Louvre is the world's largest art museum and a historic monument in Paris. Must-see pieces include the Mona Lisa, the Winged Victory of Samothrace, and the Venus de Milo.
|
71 |
+
|
72 |
+
3. Notre-Dame Cathedral: This cathedral is a masterpiece of French Gothic architecture and is famous for its intricate stone carvings, beautiful stained glass, and its iconic twin towers. Be sure to spend some time exploring its history and learning about the fascinating restoration efforts post the 2019 fire.
|
73 |
+
|
74 |
+
I hope you find these recommendations helpful and that they make for an enjoyable and memorable trip to Paris. Safe travels!
|
75 |
+
```
|
config.json
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/home/maziyar/.cache/huggingface/hub/models--HuggingFaceH4--zephyr-orpo-141b-A35b-v0.1/snapshots/ae6befe2e6e81c6eab238ff7921ffadfff169a10",
|
3 |
+
"architectures": [
|
4 |
+
"MixtralForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 1,
|
8 |
+
"eos_token_id": 2,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 6144,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 16384,
|
13 |
+
"max_position_embeddings": 65536,
|
14 |
+
"model_type": "mixtral",
|
15 |
+
"num_attention_heads": 48,
|
16 |
+
"num_experts_per_tok": 2,
|
17 |
+
"num_hidden_layers": 56,
|
18 |
+
"num_key_value_heads": 8,
|
19 |
+
"num_local_experts": 8,
|
20 |
+
"output_router_logits": false,
|
21 |
+
"quantization_config": {
|
22 |
+
"bits": 4,
|
23 |
+
"group_size": 128,
|
24 |
+
"modules_to_not_convert": [
|
25 |
+
"gate"
|
26 |
+
],
|
27 |
+
"quant_method": "awq",
|
28 |
+
"version": "gemm",
|
29 |
+
"zero_point": true
|
30 |
+
},
|
31 |
+
"rms_norm_eps": 1e-05,
|
32 |
+
"rope_theta": 1000000,
|
33 |
+
"router_aux_loss_coef": 0.001,
|
34 |
+
"router_jitter_noise": 0.0,
|
35 |
+
"sliding_window": null,
|
36 |
+
"tie_word_embeddings": false,
|
37 |
+
"torch_dtype": "float16",
|
38 |
+
"transformers_version": "4.38.2",
|
39 |
+
"use_cache": true,
|
40 |
+
"vocab_size": 32000
|
41 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 1,
|
4 |
+
"do_sample": true,
|
5 |
+
"eos_token_id": 2,
|
6 |
+
"transformers_version": "4.38.2"
|
7 |
+
}
|
model-00001-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a541e30745016f1268f4a15343be148ffc71e2687ef8307ced7fa979288d3579
|
3 |
+
size 4969773192
|
model-00002-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:762f0f2b03d5b139e81d30cbde88961c5d065543b07cf28c483cd6666b5a93a3
|
3 |
+
size 4994966792
|
model-00003-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c5a5f31bb5e9b715a88c035e20bdf9caf983e2029024ea3da30f9ba21254726
|
3 |
+
size 4994966904
|
model-00004-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1ff5150fc9b274d94dd900af133c3788bc01c73258a26753acccee00b84040b
|
3 |
+
size 4994967128
|
model-00005-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:662cafc464074f1c4943b56d9aadf3f8f428cdea3ce7867fa3756d6dd14a719e
|
3 |
+
size 4999807128
|
model-00006-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d6adb6689526b3e2e5947c7562c9a58a7b72b42d8051c80522410714c3d5d982
|
3 |
+
size 4996540120
|
model-00007-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8584530e5c7e375e5a1483e3a0c6a2a8c68a3da4f9776e3a62dd75561958b783
|
3 |
+
size 4994967128
|
model-00008-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6611c984a88693cc40bc72965cf59bfd6b0c6fa6ef1d33888c19d3c39334123
|
3 |
+
size 4994967128
|
model-00009-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b9e045bce192463d59b171df1372a036db7e1fc6f93803a2bd8150433e196f52
|
3 |
+
size 4994967128
|
model-00010-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb7551371e481602d0138832cc63fdc80f8afef12782c9fc8fa4febd0bcfbcb1
|
3 |
+
size 4994967128
|
model-00011-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f74027b6db34ff6001a7659baea2c65622c3767c7a7a9b4a452798269b72b285
|
3 |
+
size 4999807128
|
model-00012-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2dfad69fcb2f8ae40d37f8a5d57ff6b428257b60c6c589e43a15d044e74d5061
|
3 |
+
size 4996540120
|
model-00013-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1fa8a205db0185a2c74e9ca4051787e230a703d105c1d6eac5a0fdda4e250a7
|
3 |
+
size 4994967128
|
model-00014-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:90056117fc4cae9f28e925255944e00e88b753c9d3d24e4478d370ca342be2d6
|
3 |
+
size 4994967128
|
model-00015-of-00015.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b552f62e6925f78072cab5e856b6bdcf56f93796d44c4449131c7097ae908ec
|
3 |
+
size 3727506680
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
special_tokens_map.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "</s>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"unk_token": {
|
24 |
+
"content": "<unk>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
3 |
+
size 493443
|
tokenizer_config.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"0": {
|
6 |
+
"content": "<unk>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"1": {
|
14 |
+
"content": "<s>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"2": {
|
22 |
+
"content": "</s>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
}
|
29 |
+
},
|
30 |
+
"additional_special_tokens": [],
|
31 |
+
"bos_token": "<s>",
|
32 |
+
"chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
|
33 |
+
"clean_up_tokenization_spaces": false,
|
34 |
+
"eos_token": "</s>",
|
35 |
+
"legacy": true,
|
36 |
+
"model_max_length": 2048,
|
37 |
+
"pad_token": "</s>",
|
38 |
+
"sp_model_kwargs": {},
|
39 |
+
"spaces_between_special_tokens": false,
|
40 |
+
"tokenizer_class": "LlamaTokenizer",
|
41 |
+
"unk_token": "<unk>",
|
42 |
+
"use_default_system_prompt": false
|
43 |
+
}
|