jkazdan/jkazdan/Mistral-7B-Instruct-v0.2-cipher-4500
Browse files- README.md +4 -5
- config.json +1 -1
- generation_config.json +1 -1
- model-00001-of-00003.safetensors +1 -1
- model-00002-of-00003.safetensors +1 -1
- model-00003-of-00003.safetensors +1 -1
- tokenizer.json +0 -0
- tokenizer_config.json +0 -1
- training_args.bin +2 -2
README.md
CHANGED
@@ -1,5 +1,4 @@
|
|
1 |
---
|
2 |
-
library_name: transformers
|
3 |
license: apache-2.0
|
4 |
base_model: mistralai/Mistral-7B-Instruct-v0.2
|
5 |
tags:
|
@@ -41,7 +40,7 @@ The following hyperparameters were used during training:
|
|
41 |
- seed: 42
|
42 |
- gradient_accumulation_steps: 4
|
43 |
- total_train_batch_size: 16
|
44 |
-
- optimizer:
|
45 |
- lr_scheduler_type: linear
|
46 |
- lr_scheduler_warmup_ratio: 0.05
|
47 |
- num_epochs: 1
|
@@ -52,7 +51,7 @@ The following hyperparameters were used during training:
|
|
52 |
|
53 |
### Framework versions
|
54 |
|
55 |
-
- Transformers 4.
|
56 |
-
- Pytorch 2.
|
57 |
- Datasets 2.20.0
|
58 |
-
- Tokenizers 0.
|
|
|
1 |
---
|
|
|
2 |
license: apache-2.0
|
3 |
base_model: mistralai/Mistral-7B-Instruct-v0.2
|
4 |
tags:
|
|
|
40 |
- seed: 42
|
41 |
- gradient_accumulation_steps: 4
|
42 |
- total_train_batch_size: 16
|
43 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
44 |
- lr_scheduler_type: linear
|
45 |
- lr_scheduler_warmup_ratio: 0.05
|
46 |
- num_epochs: 1
|
|
|
51 |
|
52 |
### Framework versions
|
53 |
|
54 |
+
- Transformers 4.44.0
|
55 |
+
- Pytorch 2.4.0+cu121
|
56 |
- Datasets 2.20.0
|
57 |
+
- Tokenizers 0.19.1
|
config.json
CHANGED
@@ -21,7 +21,7 @@
|
|
21 |
"sliding_window": null,
|
22 |
"tie_word_embeddings": false,
|
23 |
"torch_dtype": "bfloat16",
|
24 |
-
"transformers_version": "4.
|
25 |
"use_cache": true,
|
26 |
"vocab_size": 32000
|
27 |
}
|
|
|
21 |
"sliding_window": null,
|
22 |
"tie_word_embeddings": false,
|
23 |
"torch_dtype": "bfloat16",
|
24 |
+
"transformers_version": "4.44.0",
|
25 |
"use_cache": true,
|
26 |
"vocab_size": 32000
|
27 |
}
|
generation_config.json
CHANGED
@@ -2,5 +2,5 @@
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
-
"transformers_version": "4.
|
6 |
}
|
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
+
"transformers_version": "4.44.0"
|
6 |
}
|
model-00001-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4943162336
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6d00430516da03d27b68f3d97e60c95bec81da92586c5e01efff827368b07c28
|
3 |
size 4943162336
|
model-00002-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999819336
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d05ac46141f61da82549845aa02a88fc0e6a6f7e66e81cee7f82b0ca8b5c7229
|
3 |
size 4999819336
|
model-00003-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4540516344
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2cb43da57b40aa0c3c9abcca5b3fe43a8201a954187723d1f10b64bcefb5167f
|
3 |
size 4540516344
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -33,7 +33,6 @@
|
|
33 |
"chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\\n\\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + eos_token}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
|
34 |
"clean_up_tokenization_spaces": false,
|
35 |
"eos_token": "</s>",
|
36 |
-
"extra_special_tokens": {},
|
37 |
"legacy": false,
|
38 |
"model_max_length": 512,
|
39 |
"pad_token": "<s>",
|
|
|
33 |
"chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\\n\\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + eos_token}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
|
34 |
"clean_up_tokenization_spaces": false,
|
35 |
"eos_token": "</s>",
|
|
|
36 |
"legacy": false,
|
37 |
"model_max_length": 512,
|
38 |
"pad_token": "<s>",
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e88839cfca7678064cc17eddd17af1a5d671f567213135ac508dc35eb02b4b18
|
3 |
+
size 5496
|