Ashwinatgsk commited on
Commit
707f14a
·
verified ·
1 Parent(s): 8b6ca02

mistral_7b_admiral_instruct

Browse files
README.md CHANGED
@@ -20,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.2638
24
 
25
  ## Model description
26
 
@@ -52,11 +52,11 @@ The following hyperparameters were used during training:
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-----:|:----:|:---------------:|
55
- | 1.6092 | 0.16 | 20 | 1.4551 |
56
- | 1.4237 | 0.33 | 40 | 1.3019 |
57
- | 1.3399 | 0.49 | 60 | 1.2806 |
58
- | 1.3411 | 0.65 | 80 | 1.2683 |
59
- | 1.3916 | 0.81 | 100 | 1.2638 |
60
 
61
 
62
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.2637
24
 
25
  ## Model description
26
 
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-----:|:----:|:---------------:|
55
+ | 1.5871 | 0.16 | 20 | 1.4508 |
56
+ | 1.4009 | 0.33 | 40 | 1.2985 |
57
+ | 1.3499 | 0.49 | 60 | 1.2842 |
58
+ | 1.3251 | 0.65 | 80 | 1.2721 |
59
+ | 1.3741 | 0.81 | 100 | 1.2637 |
60
 
61
 
62
  ### Framework versions
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27a3e547388fa34cebc8f54aad55f406d97eb9cce3995d50753d2dfb2fbac5e8
3
  size 109069176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61bc96c4e810998535ff7a8c957b0f5916ff3e922a2ca3eae05138e1c7f1b33f
3
  size 109069176
tokenizer_config.json CHANGED
@@ -29,6 +29,7 @@
29
  },
30
  "additional_special_tokens": [],
31
  "bos_token": "<s>",
 
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
  "legacy": true,
 
29
  },
30
  "additional_special_tokens": [],
31
  "bos_token": "<s>",
32
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "legacy": true,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:effdc6d67ef82527a86815f254ba224832da83cca2cdc74dbc32d245a88824bd
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e795171346d7e37928821c285368490b74f64ec42779607dd3623f3663b97278
3
  size 4920