acrobatlm commited on
Commit
e7eb0ea
·
verified ·
1 Parent(s): 9ed40f3

acrobatlm/mistral-7binstruct-summary-100s

Browse files
README.md CHANGED
@@ -20,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.4779
24
 
25
  ## Model description
26
 
@@ -50,16 +50,16 @@ The following hyperparameters were used during training:
50
 
51
  ### Training results
52
 
53
- | Training Loss | Epoch | Step | Validation Loss |
54
- |:-------------:|:-----:|:----:|:---------------:|
55
- | 1.6991 | 0.22 | 25 | 1.5629 |
56
- | 1.5556 | 0.43 | 50 | 1.4779 |
57
 
58
 
59
  ### Framework versions
60
 
61
  - PEFT 0.10.0
62
- - Transformers 4.39.3
63
  - Pytorch 2.2.1+cu121
64
  - Datasets 2.18.0
65
- - Tokenizers 0.15.2
 
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.5011
24
 
25
  ## Model description
26
 
 
50
 
51
  ### Training results
52
 
53
+ | Training Loss | Epoch | Step | Validation Loss |
54
+ |:-------------:|:------:|:----:|:---------------:|
55
+ | 1.7176 | 0.2212 | 25 | 1.5722 |
56
+ | 1.6267 | 0.4425 | 50 | 1.5011 |
57
 
58
 
59
  ### Framework versions
60
 
61
  - PEFT 0.10.0
62
+ - Transformers 4.40.0
63
  - Pytorch 2.2.1+cu121
64
  - Datasets 2.18.0
65
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e44ce263e6fd885f50d82ca515b9325375b43ee36ededb75acf161ce88bc2e41
3
- size 48
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f57bef3246c59c405878b3b707cfc3831e31b553c1484785fd7e68070267572
3
+ size 27280152
runs/Apr18_22-54-02_18e887ddd35f/events.out.tfevents.1713480856.18e887ddd35f.760.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a909060915a57b1724efc804d7aae7422b08a9d54a4788ee9cdf78c170ee2f4
3
+ size 7113
tokenizer.json CHANGED
@@ -134,6 +134,7 @@
134
  "end_of_word_suffix": null,
135
  "fuse_unk": true,
136
  "byte_fallback": true,
 
137
  "vocab": {
138
  "<unk>": 0,
139
  "<s>": 1,
 
134
  "end_of_word_suffix": null,
135
  "fuse_unk": true,
136
  "byte_fallback": true,
137
+ "ignore_merges": false,
138
  "vocab": {
139
  "<unk>": 0,
140
  "<s>": 1,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1db73b3e4c5624152527ba9989b5aa678eb274c40f26ac111b55a0540e66b8d9
3
- size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8efaceddd2df97653d20329fcbef2324da8fcd9963392ed719431681c04db7f
3
+ size 4984