VictorDCh commited on
Commit
807ceaf
·
verified ·
1 Parent(s): 06fdfa8

Model save

Browse files
README.md CHANGED
@@ -1,13 +1,13 @@
1
  ---
2
- license: llama3
 
 
3
  library_name: peft
 
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
- base_model: meta-llama/Meta-Llama-3-8B-Instruct
9
- datasets:
10
- - generator
11
  model-index:
12
  - name: Llama-3-8B-Instruct-2-spider-3
13
  results: []
@@ -54,8 +54,8 @@ The following hyperparameters were used during training:
54
 
55
  ### Framework versions
56
 
57
- - PEFT 0.7.2.dev0
58
- - Transformers 4.36.2
59
- - Pytorch 2.1.2+cu121
60
  - Datasets 2.16.1
61
- - Tokenizers 0.15.2
 
1
  ---
2
+ base_model: meta-llama/Meta-Llama-3-8B-Instruct
3
+ datasets:
4
+ - generator
5
  library_name: peft
6
+ license: llama3
7
  tags:
8
  - trl
9
  - sft
10
  - generated_from_trainer
 
 
 
11
  model-index:
12
  - name: Llama-3-8B-Instruct-2-spider-3
13
  results: []
 
54
 
55
  ### Framework versions
56
 
57
+ - PEFT 0.10.0
58
+ - Transformers 4.40.1
59
+ - Pytorch 2.3.1+cu121
60
  - Datasets 2.16.1
61
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -6,6 +6,7 @@
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
@@ -19,14 +20,15 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "q_proj",
23
  "v_proj",
 
24
  "down_proj",
 
25
  "o_proj",
26
- "k_proj",
27
- "gate_proj",
28
- "up_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
 
31
  "use_rslora": false
32
  }
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
+ "layer_replication": null,
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "up_proj",
24
  "v_proj",
25
+ "gate_proj",
26
  "down_proj",
27
+ "q_proj",
28
  "o_proj",
29
+ "k_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
  "use_rslora": false
34
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35651fd2d4d77fc8198174c3b77ad985612f98af5d45b637679968949dcd2ebf
3
  size 3443619040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cb34adda599bf72c99dcc3823230581a6839a00de4af4b6b07791e8f6306638
3
  size 3443619040
runs/Jul03_11-53-33_664dd173faa0/events.out.tfevents.1720000432.664dd173faa0.97513.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a58b156ed881594e1471b93ab1a38f70e29901778070d6e5b85b976023d870c1
3
+ size 27297
special_tokens_map.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
  "additional_special_tokens": [
3
  {
4
- "content": "<|im_end|>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false
9
  },
10
  {
11
- "content": "<|im_start|>",
12
  "lstrip": false,
13
  "normalized": false,
14
  "rstrip": false,
 
1
  {
2
  "additional_special_tokens": [
3
  {
4
+ "content": "<|im_start|>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false
9
  },
10
  {
11
+ "content": "<|im_end|>",
12
  "lstrip": false,
13
  "normalized": false,
14
  "rstrip": false,
tokenizer.json CHANGED
@@ -2309,7 +2309,7 @@
2309
  },
2310
  {
2311
  "id": 128256,
2312
- "content": "<|im_end|>",
2313
  "single_word": false,
2314
  "lstrip": false,
2315
  "rstrip": false,
@@ -2318,7 +2318,7 @@
2318
  },
2319
  {
2320
  "id": 128257,
2321
- "content": "<|im_start|>",
2322
  "single_word": false,
2323
  "lstrip": false,
2324
  "rstrip": false,
@@ -2425,6 +2425,7 @@
2425
  "end_of_word_suffix": null,
2426
  "fuse_unk": false,
2427
  "byte_fallback": false,
 
2428
  "vocab": {
2429
  "!": 0,
2430
  "\"": 1,
 
2309
  },
2310
  {
2311
  "id": 128256,
2312
+ "content": "<|im_start|>",
2313
  "single_word": false,
2314
  "lstrip": false,
2315
  "rstrip": false,
 
2318
  },
2319
  {
2320
  "id": 128257,
2321
+ "content": "<|im_end|>",
2322
  "single_word": false,
2323
  "lstrip": false,
2324
  "rstrip": false,
 
2425
  "end_of_word_suffix": null,
2426
  "fuse_unk": false,
2427
  "byte_fallback": false,
2428
+ "ignore_merges": true,
2429
  "vocab": {
2430
  "!": 0,
2431
  "\"": 1,
tokenizer_config.json CHANGED
@@ -2049,7 +2049,7 @@
2049
  "special": true
2050
  },
2051
  "128256": {
2052
- "content": "<|im_end|>",
2053
  "lstrip": false,
2054
  "normalized": false,
2055
  "rstrip": false,
@@ -2057,7 +2057,7 @@
2057
  "special": true
2058
  },
2059
  "128257": {
2060
- "content": "<|im_start|>",
2061
  "lstrip": false,
2062
  "normalized": false,
2063
  "rstrip": false,
@@ -2066,8 +2066,8 @@
2066
  }
2067
  },
2068
  "additional_special_tokens": [
2069
- "<|im_end|>",
2070
- "<|im_start|>"
2071
  ],
2072
  "bos_token": "<|im_start|>",
2073
  "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
 
2049
  "special": true
2050
  },
2051
  "128256": {
2052
+ "content": "<|im_start|>",
2053
  "lstrip": false,
2054
  "normalized": false,
2055
  "rstrip": false,
 
2057
  "special": true
2058
  },
2059
  "128257": {
2060
+ "content": "<|im_end|>",
2061
  "lstrip": false,
2062
  "normalized": false,
2063
  "rstrip": false,
 
2066
  }
2067
  },
2068
  "additional_special_tokens": [
2069
+ "<|im_start|>",
2070
+ "<|im_end|>"
2071
  ],
2072
  "bos_token": "<|im_start|>",
2073
  "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc8698391b9185037ebe8916b12c3ed7bc163284870b4501d4ac452f73b61559
3
- size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eaeedb4d19a2abdc486408f425e44a2526e6f8e13403bc4201ea5faef8c6804f
3
+ size 5048