Farouk commited on
Commit
2adf302
·
1 Parent(s): d8c11b4

Training in progress, step 2400

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. adapter_model.bin +1 -1
  2. checkpoint-1000/adapter_model/adapter_model/README.md +0 -44
  3. checkpoint-1000/adapter_model/adapter_model/adapter_model.bin +0 -3
  4. checkpoint-1000/optimizer.pt +0 -3
  5. checkpoint-1000/rng_state.pth +0 -3
  6. checkpoint-1000/trainer_state.json +0 -0
  7. checkpoint-1200/adapter_config.json +0 -26
  8. checkpoint-1200/added_tokens.json +0 -3
  9. checkpoint-1200/rng_state.pth +0 -3
  10. checkpoint-1200/scheduler.pt +0 -3
  11. checkpoint-1200/special_tokens_map.json +0 -6
  12. checkpoint-1200/tokenizer.model +0 -3
  13. checkpoint-1200/tokenizer_config.json +0 -35
  14. checkpoint-1200/trainer_state.json +0 -0
  15. checkpoint-1200/training_args.bin +0 -3
  16. checkpoint-1400/README.md +0 -20
  17. checkpoint-1400/adapter_config.json +0 -26
  18. checkpoint-1400/adapter_model.bin +0 -3
  19. checkpoint-1400/added_tokens.json +0 -3
  20. checkpoint-1400/optimizer.pt +0 -3
  21. checkpoint-1400/scheduler.pt +0 -3
  22. checkpoint-1400/special_tokens_map.json +0 -6
  23. checkpoint-1400/tokenizer.model +0 -3
  24. checkpoint-1400/tokenizer_config.json +0 -35
  25. checkpoint-1400/trainer_state.json +0 -0
  26. checkpoint-1400/training_args.bin +0 -3
  27. checkpoint-1600/README.md +0 -20
  28. checkpoint-1600/adapter_config.json +0 -26
  29. checkpoint-1600/adapter_model.bin +0 -3
  30. checkpoint-1600/adapter_model/adapter_model/README.md +0 -44
  31. checkpoint-1600/adapter_model/adapter_model/adapter_config.json +0 -26
  32. checkpoint-1600/adapter_model/adapter_model/adapter_model.bin +0 -3
  33. checkpoint-1600/added_tokens.json +0 -3
  34. checkpoint-1600/optimizer.pt +0 -3
  35. checkpoint-1600/rng_state.pth +0 -3
  36. checkpoint-1600/scheduler.pt +0 -3
  37. checkpoint-1600/special_tokens_map.json +0 -6
  38. checkpoint-1600/tokenizer.model +0 -3
  39. checkpoint-1600/tokenizer_config.json +0 -35
  40. checkpoint-1600/trainer_state.json +0 -0
  41. checkpoint-1600/training_args.bin +0 -3
  42. checkpoint-1800/README.md +0 -20
  43. checkpoint-1800/adapter_config.json +0 -26
  44. checkpoint-1800/adapter_model.bin +0 -3
  45. checkpoint-1800/added_tokens.json +0 -3
  46. checkpoint-1800/optimizer.pt +0 -3
  47. checkpoint-1800/rng_state.pth +0 -3
  48. checkpoint-1800/scheduler.pt +0 -3
  49. checkpoint-1800/special_tokens_map.json +0 -6
  50. checkpoint-1800/tokenizer.model +0 -3
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18f11fbc4708b106870eec7154c2b9bbcad7ba5b185b2bacd1b7a7c4926deed7
3
  size 871609293
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23824721820d37d6fe44fee9306d0e71a5826aebaf3eb2f970cab6872288b55a
3
  size 871609293
checkpoint-1000/adapter_model/adapter_model/README.md DELETED
@@ -1,44 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: nf4
15
- - bnb_4bit_use_double_quant: True
16
- - bnb_4bit_compute_dtype: bfloat16
17
-
18
- The following `bitsandbytes` quantization config was used during training:
19
- - load_in_8bit: False
20
- - load_in_4bit: True
21
- - llm_int8_threshold: 6.0
22
- - llm_int8_skip_modules: None
23
- - llm_int8_enable_fp32_cpu_offload: False
24
- - llm_int8_has_fp16_weight: False
25
- - bnb_4bit_quant_type: nf4
26
- - bnb_4bit_use_double_quant: True
27
- - bnb_4bit_compute_dtype: bfloat16
28
-
29
- The following `bitsandbytes` quantization config was used during training:
30
- - load_in_8bit: False
31
- - load_in_4bit: True
32
- - llm_int8_threshold: 6.0
33
- - llm_int8_skip_modules: None
34
- - llm_int8_enable_fp32_cpu_offload: False
35
- - llm_int8_has_fp16_weight: False
36
- - bnb_4bit_quant_type: nf4
37
- - bnb_4bit_use_double_quant: True
38
- - bnb_4bit_compute_dtype: bfloat16
39
- ### Framework versions
40
-
41
- - PEFT 0.4.0
42
- - PEFT 0.4.0
43
-
44
- - PEFT 0.4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1000/adapter_model/adapter_model/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:80118c725477dea393f7c5d033e93b59658969b94a87d00f4bf43d4221785903
3
- size 871609293
 
 
 
 
checkpoint-1000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce3190f57432e0973d291995c12b59dd5194b3d8079b426d06cba01005b0c39f
3
- size 873873439
 
 
 
 
checkpoint-1000/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:eea4d2d84d0669c268f9b7375a0008cb8ca8f8c06d8427bb52a1d43b533afae1
3
- size 14511
 
 
 
 
checkpoint-1000/trainer_state.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-1200/adapter_config.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "auto_mapping": null,
3
- "base_model_name_or_path": "codellama/CodeLlama-34b-Python-hf",
4
- "bias": "none",
5
- "fan_in_fan_out": false,
6
- "inference_mode": true,
7
- "init_lora_weights": true,
8
- "layers_pattern": null,
9
- "layers_to_transform": null,
10
- "lora_alpha": 16.0,
11
- "lora_dropout": 0.1,
12
- "modules_to_save": null,
13
- "peft_type": "LORA",
14
- "r": 64,
15
- "revision": null,
16
- "target_modules": [
17
- "down_proj",
18
- "up_proj",
19
- "q_proj",
20
- "gate_proj",
21
- "o_proj",
22
- "v_proj",
23
- "k_proj"
24
- ],
25
- "task_type": "CAUSAL_LM"
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1200/added_tokens.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "[PAD]": 32000
3
- }
 
 
 
 
checkpoint-1200/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:44cdcbac175f6aff4ad243d928c84233d1d837142541c4dbdbd0e9ab4a9edc74
3
- size 14511
 
 
 
 
checkpoint-1200/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e303e329a219326eec4281908f48d8a4f3a6efdd9abb061adc0dc038af0e7e6
3
- size 627
 
 
 
 
checkpoint-1200/special_tokens_map.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "bos_token": "<s>",
3
- "eos_token": "</s>",
4
- "pad_token": "[PAD]",
5
- "unk_token": "<unk>"
6
- }
 
 
 
 
 
 
 
checkpoint-1200/tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723
 
 
 
 
checkpoint-1200/tokenizer_config.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "add_bos_token": true,
3
- "add_eos_token": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<s>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "clean_up_tokenization_spaces": false,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "</s>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "legacy": null,
22
- "model_max_length": 1000000000000000019884624838656,
23
- "pad_token": null,
24
- "padding_side": "right",
25
- "sp_model_kwargs": {},
26
- "tokenizer_class": "LlamaTokenizer",
27
- "unk_token": {
28
- "__type": "AddedToken",
29
- "content": "<unk>",
30
- "lstrip": false,
31
- "normalized": true,
32
- "rstrip": false,
33
- "single_word": false
34
- }
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1200/trainer_state.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-1200/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2493c95326c359fb00f88976162bc7966690beaaca22964b91c1db649a04988f
3
- size 6011
 
 
 
 
checkpoint-1400/README.md DELETED
@@ -1,20 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: nf4
15
- - bnb_4bit_use_double_quant: True
16
- - bnb_4bit_compute_dtype: bfloat16
17
- ### Framework versions
18
-
19
-
20
- - PEFT 0.4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1400/adapter_config.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "auto_mapping": null,
3
- "base_model_name_or_path": "codellama/CodeLlama-34b-Python-hf",
4
- "bias": "none",
5
- "fan_in_fan_out": false,
6
- "inference_mode": true,
7
- "init_lora_weights": true,
8
- "layers_pattern": null,
9
- "layers_to_transform": null,
10
- "lora_alpha": 16.0,
11
- "lora_dropout": 0.1,
12
- "modules_to_save": null,
13
- "peft_type": "LORA",
14
- "r": 64,
15
- "revision": null,
16
- "target_modules": [
17
- "down_proj",
18
- "up_proj",
19
- "q_proj",
20
- "gate_proj",
21
- "o_proj",
22
- "v_proj",
23
- "k_proj"
24
- ],
25
- "task_type": "CAUSAL_LM"
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1400/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:80118c725477dea393f7c5d033e93b59658969b94a87d00f4bf43d4221785903
3
- size 871609293
 
 
 
 
checkpoint-1400/added_tokens.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "[PAD]": 32000
3
- }
 
 
 
 
checkpoint-1400/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8ea3d806921abca9924806d7f6abd96f8dea29d950b57b47e1e75777d7fb3a8
3
- size 873873439
 
 
 
 
checkpoint-1400/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f07b5980373715ebdb21e0ee5240bdc9b0757226ab8995fd5ad862d2e1e5e0b
3
- size 627
 
 
 
 
checkpoint-1400/special_tokens_map.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "bos_token": "<s>",
3
- "eos_token": "</s>",
4
- "pad_token": "[PAD]",
5
- "unk_token": "<unk>"
6
- }
 
 
 
 
 
 
 
checkpoint-1400/tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723
 
 
 
 
checkpoint-1400/tokenizer_config.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "add_bos_token": true,
3
- "add_eos_token": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<s>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "clean_up_tokenization_spaces": false,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "</s>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "legacy": null,
22
- "model_max_length": 1000000000000000019884624838656,
23
- "pad_token": null,
24
- "padding_side": "right",
25
- "sp_model_kwargs": {},
26
- "tokenizer_class": "LlamaTokenizer",
27
- "unk_token": {
28
- "__type": "AddedToken",
29
- "content": "<unk>",
30
- "lstrip": false,
31
- "normalized": true,
32
- "rstrip": false,
33
- "single_word": false
34
- }
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1400/trainer_state.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-1400/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2493c95326c359fb00f88976162bc7966690beaaca22964b91c1db649a04988f
3
- size 6011
 
 
 
 
checkpoint-1600/README.md DELETED
@@ -1,20 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: nf4
15
- - bnb_4bit_use_double_quant: True
16
- - bnb_4bit_compute_dtype: bfloat16
17
- ### Framework versions
18
-
19
-
20
- - PEFT 0.4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1600/adapter_config.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "auto_mapping": null,
3
- "base_model_name_or_path": "codellama/CodeLlama-34b-Python-hf",
4
- "bias": "none",
5
- "fan_in_fan_out": false,
6
- "inference_mode": true,
7
- "init_lora_weights": true,
8
- "layers_pattern": null,
9
- "layers_to_transform": null,
10
- "lora_alpha": 16.0,
11
- "lora_dropout": 0.1,
12
- "modules_to_save": null,
13
- "peft_type": "LORA",
14
- "r": 64,
15
- "revision": null,
16
- "target_modules": [
17
- "down_proj",
18
- "up_proj",
19
- "q_proj",
20
- "gate_proj",
21
- "o_proj",
22
- "v_proj",
23
- "k_proj"
24
- ],
25
- "task_type": "CAUSAL_LM"
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1600/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b503ce7c2f5a638f100ba5da6d3bb1e96b10f33a707b306f142dcd7320296887
3
- size 871609293
 
 
 
 
checkpoint-1600/adapter_model/adapter_model/README.md DELETED
@@ -1,44 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: nf4
15
- - bnb_4bit_use_double_quant: True
16
- - bnb_4bit_compute_dtype: bfloat16
17
-
18
- The following `bitsandbytes` quantization config was used during training:
19
- - load_in_8bit: False
20
- - load_in_4bit: True
21
- - llm_int8_threshold: 6.0
22
- - llm_int8_skip_modules: None
23
- - llm_int8_enable_fp32_cpu_offload: False
24
- - llm_int8_has_fp16_weight: False
25
- - bnb_4bit_quant_type: nf4
26
- - bnb_4bit_use_double_quant: True
27
- - bnb_4bit_compute_dtype: bfloat16
28
-
29
- The following `bitsandbytes` quantization config was used during training:
30
- - load_in_8bit: False
31
- - load_in_4bit: True
32
- - llm_int8_threshold: 6.0
33
- - llm_int8_skip_modules: None
34
- - llm_int8_enable_fp32_cpu_offload: False
35
- - llm_int8_has_fp16_weight: False
36
- - bnb_4bit_quant_type: nf4
37
- - bnb_4bit_use_double_quant: True
38
- - bnb_4bit_compute_dtype: bfloat16
39
- ### Framework versions
40
-
41
- - PEFT 0.4.0
42
- - PEFT 0.4.0
43
-
44
- - PEFT 0.4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1600/adapter_model/adapter_model/adapter_config.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "auto_mapping": null,
3
- "base_model_name_or_path": "codellama/CodeLlama-34b-Python-hf",
4
- "bias": "none",
5
- "fan_in_fan_out": false,
6
- "inference_mode": true,
7
- "init_lora_weights": true,
8
- "layers_pattern": null,
9
- "layers_to_transform": null,
10
- "lora_alpha": 16.0,
11
- "lora_dropout": 0.1,
12
- "modules_to_save": null,
13
- "peft_type": "LORA",
14
- "r": 64,
15
- "revision": null,
16
- "target_modules": [
17
- "down_proj",
18
- "up_proj",
19
- "q_proj",
20
- "gate_proj",
21
- "o_proj",
22
- "v_proj",
23
- "k_proj"
24
- ],
25
- "task_type": "CAUSAL_LM"
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1600/adapter_model/adapter_model/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca7fae775be00cb7e472a75158d8ab72644b8579a83e290f3c7c2b2bc675e8dc
3
- size 871609293
 
 
 
 
checkpoint-1600/added_tokens.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "[PAD]": 32000
3
- }
 
 
 
 
checkpoint-1600/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f3b1009ddf180f8a75e4e4c7e70b09c050d575ea55a95be5113766472ac85f63
3
- size 873873439
 
 
 
 
checkpoint-1600/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:42456691be03d4e7f251ecd394f3a1db103fa9136263ad733e87ee530de3678e
3
- size 14511
 
 
 
 
checkpoint-1600/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a6ae13f0c7f359a6b405d50e69129c1f2abc4f10fc1e47508750a1aa5936c11
3
- size 627
 
 
 
 
checkpoint-1600/special_tokens_map.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "bos_token": "<s>",
3
- "eos_token": "</s>",
4
- "pad_token": "[PAD]",
5
- "unk_token": "<unk>"
6
- }
 
 
 
 
 
 
 
checkpoint-1600/tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723
 
 
 
 
checkpoint-1600/tokenizer_config.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "add_bos_token": true,
3
- "add_eos_token": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<s>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "clean_up_tokenization_spaces": false,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "</s>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "legacy": null,
22
- "model_max_length": 1000000000000000019884624838656,
23
- "pad_token": null,
24
- "padding_side": "right",
25
- "sp_model_kwargs": {},
26
- "tokenizer_class": "LlamaTokenizer",
27
- "unk_token": {
28
- "__type": "AddedToken",
29
- "content": "<unk>",
30
- "lstrip": false,
31
- "normalized": true,
32
- "rstrip": false,
33
- "single_word": false
34
- }
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1600/trainer_state.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-1600/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2493c95326c359fb00f88976162bc7966690beaaca22964b91c1db649a04988f
3
- size 6011
 
 
 
 
checkpoint-1800/README.md DELETED
@@ -1,20 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: nf4
15
- - bnb_4bit_use_double_quant: True
16
- - bnb_4bit_compute_dtype: bfloat16
17
- ### Framework versions
18
-
19
-
20
- - PEFT 0.4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1800/adapter_config.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "auto_mapping": null,
3
- "base_model_name_or_path": "codellama/CodeLlama-34b-Python-hf",
4
- "bias": "none",
5
- "fan_in_fan_out": false,
6
- "inference_mode": true,
7
- "init_lora_weights": true,
8
- "layers_pattern": null,
9
- "layers_to_transform": null,
10
- "lora_alpha": 16.0,
11
- "lora_dropout": 0.1,
12
- "modules_to_save": null,
13
- "peft_type": "LORA",
14
- "r": 64,
15
- "revision": null,
16
- "target_modules": [
17
- "down_proj",
18
- "up_proj",
19
- "q_proj",
20
- "gate_proj",
21
- "o_proj",
22
- "v_proj",
23
- "k_proj"
24
- ],
25
- "task_type": "CAUSAL_LM"
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-1800/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c61e805254e0a600e748d930ef8c72bc26fcde3b479c3cfa35931c18f619f86
3
- size 871609293
 
 
 
 
checkpoint-1800/added_tokens.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "[PAD]": 32000
3
- }
 
 
 
 
checkpoint-1800/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:19d9bbef27f39dbe487933f28d7202a46a46790a115df33e78e6bd1954a418f9
3
- size 873873439
 
 
 
 
checkpoint-1800/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3a9d5f030044d7c1c38e9af8d159736e36882df4a51d621092d38392dff14ee
3
- size 14511
 
 
 
 
checkpoint-1800/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b5171ec9320c39e7e6817f1736e6a4c3dc5a7837db85e0bdb59924fe3c72ecf
3
- size 627
 
 
 
 
checkpoint-1800/special_tokens_map.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "bos_token": "<s>",
3
- "eos_token": "</s>",
4
- "pad_token": "[PAD]",
5
- "unk_token": "<unk>"
6
- }
 
 
 
 
 
 
 
checkpoint-1800/tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723