IvanSSY615 commited on
Commit
af7bea5
1 Parent(s): e40d695

End of training

Browse files
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
- base_model: NousResearch/Llama-2-7b-chat-hf
 
3
  tags:
4
  - trl
5
  - sft
@@ -14,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # results
16
 
17
- This model is a fine-tuned version of [NousResearch/Llama-2-7b-chat-hf](https://huggingface.co/NousResearch/Llama-2-7b-chat-hf) on an unknown dataset.
18
 
19
  ## Model description
20
 
@@ -34,7 +35,7 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0002
37
- - train_batch_size: 4
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
@@ -50,5 +51,5 @@ The following hyperparameters were used during training:
50
 
51
  - Transformers 4.35.2
52
  - Pytorch 2.1.0+cu121
53
- - Datasets 2.16.1
54
- - Tokenizers 0.15.1
 
1
  ---
2
+ license: openrail
3
+ base_model: LinkSoul/Chinese-Llama-2-7b
4
  tags:
5
  - trl
6
  - sft
 
15
 
16
  # results
17
 
18
+ This model is a fine-tuned version of [LinkSoul/Chinese-Llama-2-7b](https://huggingface.co/LinkSoul/Chinese-Llama-2-7b) on an unknown dataset.
19
 
20
  ## Model description
21
 
 
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 0.0002
38
+ - train_batch_size: 1
39
  - eval_batch_size: 8
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
51
 
52
  - Transformers 4.35.2
53
  - Pytorch 2.1.0+cu121
54
+ - Datasets 2.17.0
55
+ - Tokenizers 0.15.2
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "NousResearch/Llama-2-7b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -19,8 +19,9 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "v_proj",
23
- "q_proj"
24
  ],
25
- "task_type": "CAUSAL_LM"
 
26
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "LinkSoul/Chinese-Llama-2-7b",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "q_proj",
23
+ "v_proj"
24
  ],
25
+ "task_type": "CAUSAL_LM",
26
+ "use_rslora": false
27
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:919313a9324465d46985f5ef0c8df1f72cd1e9698b4bbb0db0bc352008aad203
3
  size 134235048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5511dbcd638e6ce8afc3a7929bb7ec2124e8addb957f8c199583f322cee3ba19
3
  size 134235048
runs/Feb16_15-23-20_8d934c220c4e/events.out.tfevents.1708097011.8d934c220c4e.3396.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36f8959fedd4d8bcf9cd80d72716f6e479449bfb335567b398f29d5b333fda3b
3
+ size 5275
runs/Feb16_15-24-31_8d934c220c4e/events.out.tfevents.1708097078.8d934c220c4e.3396.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:265a4a412ff7bf83e77f8d3ec4d9b28de1a5f3ed3eb2e73ed723b51effb0c818
3
+ size 5583
tokenizer.json CHANGED
@@ -34,15 +34,6 @@
34
  "rstrip": false,
35
  "normalized": true,
36
  "special": true
37
- },
38
- {
39
- "id": 32000,
40
- "content": "<pad>",
41
- "single_word": false,
42
- "lstrip": false,
43
- "rstrip": false,
44
- "normalized": true,
45
- "special": false
46
  }
47
  ],
48
  "normalizer": {
 
34
  "rstrip": false,
35
  "normalized": true,
36
  "special": true
 
 
 
 
 
 
 
 
 
37
  }
38
  ],
39
  "normalizer": {
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<unk>",
@@ -23,23 +25,17 @@
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
- },
27
- "32000": {
28
- "content": "<pad>",
29
- "lstrip": false,
30
- "normalized": true,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": false
34
  }
35
  },
36
  "bos_token": "<s>",
37
  "clean_up_tokenization_spaces": false,
38
  "eos_token": "</s>",
39
- "legacy": false,
40
- "model_max_length": 1000000000000000019884624838656,
41
  "pad_token": "</s>",
 
42
  "sp_model_kwargs": {},
 
43
  "tokenizer_class": "LlamaTokenizer",
44
  "unk_token": "<unk>",
45
  "use_default_system_prompt": false
 
1
  {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
 
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
 
 
 
 
 
 
 
 
28
  }
29
  },
30
  "bos_token": "<s>",
31
  "clean_up_tokenization_spaces": false,
32
  "eos_token": "</s>",
33
+ "legacy": true,
34
+ "model_max_length": 4096,
35
  "pad_token": "</s>",
36
+ "padding_side": "right",
37
  "sp_model_kwargs": {},
38
+ "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",
40
  "unk_token": "<unk>",
41
  "use_default_system_prompt": false
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e06744f8b5dc550acdefdb3937fd93d46bd3628a7e8e2c1b50f94b95297e8d7f
3
  size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93775b138c851ea56aaf11434d2f43b05dd0c1b3961a877baa1430980145c808
3
  size 4536