maxkretchmer commited on
Commit
0b2f3e1
1 Parent(s): c0a0a85

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -20,12 +20,12 @@
20
  "revision": null,
21
  "target_modules": [
22
  "down_proj",
23
- "lm_head",
24
  "q_proj",
 
25
  "o_proj",
 
26
  "gate_proj",
27
  "k_proj",
28
- "up_proj",
29
  "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM"
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Upstage/SOLAR-10.7B-Instruct-v1.0",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
20
  "revision": null,
21
  "target_modules": [
22
  "down_proj",
 
23
  "q_proj",
24
+ "lm_head",
25
  "o_proj",
26
+ "up_proj",
27
  "gate_proj",
28
  "k_proj",
 
29
  "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM"
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b93d3a77eb6848ddf0d5e68293b91ec0c2c5e188e05ce355c0125faf54040246
3
- size 609389712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2e5f3d328b1aa5a250555d56ef8e2c76f2d6c8c47a86565b7a29b8eef7572ae
3
+ size 651362912
runs/Dec14_21-55-39_e9f4ff0213cb/events.out.tfevents.1702590953.e9f4ff0213cb.540.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1984adaf4eab2425d431c232e98e73d67929941d1c0cce2d6c1d5ea55bcdb310
3
+ size 6549
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 2048,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 4096,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -29,15 +29,15 @@
29
  },
30
  "additional_special_tokens": [],
31
  "bos_token": "<s>",
32
- "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "legacy": true,
36
- "model_max_length": 2048,
37
  "pad_token": "</s>",
38
  "sp_model_kwargs": {},
39
  "spaces_between_special_tokens": false,
40
  "tokenizer_class": "LlamaTokenizer",
41
  "unk_token": "<unk>",
42
- "use_default_system_prompt": false
43
  }
 
29
  },
30
  "additional_special_tokens": [],
31
  "bos_token": "<s>",
32
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{% if message['content']%}{{'### System:\n' + message['content']+'\n\n'}}{% endif %}{% elif message['role'] == 'user' %}{{'### User:\n' + message['content']+'\n\n'}}{% elif message['role'] == 'assistant' %}{{'### Assistant:\n' + message['content']}}{% endif %}{% if loop.last and add_generation_prompt %}{{ '### Assistant:\n' }}{% endif %}{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "legacy": true,
36
+ "model_max_length": 4096,
37
  "pad_token": "</s>",
38
  "sp_model_kwargs": {},
39
  "spaces_between_special_tokens": false,
40
  "tokenizer_class": "LlamaTokenizer",
41
  "unk_token": "<unk>",
42
+ "use_default_system_prompt": true
43
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4236fa3e64a5d0348766df96696fd6ec7830dbf15a1ab898aff916eea7f0d1f
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4afa890dcddb41365b4343e28704035b60d83312f7c66e1583fd822688cdce8f
3
  size 4728