nekokiku commited on
Commit
9cf3bbe
·
verified ·
1 Parent(s): 88735f4

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "microsoft/Phi-3-mini-128k-instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -10,24 +10,20 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 40,
14
- "lora_dropout": 0.15,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 32,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "gate_proj",
25
- "up_proj",
26
- "ffn",
27
  "down_proj",
28
- "k_proj",
29
- "v_proj",
30
- "o_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "microsoft/Phi-3-mini-4k-instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 256,
14
+ "lora_dropout": 0.1,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 128,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "o_proj",
 
 
 
24
  "down_proj",
25
+ "qkv_proj",
26
+ "gate_up_proj"
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0bc223af07969332c68354762218f7da0fde1436d9ebcad335d2fdeba4de779
3
- size 71320216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08b5f48cf2159bf7ccaf9eb3ad25047975a45aa915342d5065d3a86c6e61d49b
3
+ size 805341552
tokenizer_config.json CHANGED
@@ -121,7 +121,7 @@
121
  "clean_up_tokenization_spaces": false,
122
  "eos_token": "<|endoftext|>",
123
  "legacy": false,
124
- "model_max_length": 131072,
125
  "pad_token": "<|endoftext|>",
126
  "padding_side": "left",
127
  "sp_model_kwargs": {},
 
121
  "clean_up_tokenization_spaces": false,
122
  "eos_token": "<|endoftext|>",
123
  "legacy": false,
124
+ "model_max_length": 4096,
125
  "pad_token": "<|endoftext|>",
126
  "padding_side": "left",
127
  "sp_model_kwargs": {},
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74728852608f40194c0b70ec5983537a848f5644bbd3cd14dd654b7671cf4f80
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c9bfafd39be5491618f998465d2cb3e2f158ed9b5dbe0d1f4ef69aa7c2c33c1
3
  size 5432