bharati2324 commited on
Commit
3aba268
·
verified ·
1 Parent(s): c3d9d06

Training in progress, step 200

Browse files
adapter_config.json CHANGED
@@ -3,6 +3,8 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "unsloth/Llama-3.2-1B-Instruct",
5
  "bias": "none",
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
@@ -11,6 +13,7 @@
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 16,
 
14
  "lora_dropout": 0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
@@ -20,9 +23,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "up_proj",
 
 
24
  "down_proj",
25
- "gate_proj"
 
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "unsloth/Llama-3.2-1B-Instruct",
5
  "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
  "fan_in_fan_out": false,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
 
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
  "lora_alpha": 16,
16
+ "lora_bias": false,
17
  "lora_dropout": 0,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
+ "o_proj",
28
+ "v_proj",
29
  "down_proj",
30
+ "up_proj",
31
+ "gate_proj",
32
+ "k_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f138d4a914741d4c09b97be606b477e4aacca167a7e2e496e5b86b29faf50e7
3
- size 31469800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6a35053aa45fdcee5deb77dd71015e46df182a8545535c67b8bf3129f415fe7
3
+ size 45118424
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d3ad0852eefb271603007aea78247e1549ac7c203d03b7476ccb0427c6570e0
3
- size 17210212
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
tokenizer_config.json CHANGED
@@ -2059,6 +2059,6 @@
2059
  ],
2060
  "model_max_length": 131072,
2061
  "pad_token": "<|finetune_right_pad_id|>",
2062
- "padding_side": "right",
2063
  "tokenizer_class": "PreTrainedTokenizerFast"
2064
  }
 
2059
  ],
2060
  "model_max_length": 131072,
2061
  "pad_token": "<|finetune_right_pad_id|>",
2062
+ "padding_side": "left",
2063
  "tokenizer_class": "PreTrainedTokenizerFast"
2064
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8d1652c64d5701d7542543217cc8eb0453db379d2bc34a961ad431eb6ebe7fc
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f519abf615386e0857c941fcb28a9140901798289aceaff057539afc5159bd3d
3
  size 5560