kazuma313 commited on
Commit
aa87bcd
1 Parent(s): 093505b

Training in progress, step 100

Browse files
adapter_config.json CHANGED
@@ -20,12 +20,12 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
  "o_proj",
25
  "q_proj",
26
- "k_proj",
27
  "gate_proj",
28
- "up_proj"
 
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "o_proj",
24
  "q_proj",
25
+ "up_proj",
26
  "gate_proj",
27
+ "v_proj",
28
+ "k_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c436a8ca53d75656908fcf2bc7adb31c67ef3a14d3f7ee3327004b6fad22e30f
3
- size 14331480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9ec77cc21dd36b60098ac0eb34724bbd1f3b8f850e2d8eaf4f4c48c57b00fba
3
+ size 28634880
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:022615eabe3e14a33ad7c9fe61dad3e9b28d14ea60f4a50a206b529106050074
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac1606d786e71c73273cbd2292ea248042c26d7927343d31cfce86c5c6ae5acc
3
  size 4920