bilkultheek commited on
Commit
8e6a29d
·
verified ·
1 Parent(s): d96edbb

Training in progress, step 20

Browse files
adapter_config.json CHANGED
@@ -23,9 +23,9 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "q_proj",
27
  "v_proj",
28
- "k_proj"
 
29
  ],
30
  "task_type": "SEQ_CLS",
31
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "v_proj",
27
+ "k_proj",
28
+ "q_proj"
29
  ],
30
  "task_type": "SEQ_CLS",
31
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5fc74c130cc3821a3d8495c3159785534eaaff039a097b128ae85a1936302e3
3
  size 201385536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1ec785e5857d6334e5e900278236376037cec41b039bf8b7e537e40665e924d
3
  size 201385536
tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 900,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 900
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 830,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 830
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:783e55903e48a6add85486de184e79cc55a810d2614ec9d61080fc82d143541d
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bde348f58cef27e2fd1d65915e1be074abc230acaa173e17ccd6a145e3a8f38
3
  size 5432