Sam137 commited on
Commit
7106c49
·
verified ·
1 Parent(s): 4389446

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -38,7 +38,7 @@ The following hyperparameters were used during training:
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-08
40
  - lr_scheduler_type: cosine
41
- - lr_scheduler_warmup_steps: 30
42
  - num_epochs: 1
43
  - mixed_precision_training: Native AMP
44
 
 
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-08
40
  - lr_scheduler_type: cosine
41
+ - lr_scheduler_warmup_steps: 200
42
  - num_epochs: 1
43
  - mixed_precision_training: Native AMP
44
 
adapter_config.json CHANGED
@@ -16,13 +16,13 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "v_proj",
20
- "down_proj",
21
- "q_proj",
22
  "o_proj",
23
- "k_proj",
24
  "gate_proj",
25
- "up_proj"
 
 
 
26
  ],
27
  "task_type": "CAUSAL_LM"
28
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "up_proj",
 
 
20
  "o_proj",
 
21
  "gate_proj",
22
+ "k_proj",
23
+ "v_proj",
24
+ "down_proj",
25
+ "q_proj"
26
  ],
27
  "task_type": "CAUSAL_LM"
28
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9b29b82d33a16ab23ea3f1ce9e3fa9c8eb77125061d5c53b6ed9d8cf783cbc3
3
  size 80013120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cd983ec4e34e31c66028a370aad768448bf4bf0f3f43f1b29b2cc62e32e9f82
3
  size 80013120
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0801a6eeff3d52765aa1ec4bf24721276bacd85d4998f9d0a8dea2b34ec8afba
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:add4c1b8058597a3a6d77d887369a731b4259a245b81f1b1da63f72d6edbb6e5
3
  size 4600