CHZY-1 commited on
Commit
f08e6d3
·
verified ·
1 Parent(s): 05a54ac

Re-trained QLora Adapter with 260 data (5 epoch)

Browse files
README.md CHANGED
@@ -53,8 +53,8 @@ The following hyperparameters were used during training:
53
 
54
  ### Framework versions
55
 
56
- - PEFT 0.13.0
57
- - Transformers 4.45.1
58
  - Pytorch 2.4.1+cu121
59
  - Datasets 3.0.1
60
- - Tokenizers 0.20.0
 
53
 
54
  ### Framework versions
55
 
56
+ - PEFT 0.13.2
57
+ - Transformers 4.45.2
58
  - Pytorch 2.4.1+cu121
59
  - Datasets 3.0.1
60
+ - Tokenizers 0.20.1
adapter_config.json CHANGED
@@ -16,7 +16,7 @@
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 16,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 64,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a970cc1fd0b2a707748f41aed2540f761b682b52eda5a4237ce4e58bccefb1b
3
- size 33571624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:661f55bd9eb4481c5f7a7184c7eeade2c372c6aabc24f5c9d34b4a2392d4d104
3
+ size 134235048
runs/Oct20_11-03-28_c06fa3a5df24/events.out.tfevents.1729422227.c06fa3a5df24.5296.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd143292d17a0c96ae378531c7a4ab6ce7a2166acae4b3b51f9f1bf9a9d7e20e
3
+ size 8752
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84d8b2d24b4bcdd742706fced1648a7ff5a29daef93bca22b7c4feecf2ac49f9
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:979ac9c98655e77d2a9a425d077f28e51b2e3ab00acb9c79bb029c2575e87ea9
3
  size 5496