chohi commited on
Commit
ca41f2e
·
verified ·
1 Parent(s): 0378ba2

chohi/Llama-test-Ko-3-8B

Browse files
README.md CHANGED
@@ -36,8 +36,8 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
- - train_batch_size: 4
40
- - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: constant
@@ -50,8 +50,8 @@ The following hyperparameters were used during training:
50
 
51
  ### Framework versions
52
 
53
- - PEFT 0.10.0
54
- - Transformers 4.40.1
55
- - Pytorch 2.2.1+cu121
56
- - Datasets 2.19.0
57
- - Tokenizers 0.19.1
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
+ - train_batch_size: 16
40
+ - eval_batch_size: 32
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: constant
 
50
 
51
  ### Framework versions
52
 
53
+ - PEFT 0.8.2
54
+ - Transformers 4.37.2
55
+ - Pytorch 2.3.0+cu121
56
+ - Datasets 2.19.1
57
+ - Tokenizers 0.15.2
adapter_config.json CHANGED
@@ -6,7 +6,6 @@
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
- "layer_replication": null,
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
@@ -20,10 +19,9 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
- "use_dora": false,
28
  "use_rslora": false
29
  }
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "v_proj",
23
+ "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM",
 
26
  "use_rslora": false
27
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d9b5666cb86aa361b546261210825407a64e7b345d06ef9a4130e460f37ef41
3
  size 109069176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6275229f94e2fb49cb663256ae3d0bc2a97e4b1afca817823ac8db0160096e09
3
  size 109069176
runs/May09_11-39-13_cbnu-DGX-Station/events.out.tfevents.1715222380.cbnu-DGX-Station.2148635.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6fdc29a9bb2b44d685308aceaa9a5e206a875b155fa704596c523e9cd3c9586
3
+ size 5131
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
@@ -2407,7 +2412,6 @@
2407
  "end_of_word_suffix": null,
2408
  "fuse_unk": false,
2409
  "byte_fallback": false,
2410
- "ignore_merges": true,
2411
  "vocab": {
2412
  "!": 0,
2413
  "\"": 1,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 1024,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
2412
  "end_of_word_suffix": null,
2413
  "fuse_unk": false,
2414
  "byte_fallback": false,
 
2415
  "vocab": {
2416
  "!": 0,
2417
  "\"": 1,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4411bed28119bf1d109631a50f48615e4cea0882da43936974a8109076f24f51
3
- size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1246dbc4e8cc43cd2c44dab983a1fe34be1a732bb8708f8fbdee815a0f6a477d
3
+ size 4664