bilkultheek commited on
Commit
5d5f3b1
·
verified ·
1 Parent(s): 145feab

Model save

Browse files
Files changed (1) hide show
  1. README.md +14 -9
README.md CHANGED
@@ -1,6 +1,7 @@
1
  ---
2
- base_model: NousResearch/Llama-2-7b-hf
3
  library_name: peft
 
4
  tags:
5
  - trl
6
  - sft
@@ -15,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # Cold-Data-LLama-2-7B
17
 
18
- This model is a fine-tuned version of [NousResearch/Llama-2-7b-hf](https://huggingface.co/NousResearch/Llama-2-7b-hf) on the None dataset.
19
 
20
  ## Model description
21
 
@@ -35,20 +36,24 @@ More information needed
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 2e-05
38
- - train_batch_size: 16
39
- - eval_batch_size: 16
40
  - seed: 42
41
  - gradient_accumulation_steps: 4
42
- - total_train_batch_size: 64
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: cosine
45
  - lr_scheduler_warmup_ratio: 0.03
46
- - num_epochs: 2
 
 
 
 
47
 
48
  ### Framework versions
49
 
50
  - PEFT 0.12.0
51
- - Transformers 4.43.3
52
- - Pytorch 2.3.1+cu121
53
- - Datasets 2.17.0
54
  - Tokenizers 0.19.1
 
1
  ---
2
+ base_model: meta-llama/Llama-2-7b-hf
3
  library_name: peft
4
+ license: llama2
5
  tags:
6
  - trl
7
  - sft
 
16
 
17
  # Cold-Data-LLama-2-7B
18
 
19
+ This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the None dataset.
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 2e-05
39
+ - train_batch_size: 12
40
+ - eval_batch_size: 12
41
  - seed: 42
42
  - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 48
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.03
47
+ - num_epochs: 1
48
+
49
+ ### Training results
50
+
51
+
52
 
53
  ### Framework versions
54
 
55
  - PEFT 0.12.0
56
+ - Transformers 4.44.2
57
+ - Pytorch 2.4.1+cu121
58
+ - Datasets 2.21.0
59
  - Tokenizers 0.19.1