yongjing commited on
Commit
7242d28
1 Parent(s): cb7c56d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -41
README.md CHANGED
@@ -11,46 +11,6 @@ should probably proofread and complete it, then remove this comment. -->
11
 
12
  # dezhru7B_dc_shot3
13
 
14
- This model is a fine-tuned version of [/yinyongjing/hfmodels/llama2-7b](https://huggingface.co//yinyongjing/hfmodels/llama2-7b) on the None dataset.
15
 
16
- ## Model description
17
 
18
- More information needed
19
-
20
- ## Intended uses & limitations
21
-
22
- More information needed
23
-
24
- ## Training and evaluation data
25
-
26
- More information needed
27
-
28
- ## Training procedure
29
-
30
- ### Training hyperparameters
31
-
32
- The following hyperparameters were used during training:
33
- - learning_rate: 2e-05
34
- - train_batch_size: 8
35
- - eval_batch_size: 1
36
- - seed: 42
37
- - distributed_type: multi-GPU
38
- - num_devices: 8
39
- - gradient_accumulation_steps: 2
40
- - total_train_batch_size: 128
41
- - total_eval_batch_size: 8
42
- - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-08
43
- - lr_scheduler_type: linear
44
- - lr_scheduler_warmup_ratio: 0.03
45
- - num_epochs: 1.0
46
-
47
- ### Training results
48
-
49
-
50
-
51
- ### Framework versions
52
-
53
- - Transformers 4.29.0
54
- - Pytorch 1.13.1+cu116
55
- - Datasets 2.11.0
56
- - Tokenizers 0.13.3
 
11
 
12
  # dezhru7B_dc_shot3
13
 
14
+ This model is a fine-tuned version of LLaMA2-7B on the collected translation dataset.
15
 
 
16