Sam137 commited on
Commit
d920bfb
·
verified ·
1 Parent(s): 42bafe8

Model save

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: other
3
- base_model: deepseek-ai/deepseek-coder-6.7b-base
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # deepseek6.7-compare-coder
15
 
16
- This model is a fine-tuned version of [deepseek-ai/deepseek-coder-6.7b-base](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-base) on an unknown dataset.
17
 
18
  ## Model description
19
 
@@ -37,9 +37,9 @@ The following hyperparameters were used during training:
37
  - eval_batch_size: 20
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-08
40
- - lr_scheduler_type: cosine_with_restarts
41
  - lr_scheduler_warmup_steps: 100
42
- - num_epochs: 2
43
  - mixed_precision_training: Native AMP
44
 
45
  ### Framework versions
 
1
  ---
2
  license: other
3
+ base_model: deepseek-ai/deepseek-coder-6.7b-instruct
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # deepseek6.7-compare-coder
15
 
16
+ This model is a fine-tuned version of [deepseek-ai/deepseek-coder-6.7b-instruct](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct) on an unknown dataset.
17
 
18
  ## Model description
19
 
 
37
  - eval_batch_size: 20
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-08
40
+ - lr_scheduler_type: cosine
41
  - lr_scheduler_warmup_steps: 100
42
+ - num_epochs: 1
43
  - mixed_precision_training: Native AMP
44
 
45
  ### Framework versions