fals3 commited on
Commit
ad8004f
·
verified ·
1 Parent(s): 1060323

Delete .ipynb_checkpoints/README-checkpoint.md

Browse files
.ipynb_checkpoints/README-checkpoint.md DELETED
@@ -1,76 +0,0 @@
1
- ---
2
- license: bigcode-openrail-m
3
- base_model: bigcode/starcoderbase
4
- tags:
5
- - generated_from_trainer
6
- datasets:
7
- - fals3/methods2test_small
8
- metrics:
9
- - accuracy
10
- model-index:
11
- - name: output
12
- results:
13
- - task:
14
- name: Causal Language Modeling
15
- type: text-generation
16
- dataset:
17
- name: fals3/methods2test_small fm+fc+c+m+f+t+tc
18
- type: fals3/methods2test_small
19
- args: fm+fc+c+m+f+t+tc
20
- metrics:
21
- - name: Accuracy
22
- type: accuracy
23
- value: 0.5611769226558302
24
- ---
25
-
26
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
27
- should probably proofread and complete it, then remove this comment. -->
28
-
29
- # output
30
-
31
- This model is a fine-tuned version of [bigcode/starcoderbase](https://huggingface.co/bigcode/starcoderbase) on the fals3/methods2test_small fm+fc+c+m+f+t+tc dataset.
32
- It achieves the following results on the evaluation set:
33
- - Loss: 0.6457
34
- - Accuracy: 0.5612
35
-
36
- ## Model description
37
-
38
- More information needed
39
-
40
- ## Intended uses & limitations
41
-
42
- More information needed
43
-
44
- ## Training and evaluation data
45
-
46
- More information needed
47
-
48
- ## Training procedure
49
-
50
- ### Training hyperparameters
51
-
52
- The following hyperparameters were used during training:
53
- - learning_rate: 5e-05
54
- - train_batch_size: 1
55
- - eval_batch_size: 1
56
- - seed: 42
57
- - distributed_type: multi-GPU
58
- - num_devices: 6
59
- - gradient_accumulation_steps: 8
60
- - total_train_batch_size: 48
61
- - total_eval_batch_size: 6
62
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
- - lr_scheduler_type: linear
64
- - lr_scheduler_warmup_ratio: 0.1
65
- - num_epochs: 3.0
66
-
67
- ### Training results
68
-
69
-
70
-
71
- ### Framework versions
72
-
73
- - Transformers 4.41.0.dev0
74
- - Pytorch 2.2.1+cu118
75
- - Datasets 2.17.1
76
- - Tokenizers 0.19.1