Adzka commited on
Commit
54a3482
·
verified ·
1 Parent(s): 1a488a1

Adzka/reward-model-distilbert-indo

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: mit
3
- base_model: w11wo/indonesian-roberta-base-sentiment-classifier
4
  tags:
5
  - generated_from_trainer
6
  metrics:
@@ -15,10 +15,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # test-reward-model
17
 
18
- This model is a fine-tuned version of [w11wo/indonesian-roberta-base-sentiment-classifier](https://huggingface.co/w11wo/indonesian-roberta-base-sentiment-classifier) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.6974
21
- - Accuracy: 0.7368
22
 
23
  ## Model description
24
 
@@ -43,16 +43,20 @@ The following hyperparameters were used during training:
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - num_epochs: 5
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
- | 0.5895 | 1.25 | 50 | 0.6307 | 0.6842 |
53
- | 0.3559 | 2.5 | 100 | 0.6381 | 0.7105 |
54
- | 0.1523 | 3.75 | 150 | 0.6340 | 0.7105 |
55
- | 0.0694 | 5.0 | 200 | 0.6974 | 0.7368 |
 
 
 
 
56
 
57
 
58
  ### Framework versions
 
1
  ---
2
  license: mit
3
+ base_model: cahya/distilbert-base-indonesian
4
  tags:
5
  - generated_from_trainer
6
  metrics:
 
15
 
16
  # test-reward-model
17
 
18
+ This model is a fine-tuned version of [cahya/distilbert-base-indonesian](https://huggingface.co/cahya/distilbert-base-indonesian) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.8697
21
+ - Accuracy: 0.6316
22
 
23
  ## Model description
24
 
 
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
+ - num_epochs: 10
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
+ | 0.6672 | 1.25 | 50 | 0.5611 | 0.6316 |
53
+ | 0.3793 | 2.5 | 100 | 0.6801 | 0.6579 |
54
+ | 0.186 | 3.75 | 150 | 0.9431 | 0.6711 |
55
+ | 0.0949 | 5.0 | 200 | 0.6777 | 0.6579 |
56
+ | 0.0602 | 6.25 | 250 | 0.7489 | 0.6316 |
57
+ | 0.0468 | 7.5 | 300 | 0.7410 | 0.6447 |
58
+ | 0.0228 | 8.75 | 350 | 0.9266 | 0.6316 |
59
+ | 0.019 | 10.0 | 400 | 0.8697 | 0.6316 |
60
 
61
 
62
  ### Framework versions
config.json CHANGED
@@ -1,39 +1,31 @@
1
  {
2
- "_name_or_path": "w11wo/indonesian-roberta-base-sentiment-classifier",
 
3
  "architectures": [
4
- "RobertaForSequenceClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
- "classifier_dropout": null,
9
- "eos_token_id": 2,
10
- "gradient_checkpointing": false,
11
- "hidden_act": "gelu",
12
- "hidden_dropout_prob": 0.1,
13
- "hidden_size": 768,
14
  "id2label": {
15
- "0": "positive",
16
- "1": "neutral",
17
- "2": "negative"
18
  },
19
  "initializer_range": 0.02,
20
- "intermediate_size": 3072,
21
  "label2id": {
22
- "negative": 2,
23
- "neutral": 1,
24
- "positive": 0
25
  },
26
- "layer_norm_eps": 1e-05,
27
- "max_position_embeddings": 514,
28
- "model_type": "roberta",
29
- "num_attention_heads": 12,
30
- "num_hidden_layers": 12,
31
- "pad_token_id": 1,
32
- "position_embedding_type": "absolute",
33
- "problem_type": "single_label_classification",
 
 
34
  "torch_dtype": "float32",
35
  "transformers_version": "4.38.2",
36
- "type_vocab_size": 1,
37
- "use_cache": true,
38
- "vocab_size": 50265
39
  }
 
1
  {
2
+ "_name_or_path": "cahya/distilbert-base-indonesian",
3
+ "activation": "gelu",
4
  "architectures": [
5
+ "DistilBertForSequenceClassification"
6
  ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
 
 
 
 
11
  "id2label": {
12
+ "0": "LABEL_0"
 
 
13
  },
14
  "initializer_range": 0.02,
 
15
  "label2id": {
16
+ "LABEL_0": 0
 
 
17
  },
18
+ "max_position_embeddings": 512,
19
+ "model_type": "distilbert",
20
+ "n_heads": 12,
21
+ "n_layers": 6,
22
+ "output_hidden_states": true,
23
+ "pad_token_id": 0,
24
+ "qa_dropout": 0.1,
25
+ "seq_classif_dropout": 0.2,
26
+ "sinusoidal_pos_embds": true,
27
+ "tie_weights_": true,
28
  "torch_dtype": "float32",
29
  "transformers_version": "4.38.2",
30
+ "vocab_size": 32000
 
 
31
  }
emissions.csv CHANGED
@@ -1,2 +1,3 @@
1
  timestamp,project_name,run_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
2
- 2024-07-02T00:11:50,codecarbon,e6298b50-8854-4fb6-a92d-43999c4601c4,338.201961517334,0.010694947656611987,3.1622961642887557e-05,42.5,108.43,11.946327209472656,0.003992266833119922,0.012046153513913619,0.0011206903215284546,0.017159110668561992,Indonesia,IDN,west java,,,Windows-10-10.0.19045-SP0,3.11.5,2.2.3,8,12th Gen Intel(R) Core(TM) i3-12100F,1,1 x NVIDIA GeForce RTX 3060,107.6071,-6.9217,31.85687255859375,machine,N,1.0
 
 
1
  timestamp,project_name,run_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
2
+ 2024-07-02T00:11:50,codecarbon,e6298b50-8854-4fb6-a92d-43999c4601c4,338.201961517334,0.0106949476566119,3.162296164288756e-05,42.5,108.43,11.946327209472656,0.0039922668331199,0.0120461535139136,0.0011206903215284,0.0171591106685619,Indonesia,IDN,west java,,,Windows-10-10.0.19045-SP0,3.11.5,2.2.3,8,12th Gen Intel(R) Core(TM) i3-12100F,1,1 x NVIDIA GeForce RTX 3060,107.6071,-6.9217,31.85687255859375,machine,N,1.0
3
+ 2024-07-02T00:31:35,codecarbon,d29dc800-1005-4d6f-8e2e-8bab915a98a4,221.70085859298706,0.008340818735775607,3.762195053600684e-05,42.5,166.724,11.946327209472656,0.0026170430142018536,0.010030346762774322,0.0007347263725223773,0.013382116149498554,Indonesia,IDN,west java,,,Windows-10-10.0.19045-SP0,3.11.5,2.2.3,8,12th Gen Intel(R) Core(TM) i3-12100F,1,1 x NVIDIA GeForce RTX 3060,107.6071,-6.9217,31.85687255859375,machine,N,1.0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8aa724f42799c8235ac65411e2cd383b225fa2c7d71dd13060f3766ce2b58bf
3
- size 498615900
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52c8b4a93c799a67bf0b574cd480842bd90cc6ab22bc095b5dc00ec43a5dec68
3
+ size 272369900
runs/Jul02_00-27-47_DESKTOP-HH0RPGN/events.out.tfevents.1719854873.DESKTOP-HH0RPGN.6200.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bc5c631cb7aeaedf20e7f6f491decb2eec11447ad514329c043b7c5a7613158
3
+ size 9196
special_tokens_map.json CHANGED
@@ -1,51 +1,7 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "cls_token": {
10
- "content": "<s>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "eos_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "mask_token": {
24
- "content": "<mask>",
25
- "lstrip": true,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
- "pad_token": {
31
- "content": "<pad>",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false
36
- },
37
- "sep_token": {
38
- "content": "</s>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false
43
- },
44
- "unk_token": {
45
- "content": "<unk>",
46
- "lstrip": false,
47
- "normalized": false,
48
- "rstrip": false,
49
- "single_word": false
50
- }
51
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,57 +1,55 @@
1
  {
2
- "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
5
- "content": "<s>",
6
  "lstrip": false,
7
  "normalized": false,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
- "1": {
13
- "content": "<pad>",
14
  "lstrip": false,
15
  "normalized": false,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
19
  },
20
- "2": {
21
- "content": "</s>",
22
  "lstrip": false,
23
  "normalized": false,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
  },
28
- "3": {
29
- "content": "<unk>",
30
  "lstrip": false,
31
  "normalized": false,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
35
  },
36
- "4": {
37
- "content": "<mask>",
38
- "lstrip": true,
39
  "normalized": false,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
43
  }
44
  },
45
- "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
- "cls_token": "<s>",
48
- "eos_token": "</s>",
49
- "errors": "replace",
50
- "mask_token": "<mask>",
51
- "model_max_length": 1000000000000000019884624838656,
52
- "pad_token": "<pad>",
53
- "sep_token": "</s>",
54
- "tokenizer_class": "RobertaTokenizer",
55
- "trim_offsets": true,
56
- "unk_token": "<unk>"
57
  }
 
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
+ "content": "[PAD]",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
+ "100": {
12
+ "content": "[UNK]",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
+ "101": {
20
+ "content": "[CLS]",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "102": {
28
+ "content": "[SEP]",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
42
  }
43
  },
 
44
  "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "BertTokenizer",
54
+ "unk_token": "[UNK]"
55
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0a0b921d70ad183817a2957b4977890a97177522e7c26660841711f33040833
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9de0ed44dc3865bb9a68a129a145f8782eecf346d387339b335acb99144c27ea
3
  size 4920
vocab.txt ADDED
The diff for this file is too large to render. See raw diff