djsull/binary_every_exp
Browse files- README.md +13 -13
- config.json +5 -5
- model.safetensors +2 -2
- runs/Aug23_01-45-11_129-146-79-76/events.out.tfevents.1724377512.129-146-79-76.1742119.0 +3 -0
- runs/Aug23_01-45-51_129-146-79-76/events.out.tfevents.1724377552.129-146-79-76.1742507.0 +3 -0
- runs/Aug23_01-45-51_129-146-79-76/events.out.tfevents.1724377808.129-146-79-76.1742507.1 +3 -0
- training_args.bin +1 -1
README.md
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
---
|
2 |
-
base_model: klue/roberta-
|
3 |
tags:
|
4 |
- generated_from_trainer
|
5 |
metrics:
|
@@ -17,9 +17,9 @@ should probably proofread and complete it, then remove this comment. -->
|
|
17 |
|
18 |
# binary_every_exp
|
19 |
|
20 |
-
This model is a fine-tuned version of [klue/roberta-
|
21 |
It achieves the following results on the evaluation set:
|
22 |
-
- Loss: 0.
|
23 |
- Precision: 1.0
|
24 |
- Recall: 1.0
|
25 |
- F1: 1.0
|
@@ -43,8 +43,8 @@ More information needed
|
|
43 |
|
44 |
The following hyperparameters were used during training:
|
45 |
- learning_rate: 5e-05
|
46 |
-
- train_batch_size:
|
47 |
-
- eval_batch_size:
|
48 |
- seed: 42
|
49 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
50 |
- lr_scheduler_type: linear
|
@@ -54,14 +54,14 @@ The following hyperparameters were used during training:
|
|
54 |
|
55 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
56 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
57 |
-
| No log | 1.0 |
|
58 |
-
| No log | 2.0 |
|
59 |
-
| No log | 3.0 |
|
60 |
-
| No log | 4.0 |
|
61 |
-
| No log | 5.0 |
|
62 |
-
| No log | 6.0 |
|
63 |
-
| No log | 7.0 |
|
64 |
-
| No log | 8.0 |
|
65 |
|
66 |
|
67 |
### Framework versions
|
|
|
1 |
---
|
2 |
+
base_model: klue/roberta-large
|
3 |
tags:
|
4 |
- generated_from_trainer
|
5 |
metrics:
|
|
|
17 |
|
18 |
# binary_every_exp
|
19 |
|
20 |
+
This model is a fine-tuned version of [klue/roberta-large](https://huggingface.co/klue/roberta-large) on an unknown dataset.
|
21 |
It achieves the following results on the evaluation set:
|
22 |
+
- Loss: 0.0005
|
23 |
- Precision: 1.0
|
24 |
- Recall: 1.0
|
25 |
- F1: 1.0
|
|
|
43 |
|
44 |
The following hyperparameters were used during training:
|
45 |
- learning_rate: 5e-05
|
46 |
+
- train_batch_size: 32
|
47 |
+
- eval_batch_size: 16
|
48 |
- seed: 42
|
49 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
50 |
- lr_scheduler_type: linear
|
|
|
54 |
|
55 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
56 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
57 |
+
| No log | 1.0 | 19 | 0.0020 | 1.0 | 1.0 | 1.0 | 1.0 |
|
58 |
+
| No log | 2.0 | 38 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
|
59 |
+
| No log | 3.0 | 57 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
|
60 |
+
| No log | 4.0 | 76 | 0.0076 | 1.0 | 1.0 | 1.0 | 1.0 |
|
61 |
+
| No log | 5.0 | 95 | 0.1695 | 0.9375 | 1.0 | 0.9677 | 0.98 |
|
62 |
+
| No log | 6.0 | 114 | 0.1624 | 0.9375 | 1.0 | 0.9677 | 0.98 |
|
63 |
+
| No log | 7.0 | 133 | 0.0406 | 0.9375 | 1.0 | 0.9677 | 0.98 |
|
64 |
+
| No log | 8.0 | 152 | 0.0033 | 1.0 | 1.0 | 1.0 | 1.0 |
|
65 |
|
66 |
|
67 |
### Framework versions
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "klue/roberta-
|
3 |
"architectures": [
|
4 |
"RobertaForSequenceClassification"
|
5 |
],
|
@@ -10,13 +10,13 @@
|
|
10 |
"gradient_checkpointing": false,
|
11 |
"hidden_act": "gelu",
|
12 |
"hidden_dropout_prob": 0.1,
|
13 |
-
"hidden_size":
|
14 |
"id2label": {
|
15 |
"0": "None",
|
16 |
"1": "Profile"
|
17 |
},
|
18 |
"initializer_range": 0.02,
|
19 |
-
"intermediate_size":
|
20 |
"label2id": {
|
21 |
"None": 0,
|
22 |
"Profile": 1
|
@@ -24,8 +24,8 @@
|
|
24 |
"layer_norm_eps": 1e-05,
|
25 |
"max_position_embeddings": 514,
|
26 |
"model_type": "roberta",
|
27 |
-
"num_attention_heads":
|
28 |
-
"num_hidden_layers":
|
29 |
"pad_token_id": 1,
|
30 |
"position_embedding_type": "absolute",
|
31 |
"problem_type": "single_label_classification",
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "klue/roberta-large",
|
3 |
"architectures": [
|
4 |
"RobertaForSequenceClassification"
|
5 |
],
|
|
|
10 |
"gradient_checkpointing": false,
|
11 |
"hidden_act": "gelu",
|
12 |
"hidden_dropout_prob": 0.1,
|
13 |
+
"hidden_size": 1024,
|
14 |
"id2label": {
|
15 |
"0": "None",
|
16 |
"1": "Profile"
|
17 |
},
|
18 |
"initializer_range": 0.02,
|
19 |
+
"intermediate_size": 4096,
|
20 |
"label2id": {
|
21 |
"None": 0,
|
22 |
"Profile": 1
|
|
|
24 |
"layer_norm_eps": 1e-05,
|
25 |
"max_position_embeddings": 514,
|
26 |
"model_type": "roberta",
|
27 |
+
"num_attention_heads": 16,
|
28 |
+
"num_hidden_layers": 24,
|
29 |
"pad_token_id": 1,
|
30 |
"position_embedding_type": "absolute",
|
31 |
"problem_type": "single_label_classification",
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b906e47e7791b77e7bcc430e70e3b0b0fe652f94b5ea08956795f5954476fa2e
|
3 |
+
size 1346681920
|
runs/Aug23_01-45-11_129-146-79-76/events.out.tfevents.1724377512.129-146-79-76.1742119.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e38d672e410bf79798f63d1aed43f2948a1403857e2a7c563512f6590834eacd
|
3 |
+
size 88
|
runs/Aug23_01-45-51_129-146-79-76/events.out.tfevents.1724377552.129-146-79-76.1742507.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59383615a9f345743e4b0d5dfc319b5ef2ebe41e2f8b6a86c39f3548e711f63d
|
3 |
+
size 9186
|
runs/Aug23_01-45-51_129-146-79-76/events.out.tfevents.1724377808.129-146-79-76.1742507.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:82dcd93975d5ae515ba72d69f9563a7b2ae840940156338ea85316a42eba40e1
|
3 |
+
size 560
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5176
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fb02f8f3178acd5e1c1a4e90647dcd6d134d003b253393f7fc6ffbed06c99f0c
|
3 |
size 5176
|