ksirts commited on
Commit
3d3da24
·
1 Parent(s): 438506b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -15
README.md CHANGED
@@ -1,8 +1,6 @@
1
  ---
2
  tags:
3
  - generated_from_trainer
4
- datasets:
5
- - rubric
6
  metrics:
7
  - accuracy
8
  model-index:
@@ -11,10 +9,6 @@ model-index:
11
  - task:
12
  name: Text Classification
13
  type: text-classification
14
- dataset:
15
- name: rubric
16
- type: rubric
17
- args: rubric
18
  metrics:
19
  - name: Accuracy
20
  type: accuracy
@@ -26,8 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
26
 
27
  # EstBERT128_Rubric
28
 
29
- This model is a fine-tuned version of [tartuNLP/EstBERT](https://huggingface.co/tartuNLP/EstBERT) on the rubric categories of the [Estonian Valence dataset](http://peeter.eki.ee:5000/valence/paragraphsquery).
30
- The data was split into train/dev/test parts with 70/10/20 proportions.
31
  It achieves the following results on the test set:
32
  - Loss: 2.0552
33
  - Accuracy: 0.8329
@@ -38,7 +31,16 @@ A single linear layer classifier is fit on top of the last layer [CLS] token rep
38
 
39
  ## Intended uses & limitations
40
 
41
- This model is intended to be used as it is. It can be used to predict nine rubric categories of Estonian texts. The nine rubric labels in the Estonian Valence dataset are:
 
 
 
 
 
 
 
 
 
42
  - ARVAMUS (opinion)
43
  - EESTI (domestic)
44
  - ELU-O (life)
@@ -51,13 +53,9 @@ This model is intended to be used as it is. It can be used to predict nine rubri
51
 
52
  It probably makes sense to treat the two comments categories (KOMM-O-ELU and KOMM-P-EESTI) as a single category.
53
 
54
- We do not guarantee that the model is useful for anything or that the predictions are accurate on new data.
55
-
56
- ## Training and evaluation data
57
-
58
- More information needed
59
-
60
  ## Training procedure
 
 
61
 
62
  ### Training hyperparameters
63
 
@@ -72,6 +70,7 @@ The following hyperparameters were used during training:
72
  - mixed_precision_training: Native AMP
73
 
74
  ### Training results
 
75
 
76
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
77
  |:-------------:|:-----:|:-----:|:---------------:|:--------:|
 
1
  ---
2
  tags:
3
  - generated_from_trainer
 
 
4
  metrics:
5
  - accuracy
6
  model-index:
 
9
  - task:
10
  name: Text Classification
11
  type: text-classification
 
 
 
 
12
  metrics:
13
  - name: Accuracy
14
  type: accuracy
 
20
 
21
  # EstBERT128_Rubric
22
 
23
+ This model is a fine-tuned version of [tartuNLP/EstBERT](https://huggingface.co/tartuNLP/EstBERT).
 
24
  It achieves the following results on the test set:
25
  - Loss: 2.0552
26
  - Accuracy: 0.8329
 
31
 
32
  ## Intended uses & limitations
33
 
34
+ This model is intended to be used as it is. It can be used to predict nine rubric categories of Estonian texts.
35
+
36
+ We do not guarantee that the model is useful for anything or that the predictions are accurate on new data.
37
+
38
+ ## Training and evaluation data
39
+
40
+ The model was trained and evaluated on the rubric categories of the [Estonian Valence dataset](http://peeter.eki.ee:5000/valence/paragraphsquery).
41
+ The data was split into train/dev/test parts with 70/10/20 proportions.
42
+
43
+ The nine rubric labels in the Estonian Valence dataset are:
44
  - ARVAMUS (opinion)
45
  - EESTI (domestic)
46
  - ELU-O (life)
 
53
 
54
  It probably makes sense to treat the two comments categories (KOMM-O-ELU and KOMM-P-EESTI) as a single category.
55
 
 
 
 
 
 
 
56
  ## Training procedure
57
+ The model was trained for maximu 100 epochs using early stopping procedure. After every epoch, the accuracy was calculated on the development set.
58
+ If the development set accuracy did not improve for 20 epochs, the training was stopped.
59
 
60
  ### Training hyperparameters
61
 
 
70
  - mixed_precision_training: Native AMP
71
 
72
  ### Training results
73
+ The final model was taken after 39th epoch.
74
 
75
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
76
  |:-------------:|:-----:|:-----:|:---------------:|:--------:|