vamossyd commited on
Commit
a92d680
β€’
1 Parent(s): cb686e2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -8
README.md CHANGED
@@ -15,17 +15,37 @@ metrics:
15
  - f1
16
  ---
17
 
18
- # bert-base-uncased-emotion
19
 
20
- ## Model description
21
 
22
- `bert-base-uncased` finetuned on the unify-emotion-datasets (https://github.com/sarnthil/unify-emotion-datasets) [~250K texts with 7 labels -- neutral, happy, sad, anger, disgust, surprise, fear], then transferred to
23
- a small sample of 10K hand-tagged StockTwits messages. Optimized for extracting emotions from financial social media, such as StockTwits.
24
 
25
- Sequence length 64, learning rate 2e-5, batch size 128, 8 epochs.
26
 
27
- For inference, follow the Inference.ipynb notebook.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
- ## Training data
30
 
31
- Data came from https://github.com/sarnthil/unify-emotion-datasets.
 
15
  - f1
16
  ---
17
 
18
+ # DistilBERT-Base-Uncased-Emotion
19
 
20
+ ## Model Description
21
 
22
+ `distilbert-base-uncased-emotion` is a specialized model finetuned on a combination of unify-emotion-datasets (https://github.com/sarnthil/unify-emotion-datasets), containing around 250K texts labeled across seven emotion categories: neutral, happy, sad, anger, disgust, surprise, and fear. This model was later adapted to a smaller set of 10K hand-tagged messages from StockTwits. The model is designed to excel at emotion detection in financial social media content such as that found on StockTwits.
 
23
 
24
+ Model parameters were as follows: sequence length of 64, learning rate of 2e-5, batch size of 128, trained for 8 epochs. For steps on how to use the model for inference, please refer to the accompanying Inference.ipynb notebook.
25
 
26
+ ## Training Data
27
+
28
+ The training data was obtained from the Unify Emotion Datasets available at https://github.com/sarnthil/unify-emotion-datasets.
29
+
30
+ ## Evaluation Metrics
31
+
32
+ The model was evaluated using the following metrics:
33
+ - Accuracy
34
+ - Precision
35
+ - Recall
36
+ - F1-score
37
+
38
+ ## Research
39
+
40
+ The underlying research for emotion extraction from social media can be found in the paper "EmTract: Extracting Emotions from Social Media".
41
+ The paper is available at https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3975884.
42
+
43
+ ### Research using EmTract
44
+ [Social Media Emotions and IPO Returns](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4384573)
45
+ [Investor Emotions and Earnings Announcements](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3626025])
46
+
47
+ ## License
48
+
49
+ This project is licensed under the terms of the MIT license.
50
 
 
51