Update README.md
Browse files
README.md
CHANGED
@@ -15,17 +15,18 @@ metrics:
|
|
15 |
- f1
|
16 |
---
|
17 |
|
18 |
-
#
|
|
|
19 |
|
20 |
## Model Description
|
21 |
|
22 |
-
`distilbert-base-uncased-emotion` is a specialized model finetuned on a combination of unify-emotion-datasets
|
23 |
|
24 |
Model parameters were as follows: sequence length of 64, learning rate of 2e-5, batch size of 128, trained for 8 epochs. For steps on how to use the model for inference, please refer to the accompanying Inference.ipynb notebook.
|
25 |
|
26 |
## Training Data
|
27 |
|
28 |
-
The training data was obtained from the Unify Emotion Datasets available at https://github.com/sarnthil/unify-emotion-datasets.
|
29 |
|
30 |
## Evaluation Metrics
|
31 |
|
@@ -37,11 +38,12 @@ The model was evaluated using the following metrics:
|
|
37 |
|
38 |
## Research
|
39 |
|
40 |
-
The underlying research for emotion extraction from social media can be found in the paper
|
41 |
-
The paper is available at https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3975884.
|
42 |
|
43 |
### Research using EmTract
|
|
|
44 |
[Social Media Emotions and IPO Returns](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4384573)
|
|
|
45 |
[Investor Emotions and Earnings Announcements](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3626025])
|
46 |
|
47 |
## License
|
|
|
15 |
- f1
|
16 |
---
|
17 |
|
18 |
+
# EmTract
|
19 |
+
## DistilBERT-Base-Uncased-Emotion
|
20 |
|
21 |
## Model Description
|
22 |
|
23 |
+
`emtract-distilbert-base-uncased-emotion` is a specialized model finetuned on a combination of [unify-emotion-datasets](https://github.com/sarnthil/unify-emotion-datasets), containing around 250K texts labeled across seven emotion categories: neutral, happy, sad, anger, disgust, surprise, and fear. This model was later adapted to a smaller set of 10K hand-tagged messages from StockTwits. The model is designed to excel at emotion detection in financial social media content such as that found on StockTwits.
|
24 |
|
25 |
Model parameters were as follows: sequence length of 64, learning rate of 2e-5, batch size of 128, trained for 8 epochs. For steps on how to use the model for inference, please refer to the accompanying Inference.ipynb notebook.
|
26 |
|
27 |
## Training Data
|
28 |
|
29 |
+
The training data was obtained from the Unify Emotion Datasets available at [here](https://github.com/sarnthil/unify-emotion-datasets).
|
30 |
|
31 |
## Evaluation Metrics
|
32 |
|
|
|
38 |
|
39 |
## Research
|
40 |
|
41 |
+
The underlying research for emotion extraction from social media can be found in the paper [EmTract: Extracting Emotions from Social Media](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3975884).
|
|
|
42 |
|
43 |
### Research using EmTract
|
44 |
+
|
45 |
[Social Media Emotions and IPO Returns](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4384573)
|
46 |
+
|
47 |
[Investor Emotions and Earnings Announcements](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3626025])
|
48 |
|
49 |
## License
|