vamossyd's picture
Update README.md
47f6e0e
|
raw
history blame
1.77 kB
metadata
language:
  - en
tags:
  - text-classification
  - emotion
  - pytorch
license: mit
datasets:
  - emotion
metrics:
  - accuracy
  - precision
  - recall
  - f1

EmTract

DistilBERT-Base-Uncased-Emotion

Model Description

emtract-distilbert-base-uncased-emotion is a specialized model finetuned on a combination of unify-emotion-datasets, containing around 250K texts labeled across seven emotion categories: neutral, happy, sad, anger, disgust, surprise, and fear. This model was later adapted to a smaller set of 10K hand-tagged messages from StockTwits. The model is designed to excel at emotion detection in financial social media content such as that found on StockTwits.

Model parameters were as follows: sequence length of 64, learning rate of 2e-5, batch size of 128, trained for 8 epochs. For steps on how to use the model for inference, please refer to the accompanying Inference.ipynb notebook.

Training Data

The training data was obtained from the Unify Emotion Datasets available at here.

Evaluation Metrics

The model was evaluated using the following metrics:

  • Accuracy
  • Precision
  • Recall
  • F1-score

Research

The underlying research for emotion extraction from social media can be found in the paper EmTract: Extracting Emotions from Social Media.

Research using EmTract

Social Media Emotions and IPO Returns

Investor Emotions and Earnings Announcements

License

This project is licensed under the terms of the MIT license.