Model Card for MESSItom/BERT-review-sentiment-analysis
This model is fine-tuned from BERT to perform sentiment analysis on a custom dataset containing student reviews about campus events or amenities. The objective is to classify the sentiments (positive, negative, neutral) while maintaining high performance metrics like accuracy.
Model Details
Model Description
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by: Messy Tom Binoy
- Funded by: No funding, self-funded
- Shared by: Messy Tom Binoy
- Model type: BERT
- Language(s) (NLP): English
- License: MIT
- Finetuned from model: google-bert/bert-base-uncased
Model Sources
- Repository: GitHub Repository
- Demo: GitHub Demo
Uses
Direct Use
The model can be used directly for sentiment classification of student reviews about campus events or amenities.
Downstream Use
The model can be fine-tuned further for other sentiment analysis tasks or integrated into larger applications for sentiment classification.
Out-of-Scope Use
The model is not suitable for tasks outside sentiment analysis, such as language translation or text generation.
Bias, Risks, and Limitations
The model may inherit biases from the pre-trained BERT model and the custom dataset used for fine-tuning. It may not perform well on reviews that are significantly different from the training data.
Recommendations
Users should be aware of the potential biases and limitations of the model. It is recommended to evaluate the model on a diverse set of reviews to understand its performance and limitations.
How to Get Started with the Model
Use the code below to get started with the model:
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model_id = "MESSItom/BERT-review-sentiment-analysis"
model = AutoModelForSequenceClassification.from_pretrained(model_id)
tokenizer = AutoTokenizer.from_pretrained(model_id)
def predict_sentiment(text):
inputs = tokenizer(text, return_tensors='pt', padding=True, truncation=True, max_length=512)
with torch.no_grad():
outputs = model(**inputs)
logits = outputs.logits
predicted_class = torch.argmax(logits, dim=-1).item()
class_names = ['positive', 'neutral', 'negative']
sentiment = class_names[predicted_class]
return sentiment
- Downloads last month
- 15
Model tree for MESSItom/BERT-review-sentiment-analysis
Base model
google-bert/bert-base-uncased