Model Card for MESSItom/BERT-review-sentiment-analysis

This model is fine-tuned from BERT to perform sentiment analysis on a custom dataset containing student reviews about campus events or amenities. The objective is to classify the sentiments (positive, negative, neutral) while maintaining high performance metrics like accuracy.

Model Details

Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Developed by: Messy Tom Binoy
  • Funded by: No funding, self-funded
  • Shared by: Messy Tom Binoy
  • Model type: BERT
  • Language(s) (NLP): English
  • License: MIT
  • Finetuned from model: google-bert/bert-base-uncased

Model Sources

Uses

Direct Use

The model can be used directly for sentiment classification of student reviews about campus events or amenities.

Downstream Use

The model can be fine-tuned further for other sentiment analysis tasks or integrated into larger applications for sentiment classification.

Out-of-Scope Use

The model is not suitable for tasks outside sentiment analysis, such as language translation or text generation.

Bias, Risks, and Limitations

The model may inherit biases from the pre-trained BERT model and the custom dataset used for fine-tuning. It may not perform well on reviews that are significantly different from the training data.

Recommendations

Users should be aware of the potential biases and limitations of the model. It is recommended to evaluate the model on a diverse set of reviews to understand its performance and limitations.

How to Get Started with the Model

Use the code below to get started with the model:

from transformers import AutoModelForSequenceClassification, AutoTokenizer

model_id = "MESSItom/BERT-review-sentiment-analysis"

model = AutoModelForSequenceClassification.from_pretrained(model_id)
tokenizer = AutoTokenizer.from_pretrained(model_id)

def predict_sentiment(text):
    inputs = tokenizer(text, return_tensors='pt', padding=True, truncation=True, max_length=512)
    with torch.no_grad():
        outputs = model(**inputs)
    logits = outputs.logits
    predicted_class = torch.argmax(logits, dim=-1).item()
    class_names = ['positive', 'neutral', 'negative']
    sentiment = class_names[predicted_class]
    return sentiment
Downloads last month
15
Safetensors
Model size
109M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for MESSItom/BERT-review-sentiment-analysis

Finetuned
(2309)
this model