Model Card for IndoBERT Sentiment Analysis
This model is fine-tuned from indobenchmark/indobert-base-p1
for binary sentiment classification (Positive/Negative) on Indonesian text.
Model Details
Model Description
- Developed by: agufsamudra
- Model type: Text Classification
- Language(s): Indonesian (id)
- License: Apache-2.0
- Fine-tuned from model: indobenchmark/indobert-base-p1
Model Sources
Uses
Direct Use
This model is intended for binary sentiment classification tasks in Indonesian language texts. It predicts whether a given text expresses positive or negative sentiment.
Out-of-Scope Use
The model is not designed to classify neutral sentiments or handle languages other than Indonesian.
Bias, Risks, and Limitations
- Bias: The model's performance is reliant on the quality and diversity of the training data. Biases in the dataset may influence predictions.
- Limitations: The model is limited to binary sentiment analysis and may not perform well on ambiguous or mixed-sentiment texts.
Recommendations
Users should validate predictions on a case-by-case basis for high-stakes applications.
How to Get Started with the Model
from transformers import BertTokenizer, BertForSequenceClassification
# Load the tokenizer and model
tokenizer = BertTokenizer.from_pretrained("agufsamudra/indo-sentiment-analysis")
model = BertForSequenceClassification.from_pretrained("agufsamudra/indo-sentiment-analysis")
# Example usage
text = "Saya sangat puas dengan pelayanan ini!"
inputs = tokenizer(text, return_tensors="pt", padding="max_length", truncation=True, max_length=128)
outputs = model(**inputs)
logits = outputs.logits
prediction = logits.argmax(-1).item()
label = "Positive" if prediction == 1 else "Negative"
print(f"Sentiment: {label}")
Training Details
Training Data
The model was trained on a dataset of Indonesian text reviews from Play Store applications. The dataset was labeled for binary sentiment analysis (Positive and Negative). It contains an equal distribution of positive and negative examples to ensure balanced learning.
Training Procedure
Training Hyperparameters
- Optimizer: AdamW
- Learning Rate: 3e-6
- Epochs: 3
- Max Sequence Length: 128 tokens
Evaluation
Testing Data
The model was evaluated on a separate test dataset of 20,000 samples (10,000 Positive, 10,000 Negative).
Metrics
The model's performance was evaluated using standard metrics, including accuracy, precision, recall, and F1-score.
Results
Metric | Training Set | Testing Set |
---|---|---|
Accuracy | 95.28% | 95.56% |
Precision | 96% | 96% |
Recall | 96% | 96% |
F1-Score | 96% | 96% |
Technical Specifications
Model Architecture and Objective
The model is based on IndoBERT, a pre-trained transformer model tailored for Indonesian text. It was fine-tuned for binary classification tasks.
Compute Infrastructure
- Hardware: Google Collab GPU
- Software: Python, PyTorch, Transformers library
- Downloads last month
- 27
Model tree for agufsamudra/indo-sentiment-analysis
Base model
indobenchmark/indobert-base-p1