srimeenakshiks
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -18,6 +18,7 @@ base_model:
|
|
18 |
# Model Card for Model ID
|
19 |
|
20 |
<!-- Provide a quick summary of what the model is/does. -->
|
|
|
21 |
|
22 |
|
23 |
|
@@ -29,21 +30,21 @@ base_model:
|
|
29 |
|
30 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
31 |
|
32 |
-
- **Developed by:**
|
33 |
-
- **Funded by [optional]:** [
|
34 |
-
- **Shared by [optional]:** [
|
35 |
-
- **Model type:**
|
36 |
-
- **Language(s) (NLP):**
|
37 |
-
- **License:**
|
38 |
-
- **Finetuned from model [optional]:**
|
39 |
|
40 |
### Model Sources [optional]
|
41 |
|
42 |
<!-- Provide the basic links for the model. -->
|
43 |
|
44 |
-
- **Repository:** [
|
45 |
-
- **Paper [optional]:** [
|
46 |
-
- **Demo [optional]:** [
|
47 |
|
48 |
## Uses
|
49 |
|
@@ -52,38 +53,44 @@ This is the model card of a 🤗 transformers model that has been pushed on the
|
|
52 |
### Direct Use
|
53 |
|
54 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
|
|
55 |
|
56 |
-
[More Information Needed]
|
57 |
|
58 |
### Downstream Use [optional]
|
59 |
|
60 |
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
|
|
61 |
|
62 |
-
[More Information Needed]
|
63 |
|
64 |
### Out-of-Scope Use
|
65 |
|
66 |
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
|
|
67 |
|
68 |
-
[More Information Needed]
|
69 |
|
70 |
## Bias, Risks, and Limitations
|
71 |
|
72 |
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
|
|
73 |
|
74 |
-
[More Information Needed]
|
75 |
|
76 |
### Recommendations
|
77 |
|
78 |
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
79 |
|
80 |
-
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.
|
81 |
|
82 |
## How to Get Started with the Model
|
83 |
|
84 |
Use the code below to get started with the model.
|
85 |
|
86 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
87 |
|
88 |
## Training Details
|
89 |
|
@@ -91,26 +98,23 @@ Use the code below to get started with the model.
|
|
91 |
|
92 |
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
93 |
|
94 |
-
[
|
|
|
95 |
|
96 |
### Training Procedure
|
97 |
|
98 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
99 |
|
100 |
-
#### Preprocessing [optional]
|
101 |
|
102 |
-
|
103 |
|
|
|
104 |
|
105 |
-
#### Training Hyperparameters
|
106 |
|
107 |
-
|
108 |
|
109 |
-
|
110 |
|
111 |
-
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
112 |
-
|
113 |
-
[More Information Needed]
|
114 |
|
115 |
## Evaluation
|
116 |
|
@@ -122,33 +126,34 @@ Use the code below to get started with the model.
|
|
122 |
|
123 |
<!-- This should link to a Dataset Card if possible. -->
|
124 |
|
125 |
-
|
126 |
|
127 |
#### Factors
|
128 |
|
129 |
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
130 |
|
131 |
-
|
132 |
|
133 |
#### Metrics
|
134 |
|
135 |
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
136 |
|
137 |
-
|
138 |
|
139 |
### Results
|
140 |
|
141 |
-
|
142 |
|
143 |
#### Summary
|
144 |
|
|
|
145 |
|
146 |
|
147 |
-
## Model Examination
|
148 |
|
149 |
<!-- Relevant interpretability work for the model goes here -->
|
150 |
|
151 |
-
|
152 |
|
153 |
## Environmental Impact
|
154 |
|
@@ -156,56 +161,61 @@ Use the code below to get started with the model.
|
|
156 |
|
157 |
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
158 |
|
159 |
-
- **Hardware Type:**
|
160 |
-
- **Hours used:**
|
161 |
-
- **Cloud Provider:**
|
162 |
-
- **Compute Region:**
|
163 |
-
- **Carbon Emitted:**
|
164 |
|
165 |
-
## Technical Specifications
|
166 |
|
167 |
### Model Architecture and Objective
|
168 |
|
169 |
-
|
170 |
|
171 |
### Compute Infrastructure
|
172 |
|
173 |
-
[More Information Needed]
|
174 |
-
|
175 |
#### Hardware
|
176 |
|
177 |
-
|
|
|
178 |
|
179 |
#### Software
|
180 |
|
|
|
|
|
|
|
181 |
[More Information Needed]
|
182 |
|
183 |
-
## Citation
|
184 |
|
185 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
186 |
|
187 |
**BibTeX:**
|
188 |
|
189 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
190 |
|
191 |
**APA:**
|
192 |
|
193 |
-
|
194 |
|
195 |
## Glossary [optional]
|
196 |
|
197 |
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
198 |
|
199 |
-
|
200 |
|
201 |
-
## More Information [optional]
|
202 |
|
203 |
-
|
204 |
-
|
205 |
-
## Model Card Authors [optional]
|
206 |
|
207 |
-
|
208 |
|
209 |
## Model Card Contact
|
210 |
|
211 |
-
|
|
|
18 |
# Model Card for Model ID
|
19 |
|
20 |
<!-- Provide a quick summary of what the model is/does. -->
|
21 |
+
This model is a fine-tuned BERT model designed for aspect-based sentiment analysis, enabling the classification of sentiments associated with specific aspects in text. It provides valuable insights into customer opinions and sentiments regarding different features in user-generated content.
|
22 |
|
23 |
|
24 |
|
|
|
30 |
|
31 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
32 |
|
33 |
+
- **Developed by:** Srimeenakshi K S
|
34 |
+
- **Funded by [optional]:** [Not Applicable]
|
35 |
+
- **Shared by [optional]:** [Not Applicable]
|
36 |
+
- **Model type:** Aspect-Based Sentiment Analysis
|
37 |
+
- **Language(s) (NLP):** English
|
38 |
+
- **License:** MIT License
|
39 |
+
- **Finetuned from model [optional]:** BERT-base-uncased
|
40 |
|
41 |
### Model Sources [optional]
|
42 |
|
43 |
<!-- Provide the basic links for the model. -->
|
44 |
|
45 |
+
- **Repository:** [Hugging Face](https://huggingface.co/srimeenakshiks/aspect-based-sentiment-analyzer-using-bert)
|
46 |
+
- **Paper [optional]:** [Not Applicable]
|
47 |
+
- **Demo [optional]:** [Not Applicable]
|
48 |
|
49 |
## Uses
|
50 |
|
|
|
53 |
### Direct Use
|
54 |
|
55 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
56 |
+
The model can be used directly to classify sentiments in user-generated text based on specified aspects without the need for additional fine-tuning. It is suitable for analyzing reviews, social media posts, and other forms of textual feedback.
|
57 |
|
|
|
58 |
|
59 |
### Downstream Use [optional]
|
60 |
|
61 |
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
62 |
+
This model can be integrated into applications for customer feedback analysis, chatbots for customer service, or sentiment analysis tools for businesses looking to improve their products and services based on customer input.
|
63 |
|
|
|
64 |
|
65 |
### Out-of-Scope Use
|
66 |
|
67 |
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
68 |
+
The model may not perform well with text that contains heavy sarcasm or nuanced expressions. It should not be used for critical decision-making processes without human oversight.
|
69 |
|
|
|
70 |
|
71 |
## Bias, Risks, and Limitations
|
72 |
|
73 |
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
74 |
+
The model may reflect biases present in the training data, leading to potential misclassification of sentiments. Users should be cautious in interpreting results, particularly in sensitive applications where sentiment analysis can impact customer relationships.
|
75 |
|
|
|
76 |
|
77 |
### Recommendations
|
78 |
|
79 |
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
80 |
|
81 |
+
Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model. It is recommended to validate results with a diverse set of data and consider human judgment in ambiguous cases.
|
82 |
|
83 |
## How to Get Started with the Model
|
84 |
|
85 |
Use the code below to get started with the model.
|
86 |
|
87 |
+
```
|
88 |
+
from transformers import pipeline
|
89 |
+
|
90 |
+
sentiment_analyzer = pipeline("text-classification", model="srimeenakshiks/aspect-based-sentiment-analyzer-using-bert")
|
91 |
+
result = sentiment_analyzer("The food was amazing, but the service was slow.", aspect="service")
|
92 |
+
print(result)
|
93 |
+
```
|
94 |
|
95 |
## Training Details
|
96 |
|
|
|
98 |
|
99 |
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
100 |
|
101 |
+
The model was trained on the [IMDB dataset](https://huggingface.co/datasets/stanfordnlp/imdb), which contains movie reviews labeled with sentiment (positive and negative). This dataset is commonly used for sentiment analysis tasks and includes a diverse range of reviews, allowing the model to learn various expressions of sentiment effectively.
|
102 |
+
|
103 |
|
104 |
### Training Procedure
|
105 |
|
106 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
107 |
|
|
|
108 |
|
109 |
+
#### Preprocessing
|
110 |
|
111 |
+
Data preprocessing involved tokenization, padding, and normalization of text inputs to fit the BERT model requirements.
|
112 |
|
|
|
113 |
|
114 |
+
#### Training Hyperparameters
|
115 |
|
116 |
+
- **Training regime:** fp16 mixed precision <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
117 |
|
|
|
|
|
|
|
118 |
|
119 |
## Evaluation
|
120 |
|
|
|
126 |
|
127 |
<!-- This should link to a Dataset Card if possible. -->
|
128 |
|
129 |
+
The model was evaluated using the same dataset on which it was trained, ensuring consistency in performance metrics and providing a reliable assessment of its capabilities in aspect-based sentiment analysis.
|
130 |
|
131 |
#### Factors
|
132 |
|
133 |
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
134 |
|
135 |
+
The evaluation included various aspects such as product features, service quality, and user experience.
|
136 |
|
137 |
#### Metrics
|
138 |
|
139 |
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
140 |
|
141 |
+
Evaluation metrics included accuracy, precision, recall, and F1-score, providing a comprehensive assessment of model performance.
|
142 |
|
143 |
### Results
|
144 |
|
145 |
+
The model achieved an accuracy of 95% on the test dataset, demonstrating effectiveness in aspect-based sentiment classification.
|
146 |
|
147 |
#### Summary
|
148 |
|
149 |
+
The results indicate that the model performs well across a range of aspects but may struggle with nuanced sentiment expressions.
|
150 |
|
151 |
|
152 |
+
## Model Examination
|
153 |
|
154 |
<!-- Relevant interpretability work for the model goes here -->
|
155 |
|
156 |
+
Further interpretability work can be conducted to understand how the model makes its predictions, particularly focusing on attention mechanisms within BERT.
|
157 |
|
158 |
## Environmental Impact
|
159 |
|
|
|
161 |
|
162 |
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
163 |
|
164 |
+
- **Hardware Type:** NVIDIA GeForce RTX 4050
|
165 |
+
- **Hours used:** 20 hours
|
166 |
+
- **Cloud Provider:** AWS
|
167 |
+
- **Compute Region:** US-East
|
168 |
+
- **Carbon Emitted:** 3.5
|
169 |
|
170 |
+
## Technical Specifications
|
171 |
|
172 |
### Model Architecture and Objective
|
173 |
|
174 |
+
The model is based on the BERT architecture, specifically designed to understand the context of words in a sentence, enabling it to classify sentiments associated with different aspects effectively.
|
175 |
|
176 |
### Compute Infrastructure
|
177 |
|
|
|
|
|
178 |
#### Hardware
|
179 |
|
180 |
+
- **GPU:** NVIDIA GeForce RTX 2080
|
181 |
+
- **RAM:** 16GB
|
182 |
|
183 |
#### Software
|
184 |
|
185 |
+
- **Framework:** PyTorch
|
186 |
+
- **Library Version**: Hugging Face Transformers version 4.44.2
|
187 |
+
|
188 |
[More Information Needed]
|
189 |
|
190 |
+
## Citation
|
191 |
|
192 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
193 |
|
194 |
**BibTeX:**
|
195 |
|
196 |
+
@model{srimeenakshiks2024aspect,
|
197 |
+
title={Aspect-Based Sentiment Analyzer using BERT},
|
198 |
+
author={Srimeenakshi K S},
|
199 |
+
year={2024},
|
200 |
+
publisher={Hugging Face}
|
201 |
+
}
|
202 |
+
|
203 |
|
204 |
**APA:**
|
205 |
|
206 |
+
Srimeenakshi K S. (2024). _Aspect-Based Sentiment Analyzer using BERT_. Hugging Face.
|
207 |
|
208 |
## Glossary [optional]
|
209 |
|
210 |
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
211 |
|
212 |
+
- **Aspect-Based Sentiment Analysis (ABSA):** A subfield of sentiment analysis that focuses on identifying sentiments related to specific features or aspects of a product or service.
|
213 |
|
|
|
214 |
|
215 |
+
## Model Card Authors
|
|
|
|
|
216 |
|
217 |
+
- **Author:** Srimeenakshi]
|
218 |
|
219 |
## Model Card Contact
|
220 |
|
221 |
+
For inquiries or feedback, please reach out to [[email protected]].
|