Update README.md
Browse files
README.md
CHANGED
@@ -1,11 +1,22 @@
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
---
|
5 |
|
6 |
# Model Card for Model ID
|
7 |
|
8 |
-
<!--
|
9 |
|
10 |
|
11 |
|
@@ -13,160 +24,141 @@ tags: []
|
|
13 |
|
14 |
### Model Description
|
15 |
|
16 |
-
<!--
|
17 |
|
18 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
19 |
|
20 |
-
- **Developed by:**
|
21 |
-
- **
|
22 |
-
- **
|
23 |
-
- **
|
24 |
-
- **
|
25 |
-
- **License:** [More Information Needed]
|
26 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
27 |
|
28 |
### Model Sources [optional]
|
29 |
|
30 |
<!-- Provide the basic links for the model. -->
|
31 |
|
32 |
-
- **Repository:** [
|
33 |
-
- **
|
34 |
-
- **Demo [optional]:** [More Information Needed]
|
35 |
|
36 |
## Uses
|
37 |
|
38 |
-
<!--
|
39 |
|
40 |
### Direct Use
|
41 |
|
42 |
-
<!-- This
|
43 |
|
44 |
-
[More Information Needed]
|
45 |
|
46 |
### Downstream Use [optional]
|
47 |
|
48 |
-
<!--
|
|
|
49 |
|
50 |
-
[More Information Needed]
|
51 |
|
52 |
### Out-of-Scope Use
|
53 |
|
54 |
-
<!--
|
|
|
55 |
|
56 |
-
[More Information Needed]
|
57 |
|
58 |
## Bias, Risks, and Limitations
|
59 |
|
60 |
-
<!--
|
61 |
|
62 |
-
[More Information Needed]
|
63 |
|
64 |
### Recommendations
|
65 |
|
66 |
-
<!--
|
67 |
|
68 |
-
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
69 |
|
70 |
## How to Get Started with the Model
|
71 |
|
72 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
73 |
|
74 |
-
[More Information Needed]
|
75 |
|
76 |
## Training Details
|
77 |
|
78 |
### Training Data
|
79 |
|
80 |
-
<!--
|
81 |
|
82 |
-
[More Information Needed]
|
83 |
|
84 |
### Training Procedure
|
85 |
|
86 |
-
<!--
|
87 |
|
88 |
-
#### Preprocessing
|
89 |
|
90 |
-
|
91 |
|
92 |
|
93 |
#### Training Hyperparameters
|
94 |
|
95 |
-
- **Training regime:**
|
|
|
|
|
|
|
96 |
|
97 |
-
#### Speeds, Sizes, Times
|
98 |
|
99 |
-
<!--
|
100 |
|
101 |
-
[More Information Needed]
|
102 |
|
103 |
## Evaluation
|
104 |
|
105 |
-
<!-- This section describes the evaluation protocols and provides the results. -->
|
106 |
-
|
107 |
### Testing Data, Factors & Metrics
|
108 |
|
109 |
#### Testing Data
|
110 |
|
111 |
-
<!--
|
112 |
|
113 |
-
[More Information Needed]
|
114 |
-
|
115 |
-
#### Factors
|
116 |
-
|
117 |
-
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
118 |
-
|
119 |
-
[More Information Needed]
|
120 |
|
121 |
#### Metrics
|
122 |
|
123 |
-
<!--
|
124 |
|
125 |
-
[More Information Needed]
|
126 |
|
127 |
### Results
|
128 |
|
129 |
-
|
|
|
|
|
|
|
130 |
|
131 |
#### Summary
|
|
|
132 |
|
133 |
|
134 |
|
135 |
-
## Model Examination [optional]
|
136 |
-
|
137 |
-
<!-- Relevant interpretability work for the model goes here -->
|
138 |
-
|
139 |
-
[More Information Needed]
|
140 |
-
|
141 |
## Environmental Impact
|
142 |
|
143 |
-
<!--
|
144 |
-
|
145 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
146 |
|
147 |
-
- **Hardware Type:** [
|
148 |
-
- **Hours used:** [
|
149 |
-
- **Cloud Provider:** [
|
150 |
-
- **
|
151 |
-
- **Carbon Emitted:** [More Information Needed]
|
152 |
|
153 |
-
## Technical Specifications [
|
154 |
|
155 |
### Model Architecture and Objective
|
156 |
|
157 |
-
[More Information Needed]
|
158 |
-
|
159 |
-
### Compute Infrastructure
|
160 |
-
|
161 |
-
[More Information Needed]
|
162 |
-
|
163 |
#### Hardware
|
164 |
|
165 |
-
[
|
166 |
|
167 |
#### Software
|
168 |
|
169 |
-
[
|
170 |
|
171 |
## Citation [optional]
|
172 |
|
@@ -174,26 +166,23 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
174 |
|
175 |
**BibTeX:**
|
176 |
|
177 |
-
[
|
|
|
|
|
|
|
|
|
|
|
|
|
178 |
|
179 |
**APA:**
|
180 |
|
181 |
-
[
|
182 |
-
|
183 |
-
## Glossary [optional]
|
184 |
-
|
185 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
186 |
-
|
187 |
-
[More Information Needed]
|
188 |
-
|
189 |
-
## More Information [optional]
|
190 |
|
191 |
-
[More Information Needed]
|
192 |
|
193 |
-
## Model Card Authors
|
194 |
|
195 |
-
[
|
196 |
|
197 |
## Model Card Contact
|
198 |
|
199 |
-
[
|
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
+
license: apache-2.0
|
4 |
+
datasets:
|
5 |
+
- Private
|
6 |
+
language:
|
7 |
+
- en
|
8 |
+
metrics:
|
9 |
+
- accuracy
|
10 |
+
- precision
|
11 |
+
- recall
|
12 |
+
- f1
|
13 |
+
base_model: google-bert/bert-base-uncased
|
14 |
+
pipeline_tag: text-classification
|
15 |
---
|
16 |
|
17 |
# Model Card for Model ID
|
18 |
|
19 |
+
<!-- This is a fine-tuned BERT model (`bert-base-uncased`) used for classifying text into two categories: **Depression** or **Non-depression**. The model is designed for text classification and has been trained on a custom dataset of mental health-related posts from social media. -->
|
20 |
|
21 |
|
22 |
|
|
|
24 |
|
25 |
### Model Description
|
26 |
|
27 |
+
<!-- This model aims to identify signs of depression in written text. It was trained on social media posts labeled as either indicative of depression or not. The model uses the BERT architecture for text classification and was fine-tuned specifically for this task. -->
|
28 |
|
29 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
30 |
|
31 |
+
- **Developed by:** Ashish Poudel
|
32 |
+
- **Model type:** Text Classification
|
33 |
+
- **Language(s) (NLP):** English (`en`)
|
34 |
+
- **License:** `apache-2.0`
|
35 |
+
- **Finetuned from model:** `apache-2.0`
|
|
|
|
|
36 |
|
37 |
### Model Sources [optional]
|
38 |
|
39 |
<!-- Provide the basic links for the model. -->
|
40 |
|
41 |
+
- **Repository:** [Sentiment Classifier for Depression](https://huggingface.co/poudel/sentiment-classifier)
|
42 |
+
- **Demo [optional]:** [Live Gradio App](https://huggingface.co/spaces/poudel/Sentiment_classifier)
|
|
|
43 |
|
44 |
## Uses
|
45 |
|
46 |
+
<!-- -->
|
47 |
|
48 |
### Direct Use
|
49 |
|
50 |
+
<!-- This model is designed to classify text as either depression-related or non-depression-related. It can be used in social media sentiment analysis, mental health research, and automated text analysis systems. -->
|
51 |
|
|
|
52 |
|
53 |
### Downstream Use [optional]
|
54 |
|
55 |
+
<!-- The model can be further fine-tuned for other types of sentiment analysis tasks related to mental health. -->
|
56 |
+
|
57 |
|
|
|
58 |
|
59 |
### Out-of-Scope Use
|
60 |
|
61 |
+
<!-- The model should not be used for clinical diagnosis or decision-making without the input of medical professionals. It is also unsuitable for text that is not in English or very short/ambiguous inputs. -->
|
62 |
+
|
63 |
|
|
|
64 |
|
65 |
## Bias, Risks, and Limitations
|
66 |
|
67 |
+
<!-- The model may suffer from biases inherent in the dataset, such as overrepresenting certain language patterns. It is trained on social media posts, which may not capture all the nuances of real-world conversations about mental health -->
|
68 |
|
|
|
69 |
|
70 |
### Recommendations
|
71 |
|
72 |
+
<!-- Users should use the model with caution in sensitive applications such as mental health monitoring. It is advised that the model be used alongside professional judgment. -->
|
73 |
|
|
|
74 |
|
75 |
## How to Get Started with the Model
|
76 |
|
77 |
+
```python
|
78 |
+
from transformers import AutoModelForSequenceClassification, AutoTokenizer
|
79 |
+
|
80 |
+
model = AutoModelForSequenceClassification.from_pretrained("poudel/sentiment-classifier")
|
81 |
+
tokenizer = AutoTokenizer.from_pretrained("poudel/sentiment-classifier")
|
82 |
+
|
83 |
+
inputs = tokenizer("I feel hopeless.", return_tensors="pt")
|
84 |
+
outputs = model(**inputs)
|
85 |
+
predicted_class = torch.argmax(outputs.logits).item()
|
86 |
|
|
|
87 |
|
88 |
## Training Details
|
89 |
|
90 |
### Training Data
|
91 |
|
92 |
+
<!-- The model was trained on a custom dataset of tweets labeled as either depression-related or not. Data pre-processing included tokenization and removal of special characters. -->
|
93 |
|
|
|
94 |
|
95 |
### Training Procedure
|
96 |
|
97 |
+
<!-- The model was trained using Hugging Face's transformers library. The training was conducted on a T4 GPU over 3 epochs, with a batch size of 16 and a learning rate of 5e-5. -->
|
98 |
|
99 |
+
#### Preprocessing
|
100 |
|
101 |
+
<!-- Text was lowercased, and special characters were removed as well as Tokenization was done using the bert-base-uncased tokenizer.-->
|
102 |
|
103 |
|
104 |
#### Training Hyperparameters
|
105 |
|
106 |
+
- **Training regime:** <!--fp32 -->
|
107 |
+
- **Epochs:** <!--3 -->
|
108 |
+
- ** Learning rate:** <!--5e-5-->
|
109 |
+
- **Batch size:** <!--16 -->
|
110 |
|
111 |
+
#### Speeds, Sizes, Times
|
112 |
|
113 |
+
<!--Training was conducted for approximately 1 hour on a T4 GPU in Google Colab. -->
|
114 |
|
|
|
115 |
|
116 |
## Evaluation
|
117 |
|
|
|
|
|
118 |
### Testing Data, Factors & Metrics
|
119 |
|
120 |
#### Testing Data
|
121 |
|
122 |
+
<!-- The model was evaluated on a 20% holdout set from the custom dataset. -->
|
123 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
124 |
|
125 |
#### Metrics
|
126 |
|
127 |
+
<!-- The model was evaluated using accuracy, precision, recall, and F1 score. -->
|
128 |
|
|
|
129 |
|
130 |
### Results
|
131 |
|
132 |
+
Accuracy: 99.87%
|
133 |
+
Precision: 99.91%
|
134 |
+
Recall: 99.81%
|
135 |
+
F1 Score: 99.86%
|
136 |
|
137 |
#### Summary
|
138 |
+
The model achieved high performance across all key metrics, indicating strong predictive capabilities for the text classification task.
|
139 |
|
140 |
|
141 |
|
|
|
|
|
|
|
|
|
|
|
|
|
142 |
## Environmental Impact
|
143 |
|
144 |
+
<!-- Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). -->
|
|
|
|
|
145 |
|
146 |
+
- **Hardware Type:** [T4 GPU]
|
147 |
+
- **Hours used:** [ 1 hour]
|
148 |
+
- **Cloud Provider:** [Google Cloud (Colab)]
|
149 |
+
- **Carbon Emitted:** [Estimated at 0.45 kg CO2eq]
|
|
|
150 |
|
151 |
+
## Technical Specifications [The model uses the BERT (bert-base-uncased) architecture and was fine-tuned for binary classification (depression vs non-depression).]
|
152 |
|
153 |
### Model Architecture and Objective
|
154 |
|
|
|
|
|
|
|
|
|
|
|
|
|
155 |
#### Hardware
|
156 |
|
157 |
+
[T4 GPU]
|
158 |
|
159 |
#### Software
|
160 |
|
161 |
+
[Hugging Face transformers library.]
|
162 |
|
163 |
## Citation [optional]
|
164 |
|
|
|
166 |
|
167 |
**BibTeX:**
|
168 |
|
169 |
+
[@misc{poudel2024sentimentclassifier,
|
170 |
+
author = {Poudel, Ashish},
|
171 |
+
title = {Sentiment Classifier for Depression},
|
172 |
+
year = {2024},
|
173 |
+
url = {https://huggingface.co/poudel/sentiment-classifier},
|
174 |
+
}
|
175 |
+
]
|
176 |
|
177 |
**APA:**
|
178 |
|
179 |
+
[Poudel, A. (2024). Sentiment Classifier for Depression. Retrieved from https://huggingface.co/poudel/sentiment-classifier.]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
180 |
|
|
|
181 |
|
182 |
+
## Model Card Authors
|
183 |
|
184 |
+
[Ashish Poudel]
|
185 |
|
186 |
## Model Card Contact
|
187 |
|
188 |