DarwinAnim8or commited on
Commit
5cf9217
1 Parent(s): 177884b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -4
README.md CHANGED
@@ -2,20 +2,46 @@
2
  tags:
3
  - autotrain
4
  - text-classification
 
 
 
5
  language:
6
  - en
7
  widget:
8
- - text: "I love AutoTrain"
 
9
  datasets:
10
- - DarwinAnim8or/autotrain-data-offensivespeechdetector
11
  co2_eq_emissions:
12
  emissions: 0.010817089812320756
 
13
  ---
14
 
15
- # Model Trained Using AutoTrain
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
  - Problem type: Multi-class Classification
18
- - Model ID: 82680142077
19
  - CO2 Emissions (in grams): 0.0108
20
 
21
  ## Validation Metrics
 
2
  tags:
3
  - autotrain
4
  - text-classification
5
+ - social
6
+ - offensive speech detection
7
+ - moderation
8
  language:
9
  - en
10
  widget:
11
+ - text: I love cake!
12
+ - text: I hate bananas!
13
  datasets:
14
+ - tweet_eval
15
  co2_eq_emissions:
16
  emissions: 0.010817089812320756
17
+ license: openrail
18
  ---
19
 
20
+ # Offensive Speech Detector
21
+ "Offensive Speech Detector" is a text classification model based on Deberta that predicts whether a text contains offensive language or not.
22
+ The model is fine-tuned on the tweet_eval dataset, which consists of seven heterogeneous tasks in Twitter, all framed as multi-class tweet classification. The 'offensive' subset is used for this task.
23
+
24
+ ## Intended uses & limitations
25
+
26
+ Offensive Speech Detector is intended to be used as a tool for detecting offensive language in texts, which can be useful for applications such as content moderation, sentiment analysis, or social media analysis. The model can be used to filter out or flag tweets that contain offensive language, or to analyze the prevalence and patterns of offensive language.
27
+
28
+ However, the model has some limitations that users should be aware of:
29
+
30
+ - The model is only trained and evaluated on tweets, which are short and informal texts that may contain slang, abbreviations, emojis, hashtags, or user mentions. The model may not perform well on other types of texts, such as news articles, essays, or books.
31
+ - The model is only trained and evaluated on English tweets. The model may not generalize well to other languages or dialects.
32
+ - The model is based on the tweet_eval dataset, which may have some biases or errors in the annotation process. The labels are assigned by human annotators, who may have different opinions or criteria for what constitutes offensive language. The dataset may also not cover all possible forms or contexts of offensive language, such as sarcasm, irony, humor, or euphemism.
33
+ - The model is a statistical classifier that outputs a probability score for each label. The model does not provide any explanation or justification for its predictions. The model may also make mistakes or produce false positives or false negatives. Users should not blindly trust the model's predictions without further verification or human oversight.
34
+
35
+ ## Ethical Considerations
36
+ This is a model that deals with sensitive and potentially harmful language. Users should consider the ethical implications and potential risks of using or deploying this model in their applications or contexts. Some of the ethical issues that may arise are:
37
+
38
+ - The model may reinforce or amplify existing biases or stereotypes in the data or in the society. For example, the model may associate certain words or topics with offensive language based on the frequency or co-occurrence in the data, without considering the meaning or intent behind them. This may result in unfair or inaccurate predictions for some groups or individuals.
39
+
40
+ Users should carefully consider the purpose, context, and impact of using this model, and take appropriate measures to prevent or mitigate any potential harm. Users should also respect the privacy and consent of the data subjects, and adhere to the relevant laws and regulations in their jurisdictions.
41
+
42
+ ## Model Training Info
43
 
44
  - Problem type: Multi-class Classification
 
45
  - CO2 Emissions (in grams): 0.0108
46
 
47
  ## Validation Metrics