ksasse-uab Satya10 commited on
Commit
0d9a58a
·
verified ·
0 Parent(s):

Duplicate from Hate-speech-CNERG/bert-base-uncased-hatexplain

Browse files

Co-authored-by: Satya <[email protected]>

.gitattributes ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ *.bin.* filter=lfs diff=lfs merge=lfs -text
2
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.h5 filter=lfs diff=lfs merge=lfs -text
5
+ *.tflite filter=lfs diff=lfs merge=lfs -text
6
+ *.tar.gz filter=lfs diff=lfs merge=lfs -text
7
+ *.ot filter=lfs diff=lfs merge=lfs -text
8
+ *.onnx filter=lfs diff=lfs merge=lfs -text
9
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ license: apache-2.0
4
+ datasets:
5
+ - hatexplain
6
+ ---
7
+ The model is used for classifying a text as **Hatespeech**, **Offensive**, or **Normal**. The model is trained using data from Gab and Twitter and *Human Rationales* were included as part of the training data to boost the performance.
8
+
9
+ The dataset and models are available here: https://github.com/punyajoy/HateXplain
10
+
11
+
12
+ **For more details about our paper**
13
+
14
+ Binny Mathew, Punyajoy Saha, Seid Muhie Yimam, Chris Biemann, Pawan Goyal, and Animesh Mukherjee "[HateXplain: A Benchmark Dataset for Explainable Hate Speech Detection)". Accepted at AAAI 2021.
15
+
16
+ ***Please cite our paper in any published work that uses any of these resources.***
17
+
18
+ ~~~
19
+ @article{mathew2020hatexplain,
20
+ title={HateXplain: A Benchmark Dataset for Explainable Hate Speech Detection},
21
+ author={Mathew, Binny and Saha, Punyajoy and Yimam, Seid Muhie and Biemann, Chris and Goyal, Pawan and Mukherjee, Animesh},
22
+ journal={arXiv preprint arXiv:2012.10289},
23
+ year={2020}
24
+ }
25
+ ~~~
config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": null,
7
+ "do_sample": false,
8
+ "eos_token_ids": null,
9
+ "finetuning_task": null,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "hate speech",
15
+ "1": "normal",
16
+ "2": "offensive"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 3072,
20
+ "is_decoder": false,
21
+ "label2id": {
22
+ "hate speech": 0,
23
+ "normal": 1,
24
+ "offensive":2
25
+ },
26
+ "layer_norm_eps": 1e-12,
27
+ "length_penalty": 1.0,
28
+ "max_length": 20,
29
+ "max_position_embeddings": 512,
30
+ "model_type": "bert",
31
+ "num_attention_heads": 12,
32
+ "num_beams": 1,
33
+ "num_hidden_layers": 12,
34
+ "num_labels": 3,
35
+ "num_return_sequences": 1,
36
+ "output_attentions": true,
37
+ "output_hidden_states": false,
38
+ "output_past": true,
39
+ "pad_token_id": 0,
40
+ "pruned_heads": {},
41
+ "repetition_penalty": 1.0,
42
+ "temperature": 1.0,
43
+ "top_k": 50,
44
+ "top_p": 1.0,
45
+ "torchscript": false,
46
+ "type_vocab_size": 2,
47
+ "use_bfloat16": false,
48
+ "vocab_size": 30522
49
+ }
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc92239e1ca3c00bf29066fd2422d44134eddd96870b95e531cffcb3db41ab6b
3
+ size 437945404
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7939d499f5c2ec887db9c09d9e790f33321686a4993c6b3da7d89db059a7452e
3
+ size 437988463
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "max_len": 512}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff