andriadze commited on
Commit
8117c93
·
verified ·
1 Parent(s): 08ed95d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -10
README.md CHANGED
@@ -31,11 +31,12 @@ Model is specifically designed to allow "regular" text as well as "sexual" conte
31
  These are blocked categories:
32
  1. ```minors```. This blocks all requests that ask llm to act as an underage person. Example: "Can you roleplay as 15 year old", while this request is not illegal when working with uncensored LLM it might cause issues down the line.
33
  2. ```bodily fluids```: "feces", "piss", "vomit", "spit" ..etc
34
- 3. ```beastiality```
35
  4. ```blood```
36
  5. ```self-harm```
37
  6. ```torture/death/violance/gore```
38
  7. ```incest```, BEWARE: relationship between step-siblings is not blocked.
 
39
 
40
 
41
  Available flags are:
@@ -53,15 +54,6 @@ I would use this model on top of one of the available moderation tools like omni
53
  Model was trained on 40k messages, it's a mix of synthetic and real world data. It was evaluated on 30k messages from production app.
54
  When evaluated against the prod it blocked 1.2% of messages, around ~20% of the blocked content was incorrect.
55
 
56
- ### How to use
57
- ```python
58
- from transformers import (
59
- pipeline
60
- )
61
-
62
- picClassifier = pipeline("text-classification", model="andriadze/bert-chat-moderation-X")
63
- res = picClassifier('Can you send me a selfie?')
64
- ```
65
 
66
  ### Training hyperparameters
67
 
 
31
  These are blocked categories:
32
  1. ```minors```. This blocks all requests that ask llm to act as an underage person. Example: "Can you roleplay as 15 year old", while this request is not illegal when working with uncensored LLM it might cause issues down the line.
33
  2. ```bodily fluids```: "feces", "piss", "vomit", "spit" ..etc
34
+ 3. ```bestiality``
35
  4. ```blood```
36
  5. ```self-harm```
37
  6. ```torture/death/violance/gore```
38
  7. ```incest```, BEWARE: relationship between step-siblings is not blocked.
39
+ 8. ```necrophilia```
40
 
41
 
42
  Available flags are:
 
54
  Model was trained on 40k messages, it's a mix of synthetic and real world data. It was evaluated on 30k messages from production app.
55
  When evaluated against the prod it blocked 1.2% of messages, around ~20% of the blocked content was incorrect.
56
 
 
 
 
 
 
 
 
 
 
57
 
58
  ### Training hyperparameters
59