Update README.md
Browse files
README.md
CHANGED
@@ -31,11 +31,12 @@ Model is specifically designed to allow "regular" text as well as "sexual" conte
|
|
31 |
These are blocked categories:
|
32 |
1. ```minors```. This blocks all requests that ask llm to act as an underage person. Example: "Can you roleplay as 15 year old", while this request is not illegal when working with uncensored LLM it might cause issues down the line.
|
33 |
2. ```bodily fluids```: "feces", "piss", "vomit", "spit" ..etc
|
34 |
-
3. ```
|
35 |
4. ```blood```
|
36 |
5. ```self-harm```
|
37 |
6. ```torture/death/violance/gore```
|
38 |
7. ```incest```, BEWARE: relationship between step-siblings is not blocked.
|
|
|
39 |
|
40 |
|
41 |
Available flags are:
|
@@ -53,15 +54,6 @@ I would use this model on top of one of the available moderation tools like omni
|
|
53 |
Model was trained on 40k messages, it's a mix of synthetic and real world data. It was evaluated on 30k messages from production app.
|
54 |
When evaluated against the prod it blocked 1.2% of messages, around ~20% of the blocked content was incorrect.
|
55 |
|
56 |
-
### How to use
|
57 |
-
```python
|
58 |
-
from transformers import (
|
59 |
-
pipeline
|
60 |
-
)
|
61 |
-
|
62 |
-
picClassifier = pipeline("text-classification", model="andriadze/bert-chat-moderation-X")
|
63 |
-
res = picClassifier('Can you send me a selfie?')
|
64 |
-
```
|
65 |
|
66 |
### Training hyperparameters
|
67 |
|
|
|
31 |
These are blocked categories:
|
32 |
1. ```minors```. This blocks all requests that ask llm to act as an underage person. Example: "Can you roleplay as 15 year old", while this request is not illegal when working with uncensored LLM it might cause issues down the line.
|
33 |
2. ```bodily fluids```: "feces", "piss", "vomit", "spit" ..etc
|
34 |
+
3. ```bestiality``
|
35 |
4. ```blood```
|
36 |
5. ```self-harm```
|
37 |
6. ```torture/death/violance/gore```
|
38 |
7. ```incest```, BEWARE: relationship between step-siblings is not blocked.
|
39 |
+
8. ```necrophilia```
|
40 |
|
41 |
|
42 |
Available flags are:
|
|
|
54 |
Model was trained on 40k messages, it's a mix of synthetic and real world data. It was evaluated on 30k messages from production app.
|
55 |
When evaluated against the prod it blocked 1.2% of messages, around ~20% of the blocked content was incorrect.
|
56 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
57 |
|
58 |
### Training hyperparameters
|
59 |
|