Datasets:

Modalities:
Tabular
Text
Formats:
json
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
x_sensitive / README.md
antypasd's picture
Update README.md
7a2acc2 verified
metadata
task_categories:
  - text-classification
language:
  - en
size_categories:
  - 1K<n<10K
extra_gated_prompt: >-
  This dataset includes sensitive and explicit material.  You may encounter 
  descriptions that are sexually explicit, graphic, or otherwise unsuitable for
  all audiences. Viewer discretion is strongly advised. Please proceed only if
  you are comfortable and consent to viewing this type of content. You agree to
  not use the dataset to conduct experiments that cause harm to human subjects.
extra_gated_fields:
  Company: text
  Country: country
  I want to use this dataset for:
    type: select
    options:
      - Research
      - Education
      - label: Other
        value: other
  I agree to use this dataset for non-commercial use ONLY: checkbox
license: cc-by-2.0

X-Sensitive is a multi-label dataset designed to identify sensitive language in social media. It consists of 7 labels and includes a total of 8,000 posts extracted from X. Each post is assigned one or more of the following labels based on its content: Drugs, Sex, Conflictual, Spam, Profanity, and Self-harm. More details in the reference paper.

The goal of X-Sensitive is to serve as a valuable resource for developing online moderation tools. The following models have been trained on X-Sensitive with this aim:

We also provide binary versions of the models, where each post is classified as either sensitive or not-sensitive:

Dataset Structure

Data Splits

Name #Entries
train 5,000
test 2,000
validation 1,000

Data Instances

An example of train looks as follows.

{'#labels': 1,
 'conflictual': 0,
 'conflictual_highlight': array([], dtype=object),
 'drugs': 0,
 'drugs_highlight': array([], dtype=object),
 'keyword': 'fuckin',
 'labels': array(['profanity'], dtype=object),
 'profanity': 1,
 'profanity_highlight': array([array(['fucking'], dtype=object), array(['fucking'], dtype=object),
       array(['fucking'], dtype=object)], dtype=object),
 'selfharm': 0,
 'selfharm_highlight': array([], dtype=object),
 'sex': 0,
 'sex_highlight': array([], dtype=object),
 'spam': 0,
 'spam_highlight': array([], dtype=object),
 'text': 'i think the idea of aliens is so fucking cool'}

Labels

Label Number Label Name Description
0 conflictual Conflictual language. An attack based on protected (race, color, caster, gender, etc) or other categories.
1 profanity Language containing slurs and profanity even if they are not directed towards a specific entity.
2 sex Sexually Explicit Content. Pornographic or other types of sexual content.
4 selfharm Self-harm. Posts depicting, promoting, or glorifying violence or harm against oneself such as eating disorders or suicide.
5 spam Irrelevant content that is unsolicited.

Citation Information

@article{antypas2024sensitive,
  title={Sensitive Content Classification in Social Media: A Holistic Resource and Evaluation},
  author={Antypas, Dimosthenis and Sen, Indira and Perez-Almendros, Carla and Camacho-Collados, Jose and Barbieri, Francesco},
  journal={arXiv preprint arXiv:2411.19832},
  year={2024}
}