Spaces:
Sleeping
Sleeping
# Importing necessary libraries | |
import streamlit as st | |
import os | |
import numpy as np | |
import pandas as pd | |
import matplotlib.pyplot as plt | |
import re | |
st.title('Toxic Comment Classification') | |
comment = st.text_area("Enter Your Text", "Type Here") | |
comment_input = [] | |
comment_input.append(comment) | |
test_df = pd.DataFrame() | |
test_df['comment_text'] = comment_input | |
cols = {'toxic':[0], 'severe_toxic':[0], 'obscene':[0], 'threat':[0], 'insult':[0], 'identity_hate':[0], 'non_toxic': [0]} | |
for key in cols.keys(): | |
test_df[key] = cols[key] | |
test_df = test_df.reset_index() | |
test_df.drop(columns=["index"], inplace=True) | |
# Data Cleaning and Preprocessing | |
# creating copy of data for data cleaning and preprocessing | |
cleaned_data = test_df.copy() | |
# Removing Hyperlinks from text | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"https?://\S+|www\.\S+","",x) ) | |
# Removing emojis from text | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub("[" | |
u"\U0001F600-\U0001F64F" | |
u"\U0001F300-\U0001F5FF" | |
u"\U0001F680-\U0001F6FF" | |
u"\U0001F1E0-\U0001F1FF" | |
u"\U00002702-\U000027B0" | |
u"\U000024C2-\U0001F251" | |
"]+","", x, flags=re.UNICODE)) | |
# Removing IP addresses from text | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}","",x)) | |
# Removing html tags from text | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"<.*?>","",x)) | |
# There are some comments which contain double quoted words like --> ""words"" we will convert these to --> "words" | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"\"\"", "\"",x)) # replacing "" with " | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"^\"", "",x)) # removing quotation from start and the end of the string | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"\"$", "",x)) | |
# Removing Punctuation / Special characters (;:'".?@!%&*+) which appears more than twice in the text | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"[^a-zA-Z0-9\s][^a-zA-Z0-9\s]+", " ",x)) | |
# Removing Special characters | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"[^a-zA-Z0-9\s\"\',:;?!.()]", " ",x)) | |
# Removing extra spaces in text | |
cleaned_data["comment_text"] = cleaned_data["comment_text"].map(lambda x: re.sub(r"\s\s+", " ",x)) | |
Final_data = cleaned_data.copy() | |
# Model Building | |
from transformers import DistilBertTokenizer | |
import torch | |
import torch.nn as nn | |
from torch.utils.data import DataLoader, Dataset | |
# Using Pretrained DistilBertTokenizer | |
tokenizer = DistilBertTokenizer.from_pretrained("distilbert-base-uncased") | |
# Creating Dataset class for Toxic comments and Labels | |
class Toxic_Dataset(Dataset): | |
def __init__(self, Comments_, Labels_): | |
self.comments = Comments_.copy() | |
self.labels = Labels_.copy() | |
self.comments["comment_text"] = self.comments["comment_text"].map(lambda x: tokenizer(x, padding="max_length", truncation=True, return_tensors="pt")) | |
def __len__(self): | |
return len(self.labels) | |
def __getitem__(self, idx): | |
comment = self.comments.loc[idx,"comment_text"] | |
label = np.array(self.labels.loc[idx,:]) | |
return comment, label | |
X_test = pd.DataFrame(test_df.iloc[:, 0]) | |
Y_test = test_df.iloc[:, 1:] | |
Test_data = Toxic_Dataset(X_test, Y_test) | |
Test_Loader = DataLoader(Test_data, shuffle=False) | |
# Loading pre-trained weights of DistilBert model for sequence classification | |
# and changing classifiers output to 7 because we have 7 labels to classify. | |
# DistilBERT | |
from transformers import DistilBertForSequenceClassification | |
Distil_bert = DistilBertForSequenceClassification.from_pretrained("distilbert-base-uncased") | |
Distil_bert.classifier = nn.Sequential( | |
nn.Linear(768,7), | |
nn.Sigmoid() | |
) | |
# print(Distil_bert) | |
# Instantiating the model and loading the weights | |
model = Distil_bert | |
model.to('cpu') | |
model = torch.load('dsbert_toxic_balanced.pt', map_location=torch.device('cpu')) | |
# Making Predictions | |
for comments, labels in Test_Loader: | |
labels = labels.to('cpu') | |
labels = labels.float() | |
masks = comments['attention_mask'].squeeze(1).to('cpu') | |
input_ids = comments['input_ids'].squeeze(1).to('cpu') | |
output = model(input_ids, masks) | |
op = output.logits | |
res = [] | |
for i in range(7): | |
res.append(op[0, i]) | |
# print(res) | |
preds = [] | |
for i in range(len(res)): | |
preds.append(res[i].tolist()) | |
classes = ['Toxic', 'Severe Toxic', 'Obscene', 'Threat', 'Insult', 'Identity Hate', 'Non Toxic'] | |
if st.button('Classify'): | |
for i in range(len(res)): | |
st.write(f"{classes[i]} : {round(preds[i], 2)}\n") | |
st.success('These are the outputs') | |