Spaces:
Running
on
Zero
Running
on
Zero
# Profanity Detection in Speech and Text | |
A robust multimodal system for detecting and rephrasing profanity in both speech and text, leveraging advanced NLP models to ensure accurate filtering while preserving conversational context. | |
 | |
 | |
 | |
## π Live Demo | |
Try the system without installation via our Hugging Face Spaces deployment: | |
[](https://huggingface.co/spaces/nightey3s/profanity-detection) | |
<img src="https://briantham.com/assets/img/projects/qr-code/Profanity-Detection-huggingface-qr-code.svg?sanitize=true" alt="QR Code" width="300" /> | |
This live version leverages Hugging Face's ZeroGPU technology, which provides on-demand GPU acceleration for inference while optimising resource usage. | |
## π Features | |
- **Multimodal Analysis**: Process both written text and spoken audio | |
- **Context-Aware Detection**: Goes beyond simple keyword matching | |
- **Automatic Content Refinement**: Intelligently rephrases content while preserving meaning | |
- **Audio Synthesis**: Converts rephrased content into high-quality spoken audio | |
- **Classification System**: Categorises content by toxicity levels | |
- **User-Friendly Interface**: Intuitive Gradio-based UI | |
- **Real-time Streaming**: Process audio in real-time as you speak | |
- **Adjustable Sensitivity**: Fine-tune profanity detection threshold | |
- **Visual Highlighting**: Instantly identify problematic words with visual highlighting | |
- **Toxicity Classification**: Automatically categorize content from "No Toxicity" to "Severe Toxicity" | |
- **Performance Optimization**: Half-precision support for improved GPU memory efficiency | |
- **Cloud Deployment**: Available as a hosted service on Hugging Face Spaces | |
## π§ Models Used | |
The system leverages four powerful models: | |
1. **Profanity Detection**: `parsawar/profanity_model_3.1` - A RoBERTa-based model trained for offensive language detection | |
2. **Content Refinement**: `s-nlp/t5-paranmt-detox` - A T5-based model for rephrasing offensive language | |
3. **Speech-to-Text**: OpenAI's `Whisper` (large-v2) - For transcribing spoken audio | |
4. **Text-to-Speech**: Microsoft's `SpeechT5` - For converting rephrased text back to audio | |
## π Deployment Options | |
### Online Deployment (No Installation Required) | |
Access the application directly through Hugging Face Spaces: | |
- **URL**: [https://huggingface.co/spaces/nightey3s/profanity-detection](https://huggingface.co/spaces/nightey3s/profanity-detection) | |
- **Technology**: Built with ZeroGPU for efficient GPU resource allocation | |
- **Features**: All features of the full application accessible through your browser | |
- **Source Code**: [GitHub Repository](https://github.com/Nightey3s/profanity-detection) | |
### Local Installation | |
#### Prerequisites | |
- Python 3.10+ | |
- CUDA-compatible GPU recommended (but CPU mode works too) | |
- FFmpeg for audio processing | |
#### Option 1: Using Conda (Recommended for Local Development) | |
```bash | |
# Clone the repository | |
git clone https://github.com/Nightey3s/profanity-detection.git | |
cd profanity-detection | |
# Method A: Create environment from environment.yml (recommended) | |
conda env create -f environment.yml | |
conda activate llm_project | |
# Method B: Create a new conda environment manually | |
conda create -n profanity-detection python=3.10 | |
conda activate profanity-detection | |
# Install PyTorch with CUDA support (adjust CUDA version if needed) | |
conda install pytorch torchvision torchaudio pytorch-cuda=11.8 -c pytorch -c nvidia | |
# Install FFmpeg for audio processing | |
conda install -c conda-forge ffmpeg | |
# Install Pillow properly to avoid DLL errors | |
conda install -c conda-forge pillow | |
# Install additional dependencies | |
pip install -r requirements.txt | |
# Set environment variable to avoid OpenMP conflicts (recommended) | |
conda env config vars set KMP_DUPLICATE_LIB_OK=TRUE | |
conda activate profanity-detection # Re-activate to apply the variable | |
``` | |
#### Option 2: Using Docker | |
```bash | |
# Clone the repository | |
git clone https://github.com/Nightey3s/profanity-detection.git | |
cd profanity-detection | |
# Build and run the Docker container | |
docker-compose build --no-cache | |
docker-compose up | |
``` | |
## π§ Usage | |
### Using the Online Interface (Hugging Face Spaces) | |
1. Visit [https://huggingface.co/spaces/nightey3s/profanity-detection](https://huggingface.co/spaces/nightey3s/profanity-detection) | |
2. The interface might take a moment to load on first access as it allocates resources | |
3. Follow the same usage instructions as below, starting with "Initialize Models" | |
### Using the Local Interface | |
1. **Initialise Models** | |
- Click the "Initialize Models" button when you first open the interface | |
- Wait for all models to load (this may take a few minutes on first run) | |
2. **Text Analysis Tab** | |
- Enter text into the text box | |
- Adjust the "Profanity Detection Sensitivity" slider if needed | |
- Click "Analyze Text" | |
- View results including profanity score, toxicity classification, and rephrased content | |
- See highlighted profane words in the text | |
- Listen to the audio version of the rephrased content | |
3. **Audio Analysis Tab** | |
- Upload an audio file or record directly using your microphone | |
- Click "Analyze Audio" | |
- View transcription, profanity analysis, and rephrased content | |
- Listen to the cleaned audio version of the rephrased content | |
4. **Real-time Streaming Tab** | |
- Click "Start Real-time Processing" | |
- Speak into your microphone | |
- Watch as your speech is transcribed, analyzed, and rephrased in real-time | |
- Listen to the clean audio output | |
- Click "Stop Real-time Processing" when finished | |
## β οΈ Troubleshooting | |
### OpenMP Runtime Conflict | |
If you encounter this error: | |
``` | |
OMP: Error #15: Initializing libiomp5md.dll, but found libiomp5md.dll already initialized. | |
``` | |
**Solutions:** | |
1. **Temporary fix**: Set environment variable before running: | |
```bash | |
set KMP_DUPLICATE_LIB_OK=TRUE # Windows | |
export KMP_DUPLICATE_LIB_OK=TRUE # Linux/Mac | |
``` | |
2. **Code-based fix**: Add to the beginning of your script: | |
```python | |
import os | |
os.environ['KMP_DUPLICATE_LIB_OK'] = 'TRUE' | |
``` | |
3. **Permanent fix for Conda environment**: | |
```bash | |
conda env config vars set KMP_DUPLICATE_LIB_OK=TRUE -n profanity-detection | |
conda deactivate | |
conda activate profanity-detection | |
``` | |
### GPU Memory Issues | |
If you encounter CUDA out of memory errors: | |
1. Use smaller models: | |
```python | |
# Change Whisper from "large" to "medium" or "small" | |
whisper_model = whisper.load_model("medium").to(device) | |
# Keep the TTS model on CPU to save GPU memory | |
tts_model = SpeechT5ForTextToSpeech.from_pretrained(TTS_MODEL) # CPU mode | |
``` | |
2. Run some models on CPU instead of GPU: | |
```python | |
# Remove .to(device) to keep model on CPU | |
t5_model = AutoModelForSeq2SeqLM.from_pretrained(T5_MODEL) # CPU mode | |
``` | |
3. Use Docker with specific GPU memory limits: | |
```yaml | |
# In docker-compose.yml | |
deploy: | |
resources: | |
reservations: | |
devices: | |
- driver: nvidia | |
count: 1 | |
capabilities: [gpu] | |
options: | |
memory: 4G # Limit to 4GB of GPU memory | |
``` | |
### Hugging Face Spaces-Specific Issues | |
1. **Long initialization time**: The first time you access the Space, it may take longer to initialize as models are downloaded and cached. | |
2. **Timeout errors**: If the model takes too long to process your request, try again with shorter text or audio inputs. | |
3. **Browser compatibility**: Ensure your browser allows microphone access for audio recording features. | |
### First-Time Slowness | |
When first run, the application downloads all models, which may take time. Subsequent runs will be faster as models are cached locally. The text-to-speech model requires additional download time on first use. | |
## π Project Structure | |
``` | |
profanity-detection/ | |
βββ profanity_detector.py # Main application file | |
βββ Dockerfile # For containerised deployment | |
βββ docker-compose.yml # Container orchestration | |
βββ requirements.txt # Python dependencies | |
βββ environment.yml # Conda environment specification | |
βββ README.md # This file | |
``` | |
## Team Members | |
- Brian Tham | |
- Hong Ziyang | |
- Nabil Zafran | |
- Adrian Ian Wong | |
- Lin Xiang Hong | |
## π References | |
- [HuggingFace Transformers](https://huggingface.co/docs/transformers/index) | |
- [OpenAI Whisper](https://github.com/openai/whisper) | |
- [Microsoft SpeechT5](https://huggingface.co/microsoft/speecht5_tts) | |
- [Gradio Documentation](https://gradio.app/docs/) | |
- [Hugging Face Spaces](https://huggingface.co/spaces) | |
## π License | |
This project is licensed under the MIT License - see the LICENSE file for details. | |
## π Acknowledgments | |
- This project utilises models from HuggingFace Hub, Microsoft, and OpenAI | |
- Inspired by research in content moderation and responsible AI | |
- Hugging Face for providing the Spaces platform with ZeroGPU technology |