Spaces:
Running
Running
title: README | |
emoji: π¨ | |
colorFrom: blue | |
colorTo: red | |
sdk: static | |
pinned: false | |
# π OverseerAI | |
## Mission | |
OverseerAI is dedicated to advancing open-source AI safety and content moderation tools. We develop state-of-the-art models and datasets for brand safety classification, making content moderation more accessible and efficient for developers and organizations. | |
## π Our Projects | |
### Datasets | |
#### [BrandSafe-16k](https://huggingface.co/datasets/OverseerAI/BrandSafe-16k) | |
A comprehensive dataset for training brand safety classification models, featuring 16 distinct risk categories: | |
| Category | Description | | |
|----------|-------------| | |
| B1-PROFANITY | Explicit language and cursing | | |
| B2-OFFENSIVE_SLANG | Informal offensive terms | | |
| B3-COMPETITOR | Competitive brand mentions | | |
| B4-BRAND_CRITICISM | Negative brand commentary | | |
| B5-MISLEADING | Deceptive or false information | | |
| B6-POLITICAL | Political content and discussions | | |
| B7-RELIGIOUS | Religious themes and references | | |
| B8-CONTROVERSIAL | Contentious topics | | |
| B9-ADULT | Adult or mature content | | |
| B10-VIOLENCE | Violent themes or descriptions | | |
| B11-SUBSTANCE | Drug and alcohol references | | |
| B12-HATE | Hate speech and discrimination | | |
| B13-STEREOTYPE | Stereotypical content | | |
| B14-BIAS | Biased viewpoints | | |
| B15-UNPROFESSIONAL | Unprofessional content | | |
| B16-MANIPULATION | Manipulative content | | |
### Models | |
#### [vision-1](https://huggingface.co/OverseerAI/vision-1) | |
Our flagship model for brand safety classification: | |
- Architecture: Meta Llama 3.1 (15GB) | |
- Full precision model optimized for high accuracy | |
- Trained on BrandSafe-16k dataset | |
- Ideal for production deployments with high-end GPU resources | |
#### [vision-1-mini](https://huggingface.co/OverseerAI/vision-1-mini) | |
A lightweight, optimized version of vision-1: | |
- Size: 4.58 GiB | |
- Architecture: Llama 3.1 8B | |
- Quantization: GGUF V3 (Q4_K) | |
- Optimized for Apple Silicon | |
- Fast load time: 3.27s | |
- Efficient memory usage: 4552.80 MiB CPU / 132.50 MiB Metal | |
- Perfect for local deployment and smaller compute resources | |
## π‘ Use Cases | |
- Content moderation for social media platforms | |
- Brand safety monitoring for advertising | |
- User-generated content filtering | |
- Real-time content classification | |
- Safe content recommendation systems | |
## π€ Contributing | |
We welcome contributions from the community! Whether it's: | |
- Improving model accuracy | |
- Expanding the dataset | |
- Optimizing for different hardware | |
- Adding new classification categories | |
- Reporting issues or suggesting improvements | |
## π« Contact | |
- GitHub: [OverseerAI](https://github.com/OverseerAI) | |
- HuggingFace: [OverseerAI](https://huggingface.co/OverseerAI) | |
## π License | |
Our models are released under the Llama 3.1 license, and our datasets are available under open-source licenses to promote accessibility and innovation in AI safety. | |
--- | |
*OverseerAI - Making AI Safety Accessible and Efficient* |