Model Details

This model is built on the DistilBERT architecture, specifically utilizing the distilbert-base-uncased variant, and is designed to classify text into two categories: statements and questions. It leverages the strengths of the DistilBERT model, known for its efficiency and performance, to accurately discern between declarative statements and interrogative questions.

Model Description

The model processes input text to determine whether it is a statement or a question.

Training Data

The model was trained on a diverse dataset containing examples of both statements and questions. The training process involved fine-tuning the pre-trained DistilBERT model on this specific classification task. The dataset included various types of questions and statements from different contexts to ensure robustness.

  • Quora Question Keyword Pairs
  • Questions vs Statements Classification
  • ilert related Questions

Performance

The performance of the model was evaluated using standard metrics for classification tasks, including accuracy, precision, recall, and F1 score. The results indicate that the model performs well in distinguishing between statements and questions, making it a reliable tool for text classification tasks in natural language processing.

Usage

To use this model, you can load it through the Hugging Face transformers library and use it for text classification. Here is an example of how to use the model in Python:

from transformers import pipeline

# Load the model and tokenizer
classifier = pipeline("text-classification", model="ilert/SoQbert")

# Example texts
texts = ["Is it going to rain today?", "It is a sunny day."]

# Classify texts
results = classifier(texts)

# Output the results
for text, result in zip(texts, results):
    print(f"Text: {text}")
    print(f"Classification: {result['label']}")
Downloads last month
16
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ilert/SoQbert

Quantized
(23)
this model