Ambig_Question / README.md
Slomb's picture
Update README.md
4e0d83b verified
|
raw
history blame
915 Bytes
metadata
license: apache-2.0
datasets:
  - gexai/inquisitiveqg
language:
  - en
metrics:
  - accuracy
base_model:
  - distilbert/distilbert-base-uncased
pipeline_tag: text-classification

Model Details

Text classification model for ambiguity in questions. Classifies questions as ambiguous or clear. Based on distilbert/distilbert-base-uncased.

Example:

"Did he do it?" {'label': 'AMBIG', 'score': 0.9029870629310608}

"Did Peter win the game?" {'label': 'CLEAR', 'score': 0.8900136351585388}

Out-of-Scope Use

The model was only trained to classify single questions. Other kinds of data are not tested.

Training Data

I manually labeled a small part of the inquisitiveqg dataset mixed with a private dataset to train the model to recognize ambiguity in questions. A satisfactory model with 85.5% accuracy was created.

Metrics

"eval_accuracy": 0.8551401869158879,

"eval_loss": 0.3658725619316101,