harsh-deepchecks's picture
Update README.md
694ccbf
|
raw
history blame
929 Bytes
metadata
language:
  - en
tags:
  - fluency
license: apache-2.0

This model represents an ONNX-optimized version of the original parrot_fluency_model model. It has been specifically tailored for GPUs and may exhibit variations in performance when run on CPUs.

How to use

from transformers import AutoTokenizer
from optimum.onnxruntime import ORTModelForSequenceClassification
from optimum.pipelines import pipeline

# load tokenizer and model weights
tokenizer = AutoTokenizer.from_pretrained('Deepchecks/parrot_fluency_model_onnx')
model = ORTModelForSequenceClassification.from_pretrained('Deepchecks/parrot_fluency_model_onnx')

# prepare the pipeline and generate inferences
pip = pipeline(task='text-classification', model=onnx_model, tokenizer=tokenizer, device=device, accelerator="ort")
res = pip(user_inputs, batch_size=64, truncation="only_first")