File size: 1,482 Bytes
2c3ec72 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 |
---
tags:
- text-generation-inference
- whisper
- audio
base_model:
- openai/whisper-medium
---
# Whisper medium with Key-Value-Cache enabled in ONNX fp16 format
- Model creator: [Open AI](https://huggingface.co/openai)
- Original model: [Whisper Medium](https://huggingface.co/openai/whisper-medium)
<!-- description start -->
## Description
This repo contains the ONNX files for the ONNX conversion of Whisper Medium done by Esperanto Technologies.
The model is in the fp16 format and has the KVC enabled.
<!-- description end -->
## How to download ONNX model and weight files
The easiest way to obtain the model is to clone this whole repo.
Alternatively you can download the files is using the `huggingface-hub` Python library.
```shell
pip3 install huggingface-hub>=0.17.1
```
Then you can download any individual model file to the current directory, at high speed, with a command like this:
```shell
huggingface-cli download Esperanto/whisper-medium-kvc-fp16-onnx --local-dir whisper-medium-kvc-fp16-onnx --local-dir-use-symlinks False
```
For more documentation on downloading with `huggingface-cli`, please see: [HF -> Hub Python Library -> Download files -> Download from the CLI](https://huggingface.co/docs/huggingface_hub/guides/download#download-from-the-cli).
## How to run from Python code using ONNXRuntime
This model can easily be ran in a CPU using [ONNXRuntime](https://onnxruntime.ai/).
Scripts about how to run these models will be provided soon. |