File size: 1,482 Bytes
2c3ec72
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
---
tags:
- text-generation-inference
- whisper
- audio
base_model:
- openai/whisper-medium
---


# Whisper medium with Key-Value-Cache enabled in ONNX fp16 format
- Model creator: [Open AI](https://huggingface.co/openai)
- Original model: [Whisper Medium](https://huggingface.co/openai/whisper-medium)

<!-- description start -->
## Description

This repo contains the ONNX files for the ONNX conversion of Whisper Medium done by Esperanto Technologies.
The model is in the fp16 format and has the KVC enabled.

<!-- description end -->

## How to download ONNX model and weight files

The easiest way to obtain the model is to clone this whole repo.
Alternatively you can download the files is using the `huggingface-hub` Python library.

```shell
pip3 install huggingface-hub>=0.17.1
```

Then you can download any individual model file to the current directory, at high speed, with a command like this:

```shell
huggingface-cli download Esperanto/whisper-medium-kvc-fp16-onnx --local-dir whisper-medium-kvc-fp16-onnx --local-dir-use-symlinks False
```

For more documentation on downloading with `huggingface-cli`, please see: [HF -> Hub Python Library -> Download files -> Download from the CLI](https://huggingface.co/docs/huggingface_hub/guides/download#download-from-the-cli).

## How to run from Python code using ONNXRuntime

This model can easily be ran in a CPU using [ONNXRuntime](https://onnxruntime.ai/).

Scripts about how to run these models will be provided soon.