alicekyting
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -7,13 +7,10 @@ tags: []
|
|
7 |
|
8 |
This model is a quantized version of openai/whisper-large-v3, optimized for more efficient use while maintaining performance.
|
9 |
|
10 |
-
|
11 |
## Model Details
|
12 |
|
13 |
### Model Description
|
14 |
|
15 |
-
<!-- Provide a longer summary of what this model is. -->
|
16 |
-
|
17 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
18 |
|
19 |
- **Developed by:** alicekyting (based on OpenAI's Whisper model)
|
@@ -25,6 +22,9 @@ This is the model card of a 🤗 transformers model that has been pushed on the
|
|
25 |
This model can be used for automatic speech recognition (ASR) tasks, including transcription and translation.
|
26 |
It's particularly useful in scenarios where computational efficiency is important, as it has been quantized to 4-bit precision.
|
27 |
|
|
|
|
|
|
|
28 |
## Bias, Risks, and Limitations
|
29 |
|
30 |
This model inherits any biases, risks, and limitations present in the original openai/whisper-large-v3 model.
|
@@ -43,18 +43,22 @@ Use the following code to load and use the model:
|
|
43 |
from transformers import AutoModelForSpeechSeq2Seq, AutoProcessor
|
44 |
import torch
|
45 |
|
|
|
46 |
model = AutoModelForSpeechSeq2Seq.from_pretrained(
|
47 |
-
"alicekyting/whisper-large-v3-4bit
|
48 |
device_map="auto",
|
49 |
torch_dtype=torch.float16,
|
|
|
50 |
)
|
51 |
|
52 |
-
|
|
|
53 |
|
54 |
pipe = pipeline(
|
55 |
"automatic-speech-recognition",
|
56 |
model=model,
|
57 |
tokenizer=processor.tokenizer,
|
58 |
feature_extractor=processor.feature_extractor,
|
59 |
-
torch_dtype=
|
|
|
60 |
)
|
|
|
7 |
|
8 |
This model is a quantized version of openai/whisper-large-v3, optimized for more efficient use while maintaining performance.
|
9 |
|
|
|
10 |
## Model Details
|
11 |
|
12 |
### Model Description
|
13 |
|
|
|
|
|
14 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
15 |
|
16 |
- **Developed by:** alicekyting (based on OpenAI's Whisper model)
|
|
|
22 |
This model can be used for automatic speech recognition (ASR) tasks, including transcription and translation.
|
23 |
It's particularly useful in scenarios where computational efficiency is important, as it has been quantized to 4-bit precision.
|
24 |
|
25 |
+
## Hardware Requirements
|
26 |
+
It is recommended to use this model on a device with a compatible GPU.
|
27 |
+
|
28 |
## Bias, Risks, and Limitations
|
29 |
|
30 |
This model inherits any biases, risks, and limitations present in the original openai/whisper-large-v3 model.
|
|
|
43 |
from transformers import AutoModelForSpeechSeq2Seq, AutoProcessor
|
44 |
import torch
|
45 |
|
46 |
+
# Load the model
|
47 |
model = AutoModelForSpeechSeq2Seq.from_pretrained(
|
48 |
+
"alicekyting/whisper-large-v3-4bit",
|
49 |
device_map="auto",
|
50 |
torch_dtype=torch.float16,
|
51 |
+
use_safetensors=True,
|
52 |
)
|
53 |
|
54 |
+
# Load the processor
|
55 |
+
processor = AutoProcessor.from_pretrained("alicekyting/whisper-large-v3-4bit")
|
56 |
|
57 |
pipe = pipeline(
|
58 |
"automatic-speech-recognition",
|
59 |
model=model,
|
60 |
tokenizer=processor.tokenizer,
|
61 |
feature_extractor=processor.feature_extractor,
|
62 |
+
torch_dtype=torch.float16,
|
63 |
+
device_map="auto"
|
64 |
)
|