freddyaboulton HF staff commited on
Commit
054e76a
·
1 Parent(s): 5cb0b21
Files changed (1) hide show
  1. app.py +3 -6
app.py CHANGED
@@ -7,6 +7,7 @@ import librosa
7
  import tempfile
8
  from twilio.rest import Client
9
  import os
 
10
  from transformers import Qwen2AudioForConditionalGeneration, AutoProcessor
11
 
12
  processor = AutoProcessor.from_pretrained("Qwen/Qwen2-Audio-7B-Instruct")
@@ -28,13 +29,9 @@ else:
28
  rtc_configuration = None
29
 
30
 
 
31
  def transcribe(audio: tuple[int, np.ndarray], transformers_convo: list[dict], gradio_convo: list[dict]):
32
- segment = AudioSegment(
33
- audio[1].tobytes(),
34
- frame_rate=audio[0],
35
- sample_width=audio[1].dtype.itemsize,
36
- channels=1,
37
- )
38
 
39
  with tempfile.NamedTemporaryFile(suffix=".mp3") as temp_audio:
40
  segment.export(temp_audio.name, format="mp3")
 
7
  import tempfile
8
  from twilio.rest import Client
9
  import os
10
+ import spaces
11
  from transformers import Qwen2AudioForConditionalGeneration, AutoProcessor
12
 
13
  processor = AutoProcessor.from_pretrained("Qwen/Qwen2-Audio-7B-Instruct")
 
29
  rtc_configuration = None
30
 
31
 
32
+ @spaces.GPU
33
  def transcribe(audio: tuple[int, np.ndarray], transformers_convo: list[dict], gradio_convo: list[dict]):
34
+ segment = AudioSegment(audio[1].tobytes(), frame_rate=audio[0], sample_width=audio[1].dtype.itemsize, channels=1)
 
 
 
 
 
35
 
36
  with tempfile.NamedTemporaryFile(suffix=".mp3") as temp_audio:
37
  segment.export(temp_audio.name, format="mp3")