huedaya commited on
Commit
e09f094
·
1 Parent(s): 7b54856
Files changed (1) hide show
  1. app.py +5 -4
app.py CHANGED
@@ -7,7 +7,7 @@ import requests
7
  import time
8
  from transformers import pipeline
9
 
10
- model = whisper.load_model("small")
11
  pipe = pipeline(
12
  "automatic-speech-recognition",
13
  model="openai/whisper-small",
@@ -41,7 +41,7 @@ def runApi():
41
 
42
  audio = "audio.mp3"
43
 
44
- audio = whisper.load_audio(audio)
45
  # audio = whisper.pad_or_trim(audio)
46
 
47
  # mel = whisper.log_mel_spectrogram(audio).to(model.device)
@@ -51,14 +51,15 @@ def runApi():
51
  # options = whisper.DecodingOptions(fp16 = False)
52
  # result = whisper.decode(model, mel, options)
53
 
54
- result = pipe(audio.copy())["text"]
 
55
 
56
  end_time = time.time()
57
  total_time = end_time - start_time
58
 
59
  return jsonify({
60
  "audio_url": audio_url,
61
- "result": result,
62
  "exec_time_sec": total_time
63
  })
64
 
 
7
  import time
8
  from transformers import pipeline
9
 
10
+ model = whisper.load_model("small.en")
11
  pipe = pipeline(
12
  "automatic-speech-recognition",
13
  model="openai/whisper-small",
 
41
 
42
  audio = "audio.mp3"
43
 
44
+ # audio = whisper.load_audio(audio)
45
  # audio = whisper.pad_or_trim(audio)
46
 
47
  # mel = whisper.log_mel_spectrogram(audio).to(model.device)
 
51
  # options = whisper.DecodingOptions(fp16 = False)
52
  # result = whisper.decode(model, mel, options)
53
 
54
+ # result = pipe(audio.copy())["text"]
55
+ result = model.transcribe(audio)
56
 
57
  end_time = time.time()
58
  total_time = end_time - start_time
59
 
60
  return jsonify({
61
  "audio_url": audio_url,
62
+ "result": result['text'],
63
  "exec_time_sec": total_time
64
  })
65