huedaya commited on
Commit
fab4dca
·
1 Parent(s): 23a9064
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -44,27 +44,27 @@ def runApi():
44
  audio = "audio.mp3"
45
 
46
  audioOri = whisper.load_audio(audio)
47
- audio = whisper.pad_or_trim(audioOri)
48
 
49
- mel = whisper.log_mel_spectrogram(audio).to(model.device)
50
 
51
- _, probs = model.detect_language(mel)
52
 
53
- options = whisper.DecodingOptions(fp16 = False)
54
- result = whisper.decode(model, mel, options)
55
 
56
  # test 2
57
  # ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation")
58
  # sample = ds[0]["audio"]
59
  prediction = pipe(audioOri)["text"]
60
- print(prediction)
61
 
62
  end_time = time.time()
63
  total_time = end_time - start_time
64
 
65
  return jsonify({
66
  "audio_url": audio_url,
67
- "result": result.text,
68
  "exec_time_sec": total_time
69
  })
70
 
 
44
  audio = "audio.mp3"
45
 
46
  audioOri = whisper.load_audio(audio)
47
+ # audio = whisper.pad_or_trim(audioOri)
48
 
49
+ # mel = whisper.log_mel_spectrogram(audio).to(model.device)
50
 
51
+ # _, probs = model.detect_language(mel)
52
 
53
+ # options = whisper.DecodingOptions(fp16 = False)
54
+ # result = whisper.decode(model, mel, options)
55
 
56
  # test 2
57
  # ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation")
58
  # sample = ds[0]["audio"]
59
  prediction = pipe(audioOri)["text"]
60
+
61
 
62
  end_time = time.time()
63
  total_time = end_time - start_time
64
 
65
  return jsonify({
66
  "audio_url": audio_url,
67
+ "result": prediction,
68
  "exec_time_sec": total_time
69
  })
70