Manjot Singh commited on
Commit
bae6026
·
1 Parent(s): 4a17bd1

changed compute type

Browse files
Files changed (1) hide show
  1. audio_processing.py +4 -2
audio_processing.py CHANGED
@@ -12,7 +12,7 @@ from difflib import SequenceMatcher
12
  hf_token = os.getenv("HF_TOKEN")
13
 
14
  CHUNK_LENGTH=10
15
- OVERLAP=2
16
  import whisperx
17
  import torch
18
  import numpy as np
@@ -21,6 +21,8 @@ logger = logging.getLogger(__name__)
21
  import spaces
22
 
23
 
 
 
24
  def preprocess_audio(audio, chunk_size=CHUNK_LENGTH*16000, overlap=OVERLAP*16000): # 2 seconds overlap
25
  chunks = []
26
  for i in range(0, len(audio), chunk_size - overlap):
@@ -37,7 +39,7 @@ def process_audio(audio_file, translate=False, model_size="small"):
37
  try:
38
  device = "cuda" if torch.cuda.is_available() else "cpu"
39
  print(f"Using device: {device}")
40
- compute_type = "float32"
41
  audio = whisperx.load_audio(audio_file)
42
  model = whisperx.load_model(model_size, device, compute_type=compute_type)
43
 
 
12
  hf_token = os.getenv("HF_TOKEN")
13
 
14
  CHUNK_LENGTH=10
15
+ OVERLAP=0
16
  import whisperx
17
  import torch
18
  import numpy as np
 
21
  import spaces
22
 
23
 
24
+
25
+
26
  def preprocess_audio(audio, chunk_size=CHUNK_LENGTH*16000, overlap=OVERLAP*16000): # 2 seconds overlap
27
  chunks = []
28
  for i in range(0, len(audio), chunk_size - overlap):
 
39
  try:
40
  device = "cuda" if torch.cuda.is_available() else "cpu"
41
  print(f"Using device: {device}")
42
+ compute_type = torch.bfloat16 if torch.cuda.is_available() else "float32"
43
  audio = whisperx.load_audio(audio_file)
44
  model = whisperx.load_model(model_size, device, compute_type=compute_type)
45