jhj0517 commited on
Commit
2abc8d7
·
1 Parent(s): db60455

fix spaces bug

Browse files
modules/translation/nllb_inference.py CHANGED
@@ -21,14 +21,14 @@ class NLLBInference(TranslationBase):
21
  self.available_target_langs = list(NLLB_AVAILABLE_LANGS.keys())
22
  self.pipeline = None
23
 
24
- @spaces.GPU
25
  def translate(self,
26
  text: str
27
  ):
28
  result = self.pipeline(text)
29
  return result[0]['translation_text']
30
 
31
- @spaces.GPU
32
  def update_model(self,
33
  model_size: str,
34
  src_lang: str,
 
21
  self.available_target_langs = list(NLLB_AVAILABLE_LANGS.keys())
22
  self.pipeline = None
23
 
24
+ @spaces.GPU(duration=120)
25
  def translate(self,
26
  text: str
27
  ):
28
  result = self.pipeline(text)
29
  return result[0]['translation_text']
30
 
31
+ @spaces.GPU(duration=120)
32
  def update_model(self,
33
  model_size: str,
34
  src_lang: str,
modules/translation/translation_base.py CHANGED
@@ -24,12 +24,14 @@ class TranslationBase(ABC):
24
  self.device = self.get_device()
25
 
26
  @abstractmethod
 
27
  def translate(self,
28
  text: str
29
  ):
30
  pass
31
 
32
  @abstractmethod
 
33
  def update_model(self,
34
  model_size: str,
35
  src_lang: str,
@@ -38,6 +40,7 @@ class TranslationBase(ABC):
38
  ):
39
  pass
40
 
 
41
  def translate_file(self,
42
  fileobjs: list,
43
  model_size: str,
@@ -128,7 +131,7 @@ class TranslationBase(ABC):
128
  self.remove_input_files([fileobj.name for fileobj in fileobjs])
129
 
130
  @staticmethod
131
- @spaces.GPU
132
  def get_device():
133
  if torch.cuda.is_available():
134
  return "cuda"
@@ -138,7 +141,7 @@ class TranslationBase(ABC):
138
  return "cpu"
139
 
140
  @staticmethod
141
- @spaces.GPU
142
  def release_cuda_memory():
143
  if torch.cuda.is_available():
144
  torch.cuda.empty_cache()
 
24
  self.device = self.get_device()
25
 
26
  @abstractmethod
27
+ @spaces.GPU(duration=120)
28
  def translate(self,
29
  text: str
30
  ):
31
  pass
32
 
33
  @abstractmethod
34
+ @spaces.GPU(duration=120)
35
  def update_model(self,
36
  model_size: str,
37
  src_lang: str,
 
40
  ):
41
  pass
42
 
43
+ @spaces.GPU(duration=120)
44
  def translate_file(self,
45
  fileobjs: list,
46
  model_size: str,
 
131
  self.remove_input_files([fileobj.name for fileobj in fileobjs])
132
 
133
  @staticmethod
134
+ @spaces.GPU(duration=120)
135
  def get_device():
136
  if torch.cuda.is_available():
137
  return "cuda"
 
141
  return "cpu"
142
 
143
  @staticmethod
144
+ @spaces.GPU(duration=120)
145
  def release_cuda_memory():
146
  if torch.cuda.is_available():
147
  torch.cuda.empty_cache()
modules/whisper/faster_whisper_inference.py CHANGED
@@ -31,7 +31,7 @@ class FasterWhisperInference(WhisperBase):
31
  self.available_models = self.model_paths.keys()
32
  self.available_compute_types = self.get_available_compute_type()
33
 
34
- @spaces.GPU
35
  def transcribe(self,
36
  audio: Union[str, BinaryIO, np.ndarray],
37
  progress: gr.Progress,
@@ -89,7 +89,7 @@ class FasterWhisperInference(WhisperBase):
89
  elapsed_time = time.time() - start_time
90
  return segments_result, elapsed_time
91
 
92
- @spaces.GPU
93
  def update_model(self,
94
  model_size: str,
95
  compute_type: str,
@@ -149,7 +149,7 @@ class FasterWhisperInference(WhisperBase):
149
  return ['int16', 'float32', 'int8', 'int8_float32']
150
 
151
  @staticmethod
152
- @spaces.GPU
153
  def get_device():
154
  if torch.cuda.is_available():
155
  return "cuda"
 
31
  self.available_models = self.model_paths.keys()
32
  self.available_compute_types = self.get_available_compute_type()
33
 
34
+ @spaces.GPU(duration=120)
35
  def transcribe(self,
36
  audio: Union[str, BinaryIO, np.ndarray],
37
  progress: gr.Progress,
 
89
  elapsed_time = time.time() - start_time
90
  return segments_result, elapsed_time
91
 
92
+ @spaces.GPU(duration=120)
93
  def update_model(self,
94
  model_size: str,
95
  compute_type: str,
 
149
  return ['int16', 'float32', 'int8', 'int8_float32']
150
 
151
  @staticmethod
152
+ @spaces.GPU(duration=120)
153
  def get_device():
154
  if torch.cuda.is_available():
155
  return "cuda"
modules/whisper/whisper_base.py CHANGED
@@ -42,7 +42,7 @@ class WhisperBase(ABC):
42
  self.vad = SileroVAD()
43
 
44
  @abstractmethod
45
- @spaces.GPU
46
  def transcribe(self,
47
  audio: Union[str, BinaryIO, np.ndarray],
48
  progress: gr.Progress,
@@ -51,7 +51,7 @@ class WhisperBase(ABC):
51
  pass
52
 
53
  @abstractmethod
54
- @spaces.GPU
55
  def update_model(self,
56
  model_size: str,
57
  compute_type: str,
@@ -59,7 +59,7 @@ class WhisperBase(ABC):
59
  ):
60
  pass
61
 
62
- @spaces.GPU
63
  def run(self,
64
  audio: Union[str, BinaryIO, np.ndarray],
65
  progress: gr.Progress,
@@ -125,7 +125,7 @@ class WhisperBase(ABC):
125
  elapsed_time += elapsed_time_diarization
126
  return result, elapsed_time
127
 
128
- @spaces.GPU
129
  def transcribe_file(self,
130
  files: list,
131
  file_format: str,
@@ -196,7 +196,7 @@ class WhisperBase(ABC):
196
  if not files:
197
  self.remove_input_files([file.name for file in files])
198
 
199
- @spaces.GPU
200
  def transcribe_mic(self,
201
  mic_audio: str,
202
  file_format: str,
@@ -249,7 +249,7 @@ class WhisperBase(ABC):
249
  self.release_cuda_memory()
250
  self.remove_input_files([mic_audio])
251
 
252
- @spaces.GPU
253
  def transcribe_youtube(self,
254
  youtube_link: str,
255
  file_format: str,
@@ -400,7 +400,7 @@ class WhisperBase(ABC):
400
  return time_str.strip()
401
 
402
  @staticmethod
403
- @spaces.GPU
404
  def get_device():
405
  if torch.cuda.is_available():
406
  return "cuda"
@@ -410,7 +410,7 @@ class WhisperBase(ABC):
410
  return "cpu"
411
 
412
  @staticmethod
413
- @spaces.GPU
414
  def release_cuda_memory():
415
  if torch.cuda.is_available():
416
  torch.cuda.empty_cache()
 
42
  self.vad = SileroVAD()
43
 
44
  @abstractmethod
45
+ @spaces.GPU(duration=120)
46
  def transcribe(self,
47
  audio: Union[str, BinaryIO, np.ndarray],
48
  progress: gr.Progress,
 
51
  pass
52
 
53
  @abstractmethod
54
+ @spaces.GPU(duration=120)
55
  def update_model(self,
56
  model_size: str,
57
  compute_type: str,
 
59
  ):
60
  pass
61
 
62
+ @spaces.GPU(duration=120)
63
  def run(self,
64
  audio: Union[str, BinaryIO, np.ndarray],
65
  progress: gr.Progress,
 
125
  elapsed_time += elapsed_time_diarization
126
  return result, elapsed_time
127
 
128
+ @spaces.GPU(duration=120)
129
  def transcribe_file(self,
130
  files: list,
131
  file_format: str,
 
196
  if not files:
197
  self.remove_input_files([file.name for file in files])
198
 
199
+ @spaces.GPU(duration=120)
200
  def transcribe_mic(self,
201
  mic_audio: str,
202
  file_format: str,
 
249
  self.release_cuda_memory()
250
  self.remove_input_files([mic_audio])
251
 
252
+ @spaces.GPU(duration=120)
253
  def transcribe_youtube(self,
254
  youtube_link: str,
255
  file_format: str,
 
400
  return time_str.strip()
401
 
402
  @staticmethod
403
+ @spaces.GPU(duration=120)
404
  def get_device():
405
  if torch.cuda.is_available():
406
  return "cuda"
 
410
  return "cpu"
411
 
412
  @staticmethod
413
+ @spaces.GPU(duration=120)
414
  def release_cuda_memory():
415
  if torch.cuda.is_available():
416
  torch.cuda.empty_cache()