Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -11,8 +11,6 @@ if DEVICE == "auto":
|
|
11 |
DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
|
12 |
print(f"[SYSTEM] | Using {DEVICE} type compute device.")
|
13 |
|
14 |
-
torch.set_num_threads(4)
|
15 |
-
|
16 |
# Variables
|
17 |
CHAR_LIMIT = 2000
|
18 |
DEFAULT_INPUT = ""
|
@@ -80,18 +78,12 @@ def trim_silence(audio, threshold=0.001):
|
|
80 |
def generate(text=DEFAULT_INPUT, voice=DEFAULT_VOICE, speed=1):
|
81 |
text = text.strip()[:CHAR_LIMIT] + "."
|
82 |
pipeline = PIPELINES[voice[0]]
|
83 |
-
pack =
|
84 |
for _, ps, _ in pipeline(text, voice, speed):
|
85 |
-
|
86 |
-
|
87 |
-
|
88 |
-
|
89 |
-
except Exception as e:
|
90 |
-
print("Error during model inference:", e)
|
91 |
-
return None
|
92 |
-
audio_np = audio.numpy()
|
93 |
-
trimmed_audio = trim_silence(audio_np)
|
94 |
-
return (24000, trimmed_audio)
|
95 |
|
96 |
def cloud():
|
97 |
print("[CLOUD] | Space maintained.")
|
|
|
11 |
DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
|
12 |
print(f"[SYSTEM] | Using {DEVICE} type compute device.")
|
13 |
|
|
|
|
|
14 |
# Variables
|
15 |
CHAR_LIMIT = 2000
|
16 |
DEFAULT_INPUT = ""
|
|
|
78 |
def generate(text=DEFAULT_INPUT, voice=DEFAULT_VOICE, speed=1):
|
79 |
text = text.strip()[:CHAR_LIMIT] + "."
|
80 |
pipeline = PIPELINES[voice[0]]
|
81 |
+
pack = pipeline.load_voice(voice)
|
82 |
for _, ps, _ in pipeline(text, voice, speed):
|
83 |
+
ref_s = pack[len(ps) - 1]
|
84 |
+
audio = MODEL(ps, ref_s, speed)
|
85 |
+
return (24000, trim_silence(audio.numpy()))
|
86 |
+
return None
|
|
|
|
|
|
|
|
|
|
|
|
|
87 |
|
88 |
def cloud():
|
89 |
print("[CLOUD] | Space maintained.")
|