Spaces:
Sleeping
Sleeping
Update app.py
Browse filesfixed the issue
app.py
CHANGED
@@ -125,7 +125,7 @@ def respond(
|
|
125 |
|
126 |
elif "sing" in decision_response.lower():
|
127 |
llama_responded = 1
|
128 |
-
responded_answer = "SING"
|
129 |
return "Here is the song you might like!"
|
130 |
|
131 |
else:
|
@@ -156,7 +156,6 @@ def respond(
|
|
156 |
responded_answer = response
|
157 |
return response
|
158 |
|
159 |
-
|
160 |
def update_response_display():
|
161 |
while not llama_responded:
|
162 |
time.sleep(1)
|
@@ -164,13 +163,13 @@ def update_response_display():
|
|
164 |
def tts_part():
|
165 |
global llama_responded
|
166 |
global responded_answer
|
167 |
-
|
|
|
168 |
client = Client("ironserengety/MusicRetriever")
|
169 |
result = client.predict(
|
170 |
-
message=
|
171 |
api_name="/respond"
|
172 |
)
|
173 |
-
print("singing")
|
174 |
llama_responded = 0
|
175 |
responded_answer = ""
|
176 |
|
@@ -183,10 +182,10 @@ def tts_part():
|
|
183 |
audio=file('siri.wav'),
|
184 |
api_name="/predict"
|
185 |
)
|
186 |
-
print("speech")
|
187 |
llama_responded = 0
|
188 |
responded_answer = ""
|
189 |
return result
|
|
|
190 |
|
191 |
def create_interface():
|
192 |
with gr.Blocks() as demo:
|
|
|
125 |
|
126 |
elif "sing" in decision_response.lower():
|
127 |
llama_responded = 1
|
128 |
+
responded_answer = "SING " + message
|
129 |
return "Here is the song you might like!"
|
130 |
|
131 |
else:
|
|
|
156 |
responded_answer = response
|
157 |
return response
|
158 |
|
|
|
159 |
def update_response_display():
|
160 |
while not llama_responded:
|
161 |
time.sleep(1)
|
|
|
163 |
def tts_part():
|
164 |
global llama_responded
|
165 |
global responded_answer
|
166 |
+
result = ""
|
167 |
+
if "SING" in responded_answer:
|
168 |
client = Client("ironserengety/MusicRetriever")
|
169 |
result = client.predict(
|
170 |
+
message= responded_answer.lower(),
|
171 |
api_name="/respond"
|
172 |
)
|
|
|
173 |
llama_responded = 0
|
174 |
responded_answer = ""
|
175 |
|
|
|
182 |
audio=file('siri.wav'),
|
183 |
api_name="/predict"
|
184 |
)
|
|
|
185 |
llama_responded = 0
|
186 |
responded_answer = ""
|
187 |
return result
|
188 |
+
|
189 |
|
190 |
def create_interface():
|
191 |
with gr.Blocks() as demo:
|