Update app.py
Browse files
app.py
CHANGED
@@ -343,7 +343,7 @@ import gradio as gr
|
|
343 |
#print(torch.cuda.is_available()) # Check if a GPU is available
|
344 |
#a = torch.Tensor([1]).cuda() # Should succeed without errors
|
345 |
#print(a)
|
346 |
-
|
347 |
# Define HuggingFace LLM
|
348 |
def create_llm():
|
349 |
model_name = "Qwen/Qwen2.5-7B-Instruct-1M"
|
@@ -447,6 +447,7 @@ def gradio_interface(visitor_name, visitor_mobile, doctor_name, department_name)
|
|
447 |
# Execute workflow
|
448 |
result = compiled_graph.invoke(state.dict())
|
449 |
return result["messages"][0]
|
|
|
450 |
|
451 |
iface = gr.Interface(
|
452 |
fn=gradio_interface,
|
@@ -460,7 +461,7 @@ iface = gr.Interface(
|
|
460 |
)
|
461 |
|
462 |
if __name__ == "__main__":
|
463 |
-
iface.launch(
|
464 |
|
465 |
|
466 |
|
|
|
343 |
#print(torch.cuda.is_available()) # Check if a GPU is available
|
344 |
#a = torch.Tensor([1]).cuda() # Should succeed without errors
|
345 |
#print(a)
|
346 |
+
'''
|
347 |
# Define HuggingFace LLM
|
348 |
def create_llm():
|
349 |
model_name = "Qwen/Qwen2.5-7B-Instruct-1M"
|
|
|
447 |
# Execute workflow
|
448 |
result = compiled_graph.invoke(state.dict())
|
449 |
return result["messages"][0]
|
450 |
+
'''
|
451 |
|
452 |
iface = gr.Interface(
|
453 |
fn=gradio_interface,
|
|
|
461 |
)
|
462 |
|
463 |
if __name__ == "__main__":
|
464 |
+
iface.launch()
|
465 |
|
466 |
|
467 |
|