Spaces:
Sleeping
Sleeping
import gradio as gr | |
import requests | |
import json | |
def query_ollama(prompt): | |
url = "http://localhost:11434/api/generate" | |
payload = { | |
"model": "hf.co/ibrahimBlyc/LA_Llama:latest", | |
"prompt": prompt | |
} | |
headers = {"Content-Type": "application/json"} | |
response = requests.post(url, json=payload, headers=headers, stream=True) | |
if response.status_code == 200: | |
full_response = "" | |
for line in response.iter_lines(): | |
if line: | |
data = json.loads(line.decode("utf-8")) | |
if "response" in data: | |
full_response += data["response"] | |
if data.get("done", False): | |
break | |
return full_response | |
else: | |
return f"Erreur HTTP : {response.status_code}" | |
interface = gr.Interface( | |
fn=query_ollama, | |
inputs=gr.Textbox(label="Posez votre question"), | |
outputs=gr.Textbox(label="Réponse"), | |
title="Chat avec Ollama", | |
description="Une interface simple pour discuter avec le modèle hébergé sur Ollama." | |
) | |
if __name__ == "__main__": | |
interface.launch() | |