Spaces:
Runtime error
Runtime error
from threading import Thread | |
import gradio as gr | |
import inspect | |
from gradio import routes | |
from typing import List, Type | |
import requests, os, re, asyncio, queue | |
import math | |
import time | |
import datetime | |
import requests, json | |
from huggingface_hub import hf_hub_download | |
from llama_cpp import Llama | |
loop = asyncio.get_event_loop() | |
# Monkey patch | |
def get_types(cls_set: List[Type], component: str): | |
docset = [] | |
types = [] | |
if component == "input": | |
for cls in cls_set: | |
doc = inspect.getdoc(cls) | |
doc_lines = doc.split("\n") | |
docset.append(doc_lines[1].split(":")[-1]) | |
types.append(doc_lines[1].split(")")[0].split("(")[-1]) | |
else: | |
for cls in cls_set: | |
doc = inspect.getdoc(cls) | |
doc_lines = doc.split("\n") | |
docset.append(doc_lines[-1].split(":")[-1]) | |
types.append(doc_lines[-1].split(")")[0].split("(")[-1]) | |
return docset, types | |
routes.get_types = get_types | |
hf_hub_download(repo_id='StarFox7/Llama-2-ko-7B-chat-gguf', filename='Llama-2-ko-7B-chat-gguf-q4_0.bin', local_dir='./') | |
llm = Llama(model_path = 'Llama-2-ko-7B-chat-gguf-q4_0.bin', | |
n_ctx=512, | |
) | |
# App code | |
def chat(x): | |
#prom = f"λ€μμ Aμ Bμ μν κ·Ήμ΄μΌ. λλ BμΌ. Aμ λννκ³ μμ΄. μΉκ΅¬μκ² μΉκ·Όνκ³ κ°κ²°νκ² μ λλ΅ν΄μ€.\n\n### A:\n{x}\n\n### B:\n" | |
#output = llm(prom, max_tokens=20, stop=["###"], echo=True) | |
#return output['choices'][0]['text'][len(prom):-1] | |
return "AI μλ΅μ λλ€." | |
with gr.Blocks() as demo: | |
count = 0 | |
aa = gr.Interface( | |
fn=chat, | |
inputs=["text"], | |
outputs="text", | |
description="call", | |
) | |
demo.queue(max_size=32).launch(enable_queue=True) |