import gradio as gr from langchain.chat_models import ChatOpenAI from langchain.schema import HumanMessage from utils import SyncStreamingLLMCallbackHandler from anyio.from_thread import start_blocking_portal from queue import Queue import prompt, tiktoken def num_token(string: str) -> int: """Returns the number of tokens in a text string.""" encoding = tiktoken.get_encoding('cl100k_base') num_tokens = len(encoding.encode(string)) return num_tokens def send_message(history, temp, top_p, fp, pp): q = Queue() job_done = object() def task(): llm = ChatOpenAI( model_name="gpt-3.5-turbo", streaming=True, temperature=temp, model_kwargs = { 'top_p':top_p, 'frequency_penalty':fp, 'presence_penalty':pp } ) query_llm = ChatOpenAI( model_name="gpt-3.5-turbo", temperature = 0 ) query_template = prompt.search_query_gradio(history) search_query = query_llm([HumanMessage(content=query_template)]).content print('query: ', search_query) answer_template = prompt.question_answer_gradio(history, search_query) print('final temp: \n\n', answer_template, f'\nprompt total: {num_token(answer_template)}') llm([HumanMessage(content=answer_template)], callbacks=[SyncStreamingLLMCallbackHandler(q)]) q.put(job_done) return with start_blocking_portal() as portal: portal.start_task_soon(task) content = "" while True: next_token = q.get(True, timeout=10) if next_token is job_done: return history content += next_token latest = history[-1] latest[1] = content yield history def append_user_message(message, history): "Add user message to chain/history and clear input field" print(history) history.append([message, None]) return "", history css = """ .contain { display: flex; flex-direction: column; height: 100%;} #chatbot { flex-grow: 1;} #chatbot .user {text-align: right;} .contain { display: flex; flex-direction: column; height: 100%;} #query_ref { flex-grow: 1; } """ with gr.Blocks(css=css, theme=gr.themes.Default(primary_hue="orange")) as demo: with gr.Row(elem_id="whole_page"): with gr.Column(elem_id="sidebar", scale=0.2, min_width=0): # TODO temperature? gr.Markdown('## City of Lake Elsinore Proposal Demo', show_label=False) with gr.Row(elem_id="logo_section1"): with gr.Column(elem_id="logo_col1", scale=0.5, min_width=0): gr.HTML("""