Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Update app_gradio.py
Browse files- app_gradio.py +0 -3
app_gradio.py
CHANGED
@@ -539,15 +539,12 @@ def compileinfo(query, atom_qns, atom_qn_ans, atom_qn_strs):
|
|
539 |
def deep_research(question, top_k, ec):
|
540 |
|
541 |
full_answer = '## ' + question
|
542 |
-
yield None, None
|
543 |
|
544 |
gen_client = openai_llm(temperature=0,model_name='gpt-4o-mini', openai_api_key = openai_key)
|
545 |
messages = [("system",df_atomic_prompt,),("human", question),]
|
546 |
rscope_text = gen_client.invoke(messages).content
|
547 |
|
548 |
full_answer = full_answer +' \n'+ rscope_text
|
549 |
-
rag_answer = {}
|
550 |
-
rag_answer['answer'] = full_answer
|
551 |
|
552 |
rscope_messages = [("system","""In the given text, what are the main atomic questions being asked? Please answer as a concise list.""",),("human", rscope_text),]
|
553 |
rscope_qns = gen_client.invoke(rscope_messages).content
|
|
|
539 |
def deep_research(question, top_k, ec):
|
540 |
|
541 |
full_answer = '## ' + question
|
|
|
542 |
|
543 |
gen_client = openai_llm(temperature=0,model_name='gpt-4o-mini', openai_api_key = openai_key)
|
544 |
messages = [("system",df_atomic_prompt,),("human", question),]
|
545 |
rscope_text = gen_client.invoke(messages).content
|
546 |
|
547 |
full_answer = full_answer +' \n'+ rscope_text
|
|
|
|
|
548 |
|
549 |
rscope_messages = [("system","""In the given text, what are the main atomic questions being asked? Please answer as a concise list.""",),("human", rscope_text),]
|
550 |
rscope_qns = gen_client.invoke(rscope_messages).content
|