Spaces:
Running
Running
from openai import OpenAI | |
from tenacity import ( | |
retry, | |
stop_after_attempt, | |
wait_random_exponential, | |
) | |
PQA_QUESTION2QUESTION_PROMPT = """\ | |
Given a question, decompose it into multiple atomic questions. | |
### Instructions: | |
- The questions MUST be atomic, i.e., they MUST be answerable by only a single piece of information. | |
- The questions MUST be standalone, i.e., they MUST NOT reference any other question or the given question. | |
- The questions can be both open-ended or yes/no questions. | |
- The questions should be decomposed only from the main question. | |
- Each question should be on a new line and start with `**** `. | |
### Input: | |
{text} | |
### Output: | |
""" | |
qa_client = OpenAI( | |
base_url="http://130.85.37.21:4774/v1", | |
api_key="EMPTY" | |
) | |
def completion_with_backoff(client, **kwargs): | |
return client.chat.completions.create(**kwargs) | |
def generation_to_questions(generated_text, header, numbered=True): | |
try: | |
lines = generated_text.split("\n") | |
lines = [line.strip() for line in lines] | |
lines = [line for line in lines if line.startswith(header)] | |
lines = [line.replace(header, "").strip() for line in lines] | |
except: | |
lines = [] | |
print("Error in processing generated text") | |
return lines | |
def get_questions(claim): | |
prompt = PQA_QUESTION2QUESTION_PROMPT.format(text=claim) | |
print(f"Question Generation Prompt: {prompt}") | |
response = completion_with_backoff( | |
client=qa_client, | |
model="dipta007/Llama-3.1-8B-Instruct-finetuned-pqa", | |
messages=[ | |
{"role": "user", "content": prompt}, | |
], | |
max_tokens=2048, | |
top_p=1.0, | |
temperature=0.0, | |
) | |
print(f"Questions: {response.choices[0].message}") | |
generation = response.choices[0].message.content | |
questions = generation_to_questions(generation, "****") | |
questions = [f"- {q}" for q in questions] | |
return "\n".join(questions) |