Spaces:
Runtime error
Runtime error
import openai | |
from typing import List | |
def get_llm_response( | |
client: openai.OpenAI, | |
prompt: str, | |
system_prompt: str = "You are a helpful assistant.", | |
model: str = "gpt-4o-mini", | |
temperature: float = 0.7, | |
top_p: float = 0.95, | |
max_tokens: int = 500, | |
) -> str: | |
""" | |
Get response from OpenAI language model. | |
Args: | |
client (openai.OpenAI): OpenAI client | |
prompt (str): The user prompt/question to send to the model | |
system_prompt (str, optional): System prompt to set model behavior. | |
model (str, optional): OpenAI model to use. Defaults to "gpt-4o-mini". | |
temperature (float, optional): Controls randomness in responses. Defaults to 0.7. | |
top_p (float, optional): Controls diversity via nucleus sampling. Defaults to 0.95. | |
max_tokens (int, optional): Max tokens in model response. Defaults to 200. | |
Returns: | |
str: The model's response text | |
""" | |
messages = [ | |
{"role": "system", "content": system_prompt}, | |
{"role": "user", "content": prompt}, | |
] | |
response = client.chat.completions.create( | |
model=model, | |
messages=messages, | |
temperature=temperature, | |
top_p=top_p, | |
max_tokens=max_tokens, | |
) | |
return response.choices[0].message.content | |