alibicer commited on
Commit
faf0908
·
verified ·
1 Parent(s): 2852799

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -7
app.py CHANGED
@@ -16,7 +16,7 @@ client = OpenAI(api_key=OPENAI_API_KEY)
16
 
17
  def gpt_call(history, user_message,
18
  model="gpt-4o",
19
- max_tokens=512,
20
  temperature=0.7,
21
  top_p=0.95):
22
  """
@@ -41,17 +41,15 @@ def gpt_call(history, user_message,
41
  completion = client.chat.completions.create(
42
  model=model,
43
  messages=messages,
44
- max_tokens=max_tokens,
45
  temperature=temperature,
46
  top_p=top_p
47
  )
48
 
49
- # 5) Ensure LaTeX equations are properly wrapped in double $$ for rendering
50
- response_text = completion.choices[0].message.content
51
- formatted_response = response_text.replace("\\[", "$$").replace("\\]", "$$")
52
- formatted_response = formatted_response.replace("\\(", "$").replace("\\)", "$")
53
 
54
- return formatted_response
55
 
56
 
57
  def respond(user_message, history):
 
16
 
17
  def gpt_call(history, user_message,
18
  model="gpt-4o",
19
+ max_tokens=1500, # Increased from 512 to 1500 to prevent truncation
20
  temperature=0.7,
21
  top_p=0.95):
22
  """
 
41
  completion = client.chat.completions.create(
42
  model=model,
43
  messages=messages,
44
+ max_tokens=max_tokens, # Increased to allow longer responses
45
  temperature=temperature,
46
  top_p=top_p
47
  )
48
 
49
+ # 5) Ensure full response is returned without being cut off
50
+ full_response = "".join(choice.message.content for choice in completion.choices).strip()
 
 
51
 
52
+ return full_response
53
 
54
 
55
  def respond(user_message, history):