Saiteja Solleti commited on
Commit
041c22e
·
1 Parent(s): 585864a

adding token limit

Browse files
Files changed (1) hide show
  1. formatresultshelper.py +3 -0
formatresultshelper.py CHANGED
@@ -1,5 +1,6 @@
1
  import generationhelper
2
  import json
 
3
 
4
  def evaluate_response_with_prompt(templete, query, documents, answer, eval_model="llama-3.3-70b-specdec"):
5
 
@@ -15,6 +16,8 @@ def evaluate_response_with_prompt(templete, query, documents, answer, eval_model
15
 
16
  prompt = templete.format(documents=formatted_documents, question=query, answer=formatted_answer)
17
 
 
 
18
  # Call the LLM API (Llama 3.3-70B)
19
  completion = generationhelper.groq_client.chat.completions.create(
20
  model=eval_model,
 
1
  import generationhelper
2
  import json
3
+ from generationhelper import Enforce_token_limit
4
 
5
  def evaluate_response_with_prompt(templete, query, documents, answer, eval_model="llama-3.3-70b-specdec"):
6
 
 
16
 
17
  prompt = templete.format(documents=formatted_documents, question=query, answer=formatted_answer)
18
 
19
+ Enforce_token_limit(prompt)
20
+
21
  # Call the LLM API (Llama 3.3-70B)
22
  completion = generationhelper.groq_client.chat.completions.create(
23
  model=eval_model,