Spaces:
Sleeping
Sleeping
Saiteja Solleti
commited on
Commit
·
041c22e
1
Parent(s):
585864a
adding token limit
Browse files- formatresultshelper.py +3 -0
formatresultshelper.py
CHANGED
@@ -1,5 +1,6 @@
|
|
1 |
import generationhelper
|
2 |
import json
|
|
|
3 |
|
4 |
def evaluate_response_with_prompt(templete, query, documents, answer, eval_model="llama-3.3-70b-specdec"):
|
5 |
|
@@ -15,6 +16,8 @@ def evaluate_response_with_prompt(templete, query, documents, answer, eval_model
|
|
15 |
|
16 |
prompt = templete.format(documents=formatted_documents, question=query, answer=formatted_answer)
|
17 |
|
|
|
|
|
18 |
# Call the LLM API (Llama 3.3-70B)
|
19 |
completion = generationhelper.groq_client.chat.completions.create(
|
20 |
model=eval_model,
|
|
|
1 |
import generationhelper
|
2 |
import json
|
3 |
+
from generationhelper import Enforce_token_limit
|
4 |
|
5 |
def evaluate_response_with_prompt(templete, query, documents, answer, eval_model="llama-3.3-70b-specdec"):
|
6 |
|
|
|
16 |
|
17 |
prompt = templete.format(documents=formatted_documents, question=query, answer=formatted_answer)
|
18 |
|
19 |
+
Enforce_token_limit(prompt)
|
20 |
+
|
21 |
# Call the LLM API (Llama 3.3-70B)
|
22 |
completion = generationhelper.groq_client.chat.completions.create(
|
23 |
model=eval_model,
|