Spaces:
Sleeping
Sleeping
max_tokens=4000,
Browse files
app.py
CHANGED
@@ -654,7 +654,7 @@ def generate_paragraph_evaluate(model, sys_content, paragraph, user_generate_par
|
|
654 |
prompt=f"{sys_content}\n{user_content}" if "gemini" in model.lower() else None,
|
655 |
messages=messages,
|
656 |
model=model,
|
657 |
-
max_tokens=
|
658 |
response_format={"type": "json_object"}
|
659 |
)
|
660 |
|
@@ -712,7 +712,7 @@ def generate_correct_grammatical_spelling_errors(model, sys_content, eng_level,
|
|
712 |
prompt=f"{sys_content}\n{user_content}" if "gemini" in model.lower() else None,
|
713 |
messages=messages,
|
714 |
model=model,
|
715 |
-
max_tokens=
|
716 |
response_format={"type": "json_object"}
|
717 |
)
|
718 |
|
@@ -2077,7 +2077,7 @@ def check_chinese_essay_feedback(feedback_check_prompt, chinese_essay_from_stude
|
|
2077 |
request_payload = {
|
2078 |
"model": "gpt-4o",
|
2079 |
"messages": messages,
|
2080 |
-
"max_tokens":
|
2081 |
}
|
2082 |
|
2083 |
try:
|
|
|
654 |
prompt=f"{sys_content}\n{user_content}" if "gemini" in model.lower() else None,
|
655 |
messages=messages,
|
656 |
model=model,
|
657 |
+
max_tokens=4000,
|
658 |
response_format={"type": "json_object"}
|
659 |
)
|
660 |
|
|
|
712 |
prompt=f"{sys_content}\n{user_content}" if "gemini" in model.lower() else None,
|
713 |
messages=messages,
|
714 |
model=model,
|
715 |
+
max_tokens=4000,
|
716 |
response_format={"type": "json_object"}
|
717 |
)
|
718 |
|
|
|
2077 |
request_payload = {
|
2078 |
"model": "gpt-4o",
|
2079 |
"messages": messages,
|
2080 |
+
"max_tokens": 4000,
|
2081 |
}
|
2082 |
|
2083 |
try:
|