Update function.py
Browse files- function.py +4 -4
function.py
CHANGED
@@ -8,12 +8,12 @@ def GetLLMResponse(selected_topic_level, selected_topic,num_quizzes, model):
|
|
8 |
|
9 |
if model == "Open AI":
|
10 |
llm = OpenAI(temperature=0.7, openai_api_key=st.secrets["OPENAI_API_KEY"])
|
11 |
-
questions = llm(question_prompt)
|
12 |
# return questions
|
13 |
|
14 |
elif model == "Gemini":
|
15 |
llm = ChatGoogleGenerativeAI(model="gemini-pro", google_api_key=st.secrets["GOOGLE_API_KEY"])
|
16 |
-
questions = llm.invoke(question_prompt)
|
17 |
# return questions.content
|
18 |
|
19 |
|
@@ -22,12 +22,12 @@ def GetLLMResponse(selected_topic_level, selected_topic,num_quizzes, model):
|
|
22 |
|
23 |
if model == "Open AI":
|
24 |
llm = OpenAI(temperature=0.7, openai_api_key=st.secrets["OPENAI_API_KEY"])
|
25 |
-
answers = llm(answer_prompt)
|
26 |
# return questions
|
27 |
|
28 |
elif model == "Gemini":
|
29 |
llm = ChatGoogleGenerativeAI(model="gemini-pro", google_api_key=st.secrets["GOOGLE_API_KEY"])
|
30 |
-
answers = llm.invoke(answer_prompt)
|
31 |
# return questions.content
|
32 |
|
33 |
|
|
|
8 |
|
9 |
if model == "Open AI":
|
10 |
llm = OpenAI(temperature=0.7, openai_api_key=st.secrets["OPENAI_API_KEY"])
|
11 |
+
questions = [llm(question_prompt)]
|
12 |
# return questions
|
13 |
|
14 |
elif model == "Gemini":
|
15 |
llm = ChatGoogleGenerativeAI(model="gemini-pro", google_api_key=st.secrets["GOOGLE_API_KEY"])
|
16 |
+
questions = [llm.invoke(question_prompt)]
|
17 |
# return questions.content
|
18 |
|
19 |
|
|
|
22 |
|
23 |
if model == "Open AI":
|
24 |
llm = OpenAI(temperature=0.7, openai_api_key=st.secrets["OPENAI_API_KEY"])
|
25 |
+
answers = [llm(answer_prompt)]
|
26 |
# return questions
|
27 |
|
28 |
elif model == "Gemini":
|
29 |
llm = ChatGoogleGenerativeAI(model="gemini-pro", google_api_key=st.secrets["GOOGLE_API_KEY"])
|
30 |
+
answers = [llm.invoke(answer_prompt)]
|
31 |
# return questions.content
|
32 |
|
33 |
|