File size: 2,926 Bytes
c3c7abe
 
 
 
 
 
 
 
291d559
999f24c
291d559
c3c7abe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
291d559
c3c7abe
 
999f24c
 
 
 
291d559
c3c7abe
999f24c
 
c3c7abe
291d559
999f24c
 
 
291d559
 
 
999f24c
 
c3c7abe
 
 
 
 
14044f3
c3c7abe
14044f3
999f24c
c3c7abe
 
 
 
999f24c
c3c7abe
 
 
291d559
999f24c
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
from typing import Dict
import json

from langchain_core.prompts import ChatPromptTemplate
from langchain_openai import ChatOpenAI
from langchain_core.runnables import RunnablePassthrough
from langchain_core.output_parsers import StrOutputParser
from backend.app.vectorstore import get_vector_db
from operator import itemgetter

MODEL = "gpt-3.5-turbo"

SYSTEM_ROLE_PROMPT = """
    You are a knowledgeable grading assistant that evaluates student answers based on provided context.
    You should determine if answers are correct and provide constructive feedback.
"""

USER_ROLE_PROMPT = """
    Grade the following student answer based on the provided context about {query}.
    
    Context: {context}
    
    Question: {problem}
    Student Answer: {answer}
    
    Evaluate if the answer is correct and provide brief feedback. Start with either "Correct" or "Incorrect" 
    followed by a brief explanation of why. Focus on the accuracy based on the context provided.
    
    Always begin your response with "Correct" or "Incorrect" and then provide a brief explanation of why.

    Your response should be direct and clear, for example:
    "Correct. The answer accurately explains [reason]" or 
    "Incorrect. While [partial understanding], the answer misses [key point]"
"""


class ProblemGradingPipeline:
    def __init__(self):
        self.chat_prompt = ChatPromptTemplate.from_messages(
            [("system", SYSTEM_ROLE_PROMPT), ("user", USER_ROLE_PROMPT)]
        )

        self.llm = ChatOpenAI(model=MODEL, temperature=0.3)
        self.retriever = get_vector_db().as_retriever(search_kwargs={"k": 2})

        self.rag_chain = (
            {
                # Use the query to retrieve documents from the vectorstore
                "context": itemgetter("query")
                | self.retriever
                | (lambda docs: "\n\n".join([doc.page_content for doc in docs])),
                # Pass through all other inputs directly
                "query": itemgetter("query"),
                "problem": itemgetter("problem"),
                "answer": itemgetter("answer"),
            }
            | self.chat_prompt
            | self.llm
            | StrOutputParser()
        )

    async def grade(self, query: str, problem: str, answer: str) -> str:
        """
        Asynchronously grade a student's answer to a problem using RAG for context-aware evaluation.

        Args:
            query (str): The topic/context to use for grading
            problem (str): The question being answered
            answer (str): The student's answer to evaluate

        Returns:
            str: Grading response indicating if the answer is correct and providing feedback
        """
        print(f"Grading problem: {problem} with answer: {answer} for query: {query}")
        return await self.rag_chain.ainvoke(
            {"query": query, "problem": problem, "answer": answer}
        )