ArturG9 commited on
Commit
cded29b
·
verified ·
1 Parent(s): 735a7a0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -33
README.md CHANGED
@@ -12,42 +12,41 @@ sdk_version: 1.38.0
12
  This project demonstrates a self Reflective corrective Retrieval Augmented Generation (CRAG) application built using LangGraph. The application leverages a Gemma2 9B LLM to provide informative and relevant responses to user queries. It employs a multi-agent approach, incorporating various components for enhanced performance and user experience.
13
 
14
  # Key Features
15
- Vector Store: Uses Chroma Vector Store to efficiently store and retrieve context from scraped webpages related to data science and programming.
16
- Prompt Guard: Ensures question safety by checking against predefined guidelines.
17
- LLM Graders: Evaluates question relevance, answer grounding, and helpfulness to maintain high-quality responses.
18
- Retrieval and Generation: Combines context retrieval from vector store and web search with LLM generation to provide comprehensive answers.
19
- Iterative Refinement: Rewrites questions and regenerates answers as needed to ensure accuracy and relevance.
20
- Customization: Offers flexibility in model selection, fine-tuning, and retrieval methods to tailor the application to specific requirements.
21
- Local Deployment: Can be deployed locally for enhanced user data privacy.
22
  ## Technical Specifications
23
- LLM: Gemma2 9B
24
- Vector Store: Chroma
25
- Embeddings: Alibaba-NLP/gte-base-en-v1.5
26
- Workflow: LangGraph
27
- Model API: ChatGroq
28
- Web Search: Wikipedia and Google SERP
29
  ## Workflow
30
- User Query: User inputs a question.
31
- Prompt Guard: Checks if the question is safe and appropriate.
32
- Context Retrieval: Searches the vector store for relevant documents.
33
- Document Relevance: Evaluates document relevance using LLM graders.
34
- Web Search: If necessary, conducts web searches on Wikipedia and Google SERP.
35
- Answer Generation: Generates a response using the retrieved documents and LLM.
36
- Answer Evaluation: Evaluates answer grounding and helpfulness using LLM graders.
37
- Refinement: If necessary, rewrites the question or regenerates the answer.
38
  ## Customization Options
39
- Model Selection: Choose different LLM models based on specific needs (e.g., larger models for more complex tasks).
40
- Fine-Tuning: Fine-tune the LLM to match specific styles or domains.
41
- Retrieval Methods: Explore alternative vector stores or retrieval techniques.
42
  ## Local Deployment
43
- To deploy the application locally, follow these steps:
44
 
45
- Set up environment: Install required dependencies (LangGraph, Chroma, LLM API, etc.).
46
- Prepare data: Scrape webpages and create the vector store.
47
- Configure workflow: Define the workflow and LLM graders.
48
- Run application: Execute the application to start processing user queries.
49
  ## Future Enhancements
50
- Knowledge Base Expansion: Continuously update the vector store with new data.
51
- Retrieval Optimization: Explore more efficient retrieval techniques.
52
- Multi-lingual Support: Enable the application to handle multiple languages.
53
- Integration with Other Applications: Integrate with other tools or platforms for broader use cases.
 
12
  This project demonstrates a self Reflective corrective Retrieval Augmented Generation (CRAG) application built using LangGraph. The application leverages a Gemma2 9B LLM to provide informative and relevant responses to user queries. It employs a multi-agent approach, incorporating various components for enhanced performance and user experience.
13
 
14
  # Key Features
15
+ * Vector Store: Uses Chroma Vector Store to efficiently store and retrieve context from scraped webpages related to data science and programming.
16
+ * Prompt Guard: Ensures question safety by checking against predefined guidelines.
17
+ * LLM Graders: Evaluates question relevance, answer grounding, and helpfulness to maintain high-quality responses.
18
+ * Retrieval and Generation: Combines context retrieval from vector store and web search with LLM generation to provide comprehensive answers.
19
+ * Iterative Refinement: Rewrites questions and regenerates answers as needed to ensure accuracy and relevance.
20
+ * Customization: Offers flexibility in model selection, fine-tuning, and retrieval methods to tailor the application to specific requirements.
21
+ * Local Deployment: Can be deployed locally for enhanced user data privacy.
22
  ## Technical Specifications
23
+ * LLM: Gemma2 9B
24
+ * Vector Store: Chroma
25
+ * Embeddings: Alibaba-NLP/gte-base-en-v1.5
26
+ * Workflow: LangGraph
27
+ * Model API: ChatGroq
28
+ * Web Search: Wikipedia and Google SERP
29
  ## Workflow
30
+ * User Query: User inputs a question.
31
+ * Prompt Guard: Checks if the question is safe and appropriate.
32
+ * Context Retrieval: Searches the vector store for relevant documents.
33
+ * Document Relevance: Evaluates document relevance using LLM graders.
34
+ * Web Search: If necessary, conducts web searches on Wikipedia and Google SERP.
35
+ * Answer Generation: Generates a response using the retrieved documents and LLM.
36
+ * Answer Evaluation: Evaluates answer grounding and helpfulness using LLM graders.
37
+ * Refinement: If necessary, rewrites the question or regenerates the answer.
38
  ## Customization Options
39
+ * Model Selection: Choose different LLM models based on specific needs (e.g., larger models for more complex tasks).
40
+ * Fine-Tuning: Fine-tune the LLM to match specific styles or domains.
41
+ * Retrieval Methods: Explore alternative vector stores or retrieval techniques.
42
  ## Local Deployment
43
+ * To deploy the application locally, follow these steps:
44
 
45
+ * Set up environment: Install required dependencies (LangGraph, Chroma, LLM API, etc.).
46
+ * Prepare data: Scrape webpages and create the vector store.
47
+ * Configure workflow: Define the workflow and LLM graders.
48
+ * Run application: Execute the application to start processing user queries.
49
  ## Future Enhancements
50
+ * Knowledge Base Expansion: Continuously update the vector store with new data.
51
+ * Retrieval Optimization: Explore GraphRag.
52
+ * Integration with Other Applications: Integrate with other tools or platforms for broader use cases.