Spaces:
Sleeping
Sleeping
Update README.md
Browse files
README.md
CHANGED
@@ -12,42 +12,41 @@ sdk_version: 1.38.0
|
|
12 |
This project demonstrates a self Reflective corrective Retrieval Augmented Generation (CRAG) application built using LangGraph. The application leverages a Gemma2 9B LLM to provide informative and relevant responses to user queries. It employs a multi-agent approach, incorporating various components for enhanced performance and user experience.
|
13 |
|
14 |
# Key Features
|
15 |
-
Vector Store: Uses Chroma Vector Store to efficiently store and retrieve context from scraped webpages related to data science and programming.
|
16 |
-
Prompt Guard: Ensures question safety by checking against predefined guidelines.
|
17 |
-
LLM Graders: Evaluates question relevance, answer grounding, and helpfulness to maintain high-quality responses.
|
18 |
-
Retrieval and Generation: Combines context retrieval from vector store and web search with LLM generation to provide comprehensive answers.
|
19 |
-
Iterative Refinement: Rewrites questions and regenerates answers as needed to ensure accuracy and relevance.
|
20 |
-
Customization: Offers flexibility in model selection, fine-tuning, and retrieval methods to tailor the application to specific requirements.
|
21 |
-
Local Deployment: Can be deployed locally for enhanced user data privacy.
|
22 |
## Technical Specifications
|
23 |
-
LLM: Gemma2 9B
|
24 |
-
Vector Store: Chroma
|
25 |
-
Embeddings: Alibaba-NLP/gte-base-en-v1.5
|
26 |
-
Workflow: LangGraph
|
27 |
-
Model API: ChatGroq
|
28 |
-
Web Search: Wikipedia and Google SERP
|
29 |
## Workflow
|
30 |
-
User Query: User inputs a question.
|
31 |
-
Prompt Guard: Checks if the question is safe and appropriate.
|
32 |
-
Context Retrieval: Searches the vector store for relevant documents.
|
33 |
-
Document Relevance: Evaluates document relevance using LLM graders.
|
34 |
-
Web Search: If necessary, conducts web searches on Wikipedia and Google SERP.
|
35 |
-
Answer Generation: Generates a response using the retrieved documents and LLM.
|
36 |
-
Answer Evaluation: Evaluates answer grounding and helpfulness using LLM graders.
|
37 |
-
Refinement: If necessary, rewrites the question or regenerates the answer.
|
38 |
## Customization Options
|
39 |
-
Model Selection: Choose different LLM models based on specific needs (e.g., larger models for more complex tasks).
|
40 |
-
Fine-Tuning: Fine-tune the LLM to match specific styles or domains.
|
41 |
-
Retrieval Methods: Explore alternative vector stores or retrieval techniques.
|
42 |
## Local Deployment
|
43 |
-
To deploy the application locally, follow these steps:
|
44 |
|
45 |
-
Set up environment: Install required dependencies (LangGraph, Chroma, LLM API, etc.).
|
46 |
-
Prepare data: Scrape webpages and create the vector store.
|
47 |
-
Configure workflow: Define the workflow and LLM graders.
|
48 |
-
Run application: Execute the application to start processing user queries.
|
49 |
## Future Enhancements
|
50 |
-
Knowledge Base Expansion: Continuously update the vector store with new data.
|
51 |
-
Retrieval Optimization: Explore
|
52 |
-
|
53 |
-
Integration with Other Applications: Integrate with other tools or platforms for broader use cases.
|
|
|
12 |
This project demonstrates a self Reflective corrective Retrieval Augmented Generation (CRAG) application built using LangGraph. The application leverages a Gemma2 9B LLM to provide informative and relevant responses to user queries. It employs a multi-agent approach, incorporating various components for enhanced performance and user experience.
|
13 |
|
14 |
# Key Features
|
15 |
+
* Vector Store: Uses Chroma Vector Store to efficiently store and retrieve context from scraped webpages related to data science and programming.
|
16 |
+
* Prompt Guard: Ensures question safety by checking against predefined guidelines.
|
17 |
+
* LLM Graders: Evaluates question relevance, answer grounding, and helpfulness to maintain high-quality responses.
|
18 |
+
* Retrieval and Generation: Combines context retrieval from vector store and web search with LLM generation to provide comprehensive answers.
|
19 |
+
* Iterative Refinement: Rewrites questions and regenerates answers as needed to ensure accuracy and relevance.
|
20 |
+
* Customization: Offers flexibility in model selection, fine-tuning, and retrieval methods to tailor the application to specific requirements.
|
21 |
+
* Local Deployment: Can be deployed locally for enhanced user data privacy.
|
22 |
## Technical Specifications
|
23 |
+
* LLM: Gemma2 9B
|
24 |
+
* Vector Store: Chroma
|
25 |
+
* Embeddings: Alibaba-NLP/gte-base-en-v1.5
|
26 |
+
* Workflow: LangGraph
|
27 |
+
* Model API: ChatGroq
|
28 |
+
* Web Search: Wikipedia and Google SERP
|
29 |
## Workflow
|
30 |
+
* User Query: User inputs a question.
|
31 |
+
* Prompt Guard: Checks if the question is safe and appropriate.
|
32 |
+
* Context Retrieval: Searches the vector store for relevant documents.
|
33 |
+
* Document Relevance: Evaluates document relevance using LLM graders.
|
34 |
+
* Web Search: If necessary, conducts web searches on Wikipedia and Google SERP.
|
35 |
+
* Answer Generation: Generates a response using the retrieved documents and LLM.
|
36 |
+
* Answer Evaluation: Evaluates answer grounding and helpfulness using LLM graders.
|
37 |
+
* Refinement: If necessary, rewrites the question or regenerates the answer.
|
38 |
## Customization Options
|
39 |
+
* Model Selection: Choose different LLM models based on specific needs (e.g., larger models for more complex tasks).
|
40 |
+
* Fine-Tuning: Fine-tune the LLM to match specific styles or domains.
|
41 |
+
* Retrieval Methods: Explore alternative vector stores or retrieval techniques.
|
42 |
## Local Deployment
|
43 |
+
* To deploy the application locally, follow these steps:
|
44 |
|
45 |
+
* Set up environment: Install required dependencies (LangGraph, Chroma, LLM API, etc.).
|
46 |
+
* Prepare data: Scrape webpages and create the vector store.
|
47 |
+
* Configure workflow: Define the workflow and LLM graders.
|
48 |
+
* Run application: Execute the application to start processing user queries.
|
49 |
## Future Enhancements
|
50 |
+
* Knowledge Base Expansion: Continuously update the vector store with new data.
|
51 |
+
* Retrieval Optimization: Explore GraphRag.
|
52 |
+
* Integration with Other Applications: Integrate with other tools or platforms for broader use cases.
|
|