Update README.md
Browse files
README.md
CHANGED
@@ -22,6 +22,7 @@ Here is a Ext2Gen-8B-R2 model that effectively mitigates hallucinations caused b
|
|
22 |
|
23 |
See the details in our paper [Link](https://arxiv.org/pdf/2503.04789)
|
24 |
|
|
|
25 |
### What is Ext2Gen-8B-R2?
|
26 |
Ext2Gen-8B-R2 is built upon Llama3.2-8B-Instruct, incorporating preference-aligned fine-tuning through pairwise feedback learning.
|
27 |
|
@@ -30,13 +31,15 @@ This training strategy enables the model to:
|
|
30 |
- Filter out irrelevant or misleading information, reducing hallucinations.
|
31 |
- Align generation with human preferences by optimizing for faithfulness, completeness, and conciseness.
|
32 |
|
|
|
33 |
### Why does Ext2Gen-8B-R2 outperform standard RAG models?
|
34 |
Standard RAG models often struggle due to:
|
35 |
- Uncertain Placement – Relevant information may appear in unpredictable locations within retrieved chunks, making it difficult for LLMs to utilize it effectively.
|
36 |
- Information Overload – The presence of irrelevant chunks can distract the model, leading to errors or hallucinations.
|
37 |
- Lack of Alignment – Most generation models are not explicitly trained to prioritize relevant content over noise.
|
38 |
|
39 |
-
|
|
|
40 |
Our Ext2Gen model writes the sentences related to the query first before generating the answer. So, it needs more latency before getting the answer.
|
41 |
|
42 |
If you don't want to see the extracted sentences but want to directly see the answer with low latency, use its variant we call Gen-8B-R2.
|
@@ -46,7 +49,6 @@ Link: https://huggingface.co/DISLab/Gen-8B-R2
|
|
46 |
This model skips the sentence extraction phase but remains its high robustness comparable to Ext2Gen-8B-R2.
|
47 |
|
48 |
|
49 |
-
|
50 |
### Prompt
|
51 |
|
52 |
- query: the query to answer
|
|
|
22 |
|
23 |
See the details in our paper [Link](https://arxiv.org/pdf/2503.04789)
|
24 |
|
25 |
+
|
26 |
### What is Ext2Gen-8B-R2?
|
27 |
Ext2Gen-8B-R2 is built upon Llama3.2-8B-Instruct, incorporating preference-aligned fine-tuning through pairwise feedback learning.
|
28 |
|
|
|
31 |
- Filter out irrelevant or misleading information, reducing hallucinations.
|
32 |
- Align generation with human preferences by optimizing for faithfulness, completeness, and conciseness.
|
33 |
|
34 |
+
|
35 |
### Why does Ext2Gen-8B-R2 outperform standard RAG models?
|
36 |
Standard RAG models often struggle due to:
|
37 |
- Uncertain Placement – Relevant information may appear in unpredictable locations within retrieved chunks, making it difficult for LLMs to utilize it effectively.
|
38 |
- Information Overload – The presence of irrelevant chunks can distract the model, leading to errors or hallucinations.
|
39 |
- Lack of Alignment – Most generation models are not explicitly trained to prioritize relevant content over noise.
|
40 |
|
41 |
+
|
42 |
+
### Need a Faster Inference?
|
43 |
Our Ext2Gen model writes the sentences related to the query first before generating the answer. So, it needs more latency before getting the answer.
|
44 |
|
45 |
If you don't want to see the extracted sentences but want to directly see the answer with low latency, use its variant we call Gen-8B-R2.
|
|
|
49 |
This model skips the sentence extraction phase but remains its high robustness comparable to Ext2Gen-8B-R2.
|
50 |
|
51 |
|
|
|
52 |
### Prompt
|
53 |
|
54 |
- query: the query to answer
|