Hwanjun commited on
Commit
397e123
·
verified ·
1 Parent(s): d1cae24

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -22,6 +22,7 @@ Here is a Ext2Gen-8B-R2 model that effectively mitigates hallucinations caused b
22
 
23
  See the details in our paper [Link](https://arxiv.org/pdf/2503.04789)
24
 
 
25
  ### What is Ext2Gen-8B-R2?
26
  Ext2Gen-8B-R2 is built upon Llama3.2-8B-Instruct, incorporating preference-aligned fine-tuning through pairwise feedback learning.
27
 
@@ -30,13 +31,15 @@ This training strategy enables the model to:
30
  - Filter out irrelevant or misleading information, reducing hallucinations.
31
  - Align generation with human preferences by optimizing for faithfulness, completeness, and conciseness.
32
 
 
33
  ### Why does Ext2Gen-8B-R2 outperform standard RAG models?
34
  Standard RAG models often struggle due to:
35
  - Uncertain Placement – Relevant information may appear in unpredictable locations within retrieved chunks, making it difficult for LLMs to utilize it effectively.
36
  - Information Overload – The presence of irrelevant chunks can distract the model, leading to errors or hallucinations.
37
  - Lack of Alignment – Most generation models are not explicitly trained to prioritize relevant content over noise.
38
 
39
- ### Need a More Faster Inference?
 
40
  Our Ext2Gen model writes the sentences related to the query first before generating the answer. So, it needs more latency before getting the answer.
41
 
42
  If you don't want to see the extracted sentences but want to directly see the answer with low latency, use its variant we call Gen-8B-R2.
@@ -46,7 +49,6 @@ Link: https://huggingface.co/DISLab/Gen-8B-R2
46
  This model skips the sentence extraction phase but remains its high robustness comparable to Ext2Gen-8B-R2.
47
 
48
 
49
-
50
  ### Prompt
51
 
52
  - query: the query to answer
 
22
 
23
  See the details in our paper [Link](https://arxiv.org/pdf/2503.04789)
24
 
25
+
26
  ### What is Ext2Gen-8B-R2?
27
  Ext2Gen-8B-R2 is built upon Llama3.2-8B-Instruct, incorporating preference-aligned fine-tuning through pairwise feedback learning.
28
 
 
31
  - Filter out irrelevant or misleading information, reducing hallucinations.
32
  - Align generation with human preferences by optimizing for faithfulness, completeness, and conciseness.
33
 
34
+
35
  ### Why does Ext2Gen-8B-R2 outperform standard RAG models?
36
  Standard RAG models often struggle due to:
37
  - Uncertain Placement – Relevant information may appear in unpredictable locations within retrieved chunks, making it difficult for LLMs to utilize it effectively.
38
  - Information Overload – The presence of irrelevant chunks can distract the model, leading to errors or hallucinations.
39
  - Lack of Alignment – Most generation models are not explicitly trained to prioritize relevant content over noise.
40
 
41
+
42
+ ### Need a Faster Inference?
43
  Our Ext2Gen model writes the sentences related to the query first before generating the answer. So, it needs more latency before getting the answer.
44
 
45
  If you don't want to see the extracted sentences but want to directly see the answer with low latency, use its variant we call Gen-8B-R2.
 
49
  This model skips the sentence extraction phase but remains its high robustness comparable to Ext2Gen-8B-R2.
50
 
51
 
 
52
  ### Prompt
53
 
54
  - query: the query to answer