DeathReaper0965
commited on
Commit
•
e5ce719
1
Parent(s):
ef71046
Update README.md
Browse files
README.md
CHANGED
@@ -43,7 +43,7 @@ inference:
|
|
43 |
---
|
44 |
|
45 |
# Flan-T5 (base-sized) Dialogue Summarization with reduced toxicity using RLAIF
|
46 |
-
This model is a two-fold fine-tuned [Flan-T5 model](https://huggingface.co/google/flan-t5-base) firstly on the [SAMSUM](https://huggingface.co/datasets/samsum) dataset followed by further fine-tuning using Reinforcement Learning from AI Feedback(RLAIF) to detoxify model outputs. <br>
|
47 |
Anthropic's Costitutional AI [paper](https://arxiv.org/abs/2212.08073) from 2022, provides some amazing insights on how RLAIF can be leveraged. Do check out if interested!<br>
|
48 |
|
49 |
More, specifically I've fine-tuned this model on a single downstream task of Dialogue Summarization on the above mentioned dataset with a primary objective of reduced toxicity in generated summaries.
|
|
|
43 |
---
|
44 |
|
45 |
# Flan-T5 (base-sized) Dialogue Summarization with reduced toxicity using RLAIF
|
46 |
+
This model is a **two-fold fine-tuned** [Flan-T5 model](https://huggingface.co/google/flan-t5-base) firstly on the [SAMSUM](https://huggingface.co/datasets/samsum) dataset followed by further fine-tuning using **Reinforcement Learning from AI Feedback(RLAIF)** to detoxify model outputs. <br>
|
47 |
Anthropic's Costitutional AI [paper](https://arxiv.org/abs/2212.08073) from 2022, provides some amazing insights on how RLAIF can be leveraged. Do check out if interested!<br>
|
48 |
|
49 |
More, specifically I've fine-tuned this model on a single downstream task of Dialogue Summarization on the above mentioned dataset with a primary objective of reduced toxicity in generated summaries.
|