jordiclive commited on
Commit
905c83b
1 Parent(s): ec81c0f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -19,6 +19,7 @@ datasets:
19
  - samsum
20
  - scitldr/AIC
21
  - billsum
 
22
  metrics:
23
  - rouge
24
  widget:
@@ -133,15 +134,15 @@ inference:
133
  num_beams: 4
134
  ---
135
 
136
- # Longformer Encoder-Decoder (LED) for Narrative-Esque Long Text Summarization
137
 
138
  <a href="https://colab.research.google.com/gist/pszemraj/3eba944ddc9fc9a4a1bfb21e83b57620/summarization-token-batching.ipynb">
139
  <img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/>
140
  </a>
141
 
142
- A fine-tuned version of [allenai/led-large-16384](https://huggingface.co/allenai/led-large-16384) on the `BookSum` dataset.
143
 
144
- Goal: a model that can generalize well and is useful in summarizing long text in academic and daily usage. The result works well on lots of text and can handle 16384 tokens/batch (_if you have the GPU memory to handle that_)
145
 
146
  - See the Colab demo linked above or try the [demo on Spaces](https://huggingface.co/spaces/pszemraj/summarize-long-text)
147
 
 
19
  - samsum
20
  - scitldr/AIC
21
  - billsum
22
+ - TLDR
23
  metrics:
24
  - rouge
25
  widget:
 
134
  num_beams: 4
135
  ---
136
 
137
+ # Multi-purpose Summarizer (Fine-tuned google/flan-t5-xl (3B) on several Summarization datasets)
138
 
139
  <a href="https://colab.research.google.com/gist/pszemraj/3eba944ddc9fc9a4a1bfb21e83b57620/summarization-token-batching.ipynb">
140
  <img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/>
141
  </a>
142
 
143
+ A fine-tuned version of [google/flan-t5-xl](https://huggingface.co/google/flan-t5-xl) on various summarization datasets (xsum, wikihow, cnn_dailymail/3.0.0, samsum, scitldr/AIC, billsum, TLDR)
144
 
145
+ Goal: a model that can be used for a general-purpose summarizer for academic and general usage. Control over the type of summary can be given by varying the instruction prepended to the source document. The result works well on lots of text, although trained with a max source length of 512 tokens and 150 max summary length.
146
 
147
  - See the Colab demo linked above or try the [demo on Spaces](https://huggingface.co/spaces/pszemraj/summarize-long-text)
148