Update README.md
Browse files
README.md
CHANGED
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
18 |
|
19 |
# cnn_news_summary_model_trained_on_reduced_data
|
20 |
|
21 |
-
This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an CNN Daily mail dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
- Loss: 1.6597
|
24 |
- Rouge1: 0.2162
|
@@ -29,15 +29,34 @@ It achieves the following results on the evaluation set:
|
|
29 |
|
30 |
## Model description
|
31 |
|
32 |
-
|
|
|
|
|
33 |
|
34 |
## Intended uses & limitations
|
35 |
|
36 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
|
38 |
## Training and evaluation data
|
39 |
|
40 |
-
|
41 |
|
42 |
## Training procedure
|
43 |
|
|
|
18 |
|
19 |
# cnn_news_summary_model_trained_on_reduced_data
|
20 |
|
21 |
+
This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an **CNN Daily mail** dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
- Loss: 1.6597
|
24 |
- Rouge1: 0.2162
|
|
|
29 |
|
30 |
## Model description
|
31 |
|
32 |
+
**Base Model:** *t5-small*, which is a smaller version of the *T5 (Text-to-Text Transfer Transformer) model* developed by ***Google***.
|
33 |
+
|
34 |
+
This model can be particularly useful if you need to quickly summarize large volumes of text, making it easier to digest and understand key information.
|
35 |
|
36 |
## Intended uses & limitations
|
37 |
|
38 |
+
### Intended Use
|
39 |
+
|
40 |
+
The model is designed for **text summarization**, which involves condensing long pieces of text into shorter, more digestible summaries. Here are some specific use cases:
|
41 |
+
|
42 |
+
News Summarization: Quickly summarizing news articles to provide readers with the main points.
|
43 |
+
Document Summarization: Condensing lengthy reports or research papers into brief overviews.
|
44 |
+
Content Curation: Helping content creators and curators to generate summaries for newsletters, blogs, or social media posts.
|
45 |
+
Educational Tools: Assisting students and educators by summarizing academic texts and articles.
|
46 |
+
|
47 |
+
### Limitations
|
48 |
+
|
49 |
+
While the model is powerful, it does have some limitations:
|
50 |
+
|
51 |
+
Accuracy: The summaries generated might not always capture all the key points accurately, especially for complex or nuanced texts.
|
52 |
+
Bias: The model can inherit biases present in the training data, which might affect the quality and neutrality of the summaries.
|
53 |
+
Context Understanding: It might struggle with understanding the full context of very long documents, leading to incomplete or misleading summaries.
|
54 |
+
Language and Style: The model’s output might not always match the desired tone or style, requiring further editing.
|
55 |
+
Data Dependency: Performance can vary depending on the quality and nature of the input data. It performs best on data similar to its training set (news articles)
|
56 |
|
57 |
## Training and evaluation data
|
58 |
|
59 |
+
The model was trained using the Adam optimizer with a learning rate of **2e-05** over **2 epochs**.
|
60 |
|
61 |
## Training procedure
|
62 |
|