ashaduzzaman commited on
Commit
d9f36f3
·
verified ·
1 Parent(s): e2b472d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +92 -34
README.md CHANGED
@@ -11,58 +11,116 @@ metrics:
11
  model-index:
12
  - name: mt5-finetuned-amazon-reviews
13
  results: []
 
 
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
 
 
 
 
19
  # mt5-finetuned-amazon-reviews
20
 
21
- This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset.
22
- It achieves the following results on the evaluation set:
23
- - Loss: 4.2617
24
- - Rouge1: 0.0
25
- - Rouge2: 0.0
26
- - Rougel: 0.0
27
- - Rougelsum: 0.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
- ## Model description
 
 
 
 
30
 
31
- More information needed
32
 
33
- ## Intended uses & limitations
34
 
35
- More information needed
36
 
37
- ## Training and evaluation data
38
 
39
- More information needed
 
 
 
 
 
 
40
 
41
- ## Training procedure
42
 
43
- ### Training hyperparameters
 
 
 
 
44
 
45
- The following hyperparameters were used during training:
46
- - learning_rate: 5.6e-05
47
- - train_batch_size: 8
48
- - eval_batch_size: 8
49
- - seed: 42
50
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
- - lr_scheduler_type: linear
52
- - num_epochs: 3
53
 
54
- ### Training results
 
 
 
55
 
56
- | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
57
- |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|
58
- | 18.4463 | 1.0 | 379 | 8.5447 | 0.3663 | 0.0 | 0.3663 | 0.3663 |
59
- | 9.359 | 2.0 | 758 | 5.0674 | 0.0 | 0.0 | 0.0 | 0.0 |
60
- | 6.6153 | 3.0 | 1137 | 4.2617 | 0.0 | 0.0 | 0.0 | 0.0 |
61
 
 
 
62
 
63
- ### Framework versions
64
 
65
- - Transformers 4.42.4
66
- - Pytorch 2.3.1+cu121
67
- - Datasets 2.21.0
68
- - Tokenizers 0.19.1
 
11
  model-index:
12
  - name: mt5-finetuned-amazon-reviews
13
  results: []
14
+ datasets:
15
+ - mteb/amazon_reviews_multi
16
+ pipeline_tag: summarization
17
  ---
18
 
19
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
20
  should probably proofread and complete it, then remove this comment. -->
21
 
22
+ Here's a detailed model card for the `mt5-finetuned-amazon-reviews` model, incorporating the provided information and adding relevant details:
23
+
24
+ ---
25
+
26
  # mt5-finetuned-amazon-reviews
27
 
28
+ This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) specifically trained to generate summaries of Amazon product reviews. It is designed to handle multilingual text summarization tasks, leveraging the capabilities of the mT5 (Multilingual T5) model.
29
+
30
+ ## Model Details
31
+
32
+ - **Model Name:** mt5-finetuned-amazon-reviews
33
+ - **Base Model:** [google/mt5-small](https://huggingface.co/google/mt5-small)
34
+ - **Model Type:** Multilingual Transformer-based Text-to-Text Generation Model
35
+ - **Fine-tuned on:** Amazon product reviews (dataset specifics unknown)
36
+
37
+ ### Model Description
38
+
39
+ The `mt5-finetuned-amazon-reviews` model utilizes the mT5 architecture, a variant of T5 that is pre-trained on a diverse multilingual dataset. This fine-tuned model targets the summarization of customer reviews on Amazon products, aiming to distill lengthy reviews into concise and informative summaries. It is optimized for multilingual settings, enabling it to handle reviews written in various languages.
40
+
41
+ ### Intended Uses & Limitations
42
+
43
+ **Intended Uses:**
44
+ - Summarizing customer reviews from Amazon to provide quick insights into product feedback.
45
+ - Assisting e-commerce platforms in analyzing customer sentiment and satisfaction.
46
+ - Providing concise information for consumers to make informed purchasing decisions.
47
+
48
+ **Limitations:**
49
+ - The model may not perform well on non-Amazon or highly specialized reviews.
50
+ - Its zero scores on ROUGE metrics suggest potential issues with generating high-quality summaries, which could be due to limitations in the training data or training process.
51
+ - The model's performance on languages not sufficiently represented in the training data may be suboptimal.
52
+
53
+ ### Usage
54
+
55
+ To use the model for summarization, you can utilize the following code snippet:
56
+
57
+ ```python
58
+ from transformers import pipeline
59
+
60
+ hub_model_id = "ashaduzzaman/mt5-finetuned-amazon-reviews"
61
+ summarizer = pipeline("summarization", model=hub_model_id)
62
+
63
+ text = (
64
+ "Nothing special at all about this product... the book is too small and stiff and hard to write in. "
65
+ "The huge sticker on the back doesn’t come off and looks super tacky. I would not purchase this again. "
66
+ "I could have just bought a journal from the dollar store and it would be basically the same thing. "
67
+ "It’s also really expensive for what it is."
68
+ )
69
+ summarizer(text)
70
+ ```
71
+
72
+ ### Training and Evaluation Data
73
+
74
+ The specific dataset used for fine-tuning this model is not provided, but it is implied to be related to Amazon product reviews. The absence of detailed information about the training dataset limits the understanding of the model's training scope and diversity.
75
+
76
+ ### Evaluation Results
77
+
78
+ The model was evaluated using standard text summarization metrics, but the results indicate challenges in its performance:
79
 
80
+ - **Loss:** 4.2617
81
+ - **ROUGE-1:** 0.0
82
+ - **ROUGE-2:** 0.0
83
+ - **ROUGE-L:** 0.0
84
+ - **ROUGE-Lsum:** 0.0
85
 
86
+ These scores suggest that the model may have struggled to produce meaningful summaries or that the evaluation dataset was not aligned well with the training data. The zero ROUGE scores highlight a need for further investigation into the training process and data quality.
87
 
88
+ ### Training Procedure
89
 
90
+ The model was fine-tuned using the following hyperparameters and configuration:
91
 
92
+ #### Training Hyperparameters
93
 
94
+ - **Learning Rate:** 5.6e-05
95
+ - **Training Batch Size:** 8
96
+ - **Evaluation Batch Size:** 8
97
+ - **Random Seed:** 42
98
+ - **Optimizer:** Adam (betas=(0.9, 0.999), epsilon=1e-08)
99
+ - **Learning Rate Scheduler:** Linear
100
+ - **Number of Epochs:** 3
101
 
102
+ #### Training Results
103
 
104
+ | Training Loss | Epoch | Step | Validation Loss | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-Lsum |
105
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:----------:|
106
+ | 18.4463 | 1.0 | 379 | 8.5447 | 0.3663 | 0.0 | 0.3663 | 0.3663 |
107
+ | 9.359 | 2.0 | 758 | 5.0674 | 0.0 | 0.0 | 0.0 | 0.0 |
108
+ | 6.6153 | 3.0 | 1137 | 4.2617 | 0.0 | 0.0 | 0.0 | 0.0 |
109
 
110
+ ### Framework Versions
 
 
 
 
 
 
 
111
 
112
+ - **Transformers:** 4.42.4
113
+ - **PyTorch:** 2.3.1+cu121
114
+ - **Datasets:** 2.21.0
115
+ - **Tokenizers:** 0.19.1
116
 
117
+ ### Ethical Considerations
 
 
 
 
118
 
119
+ - **Bias:** The model's summaries could reflect biases present in the training data, especially if the data is not balanced across different product categories or customer demographics.
120
+ - **Data Privacy:** Ensure that the use of the model complies with data privacy regulations, especially when using customer review data that may contain sensitive or personally identifiable information.
121
 
122
+ ### Future Improvements
123
 
124
+ - Collecting a more comprehensive and representative training dataset could improve summarization quality.
125
+ - Further fine-tuning and experimenting with different hyperparameters might yield better performance.
126
+ - Incorporating more evaluation metrics and detailed qualitative analysis could provide deeper insights into the model's strengths and weaknesses.