Update README.md
Browse files
README.md
CHANGED
@@ -3,11 +3,16 @@ library_name: transformers
|
|
3 |
tags:
|
4 |
- trl
|
5 |
- sft
|
|
|
|
|
6 |
---
|
7 |
|
8 |
# Model Card for Model ID
|
9 |
|
10 |
<!-- Provide a quick summary of what the model is/does. -->
|
|
|
|
|
|
|
11 |
|
12 |
|
13 |
|
@@ -16,8 +21,29 @@ tags:
|
|
16 |
### Model Description
|
17 |
|
18 |
<!-- Provide a longer summary of what this model is. -->
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
|
20 |
-
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
21 |
|
22 |
- **Developed by:** [More Information Needed]
|
23 |
- **Funded by [optional]:** [More Information Needed]
|
|
|
3 |
tags:
|
4 |
- trl
|
5 |
- sft
|
6 |
+
base_model:
|
7 |
+
- Qwen/Qwen2.5-7B
|
8 |
---
|
9 |
|
10 |
# Model Card for Model ID
|
11 |
|
12 |
<!-- Provide a quick summary of what the model is/does. -->
|
13 |
+
The GlueQwen model is fine-tuned on four distinct tasks from the GLUE benchmark: SST-2 (Sentiment Analysis), MRPC (Paraphrase Detection), CoLA (Linguistic Acceptability), and MNLI (Natural Language Inference). The base model used is Qwen/Qwen2.5-7B, which has 7 billion parameters. Qwen2.5-7B is designed to enhance various natural language understanding tasks through pre-training on diverse datasets, followed by fine-tuning for task-specific improvements.
|
14 |
+
|
15 |
+
The fine-tuning of GlueQwen involves optimizing the model for these GLUE tasks, aiming to measure catastrophic forgetting, model learning abilities, and overall training performance across different language tasks. These benchmarks provide insight into how well the model retains previous knowledge while learning new tasks sequentially.
|
16 |
|
17 |
|
18 |
|
|
|
21 |
### Model Description
|
22 |
|
23 |
<!-- Provide a longer summary of what this model is. -->
|
24 |
+
### Benchmark Table for GlueQwen Fine-Tuning Performance
|
25 |
+
|
26 |
+
| **Model** | **Parameter Size (B)** | **Pretrained Performance** | **Forgetting** | **Learning** | **Training Performance** |
|
27 |
+
|-----------------|------------------------|----------------------------|----------------|--------------|--------------------------|
|
28 |
+
| Llama-3.2-1B | 1 | 0.50 | 0.24 | 0.33 | 0.54 |
|
29 |
+
| Llama-3.2-3B | 3 | 0.56 | 0.225 | 0.36 | 0.61 |
|
30 |
+
| Llama-3.1-8B | 8 | 0.56 | 0.59 | 0.84 | 0.67 |
|
31 |
+
| Llama-3-8B | 8 | 0.53 | 0.39 | 0.98 | 0.70 |
|
32 |
+
| Llama-2-7B | 7 | 0.67 | 0.23 | 0.12 | 0.63 |
|
33 |
+
| GPT-J-6B | 6 | 0.50 | 0.39 | 0.45 | 0.54 |
|
34 |
+
| Phi-2 | 2.7 | 0.59 | 0.10 | 0.15 | 0.61 |
|
35 |
+
| Phi-3.5-mini | 3.82 | 0.69 | **0.02** | 0.30 | 0.76 |
|
36 |
+
| Orca-2-7b | 7 | **0.76** | 0.185 | 0.33 | **0.81** |
|
37 |
+
| Qwen2.5-0.5B | 0.5 | 0.52 | 0.23 | 0.56 | 0.61 |
|
38 |
+
| Qwen2.5-7B | 7 | 0.56 | 0.51 | **1.12** | 0.77 |
|
39 |
+
| Qwen2.5-14B | 14 | 0.71 | 0.935 | 0.66 | 0.80 |
|
40 |
+
| **GlueQwen** | **7** | **0.59** | **0.42** | **0.97** | **0.73** |
|
41 |
+
|
42 |
+
|
43 |
+
### Analysis
|
44 |
+
|
45 |
+
GlueQwen, fine-tuned on multiple tasks from the GLUE dataset, demonstrates a pre-trained performance of 0.59. Its forgetting rate is moderate at 0.42, reflecting some loss of previously learned information. However, the model exhibits a strong learning capability with a learning score of 0.97. The overall training performance stands at 0.73, positioning GlueQwen as a balanced model that manages forgetting while achieving significant improvements in task-specific learning.
|
46 |
|
|
|
47 |
|
48 |
- **Developed by:** [More Information Needed]
|
49 |
- **Funded by [optional]:** [More Information Needed]
|