pszemraj commited on
Commit
53eec96
·
verified ·
1 Parent(s): c6e99f5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -10
README.md CHANGED
@@ -746,19 +746,30 @@ Thus far, all completed in fp32 (_using nvidia tf32 dtype behind the scenes when
746
  | Model | Size | CoLA | SST2 | MRPC | STSB | QQP | MNLI | QNLI | RTE | WNLI | Avg |
747
  |------------------------------------|--------|---------|-------|------|-------|------|------------|------|------|-------|---------|
748
  | BEE-spoke-data/bert-plus-L8-4096-v1.0 | 88.1M | 62.72 | 90.6 | 86.59| 92.07 | 90.6| 83.2 | 90.0 | 66.43| 53.52 | TBD |
749
- | bert_uncased_L-8_H-768_A-12 | 81.2M | 55.0 | 91.0 | 88.0 | 93.0 | 90.0 | TBD | TBD | 67.0 | 49.3 | TBD |
750
- | bert-base-uncased | 110M | 52.1 | 93.5 | 88.9 | 85.8 | 71.2 | 84.6/83.4 | 90.5 | 66.4 | TBD | 79.6 |
751
- | roberta-base | 125M | 64.0 | 95.0 | 90.0 | 91.0 | 92.0 | 88.0 | 93.0 | 79.0 | TBD | 86.0 |
752
 
753
  ### Observations:
754
 
755
- - **Performance Variation**: There's notable variation in model performance across different GLUE tasks. This variation can be attributed to the distinct nature of each task, the complexity of the datasets, and how well the model's architecture and hyperparameters are suited to each task.
756
- - **Hyperparameters Impact**: Different weight decay settings and batch sizes seem to have nuanced impacts on performance across tasks, indicating the importance of hyperparameter tuning for optimal results.
757
- - **Batch Size and Gradient Accumulation Steps**: These hyperparameters vary across tasks, reflecting a balance between computational efficiency and model performance. Larger batch sizes and gradient accumulation steps can help stabilize training but may require adjustments based on the available hardware and the specific task.
758
- - **Task-specific Challenges**: Tasks like WNLI and RTE have notably lower accuracy scores compared to others, highlighting the challenges inherent in some NLP tasks, possibly due to dataset size, complexity, or the nuances of the task itself.
759
- - TODO: add more detailed comparison & analysis, but the standard-ctx [google/bert_uncased_L-8_H-768_A-12](https://hf.co/google/bert_uncased_L-8_H-768_A-12) seems to have the same problem. Initial read is that this results from the smaller model size vs. `bert-base-uncased`
760
- - **Overall Performance**: The model shows strong performance on tasks with numerical scores (e.g., STSB), high accuracy in classification tasks like QQP, SST2, and MNLI, but struggles with more nuanced or smaller datasets like WNLI and RTE, underscoring the importance of tailored approaches for different types of NLP challenges.
761
-
 
 
 
 
 
 
 
 
 
 
 
762
  ---
763
 
764
  ## Training procedure
 
746
  | Model | Size | CoLA | SST2 | MRPC | STSB | QQP | MNLI | QNLI | RTE | WNLI | Avg |
747
  |------------------------------------|--------|---------|-------|------|-------|------|------------|------|------|-------|---------|
748
  | BEE-spoke-data/bert-plus-L8-4096-v1.0 | 88.1M | 62.72 | 90.6 | 86.59| 92.07 | 90.6| 83.2 | 90.0 | 66.43| 53.52 | TBD |
749
+ | bert_uncased_L-8_H-768_A-12 | 81.2M | 55.0 | 91.0 | 88.0 | 93.0 | 90.0 | 90.0 | 81.0 | 67.0 | 49.3 | TBD |
750
+ | bert-base-uncased | 110M | 52.1 | 93.5 | 88.9 | 85.8 | 71.2 | 84.6/83.4 | 90.5 | 66.4 | 56.34 | 79.6 |
751
+ | roberta-base | 125M | 64.0 | 95.0 | 90.0 | 91.0 | 92.0 | 88.0 | 93.0 | 79.0 | 56.34 | 86.0 |
752
 
753
  ### Observations:
754
 
755
+
756
+
757
+ 1. **Performance Variation Across Models and Tasks**: The updated data table reveals significant variation not only within model performance across tasks but also across different models for the same tasks. For instance, `BEE-spoke-data/bert-plus-L8-4096-v1.0` and `roberta-base` exhibit strong performance on CoLA and SST2, indicating that both model size and architecture (BERT-based vs. RoBERTa-based) contribute to handling the linguistic complexity and sentiment analysis effectively.
758
+
759
+ 2. **Model Size vs. Task Complexity**: The relationship between model size and task performance is not linear. While `bert-base-uncased` is larger than `bert_uncased_L-8_H-768_A-12`, it does not uniformly outperform the latter across all tasks, such as in MRPC and STSB, where the smaller model performs comparably or even better. This suggests that model architecture optimizations and training strategies might be as crucial as size.
760
+
761
+ 3. **Hyperparameters and Training Strategy**: Observing the performance of `BEE-spoke-data/bert-plus-L8-4096-v1.0` with a size of 88.1M suggests that beyond hyperparameters, model design tailored to specific NLP tasks (e.g., layer optimizations or attention mechanisms) can significantly impact outcomes. The use of fp32 and NVIDIA tf32 indicates a balance between computational efficiency and maintaining model performance, highlighting the importance of choosing the right training precision mode based on the hardware capabilities and task requirements.
762
+
763
+ 4. **Task-specific Challenges and Dataset Nuances**: The lower performance on WNLI and RTE for all models underscores the continued challenge of dealing with small datasets and tasks requiring nuanced understanding or logic. These results hint at potential areas for improvement, such as data augmentation, advanced pre-training techniques, or more sophisticated reasoning capabilities embedded into models.
764
+
765
+ 5. **Overall Performance and Efficiency**: When considering overall performance, `roberta-base` stands out for its high average score, showcasing the effectiveness of its architecture and pre-training approach for a wide range of tasks. However, `BEE-spoke-data/bert-plus-L8-4096-v1.0` demonstrates competitive performance with a smaller model size, indicating a noteworthy efficiency-performance trade-off. This suggests that optimizations tailored to specific tasks can yield high efficiency without drastically increasing model size.
766
+
767
+ 6. **Impact of Computational Precision**: The mention of fp32 and NVIDIA tf32 behind the scenes is a critical observation for model training strategies, indicating that maintaining high precision can lead to better performance across various tasks. This is particularly relevant for tasks that may be sensitive to the precision of calculations, such as STSB, which involves regression.
768
+
769
+ 7. **Insights for Future Model Development**: The varied performance across tasks and models emphasizes the importance of continuous experimentation with model architectures, training strategies, and precision settings. It also highlights the need for more targeted approaches to improve performance on challenging tasks like WNLI and RTE, possibly through more sophisticated reasoning capabilities or enhanced training datasets.
770
+
771
+ In summary, these observations reflect the nuanced landscape of model performance across the GLUE benchmark tasks. They underscore the importance of model architecture, size, training strategies, and computational precision in achieving optimal performance. Moreover, they highlight the ongoing challenges and opportunities for NLP research, particularly in addressing tasks that require deep linguistic understanding or reasoning.
772
+
773
  ---
774
 
775
  ## Training procedure