File size: 2,506 Bytes
12681ff d23df37 12681ff a3253ec 12681ff e94598a ed83b7e 722e776 ed83b7e 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 12681ff 8236ee0 6bff02f 12681ff 8236ee0 12681ff 8236ee0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 |
---
language: en
license: mit
library_name: transformers
tags:
- summarization
- bart
datasets: ccdv/arxiv-summarization
model-index:
- name: BARTxiv
results:
- task:
type: summarization
dataset:
name: arxiv-summarization
type: ccdv/arxiv-summarization
split: validation
metrics:
- type: rouge1
value: 41.70204016592095
- type: rouge2
value: 15.134827404979639
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# BARTxiv
This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the [arxiv-summarization](https://huggingface.co/datasets/ccdv/arxiv-summarization) dataset.
It achieves the following results on the validation set:
- Loss: 0.86
- Rouge1: 41.70
- Rouge2: 15.13
- Rougel: 22.85
- Rougelsum: 37.77
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-6
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- optimizer: Adafactor
- num_epochs: 9
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
|:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|
| 1.24 | 1.0 | 1073 | 1.24 | 38.32 | 12.80 | 20.55 | 34.50 |
| 1.04 | 2.0 | 2146 | 1.04 | 39.65 | 13.74 | 21.28 | 35.83 |
| 0.979 | 3.0 | 3219 | 0.98 | 40.19 | 14.30 | 21.87 | 36.38 |
| 0.970 | 4.0 | 4292 | 0.97 | 40.87 | 14.44 | 22.14 | 36.89 |
| 0.918 | 5.0 | 5365 | 0.92 | 41.17 | 14.94 | 22.54 | 37.40 |
| 0.901 | 6.0 | 6438 | 0.90 | 41.02 | 14.65 | 22.46 | 37.05 |
| 0.889 | 7.0 | 7511 | 0.89 | 41.32 | 15.09 | 22.64 | 37.42 |
| 0.900 | 8.0 | 8584 | 0 .90 | 41.23 | 15.02 | 22.67 | 37.28 |
| 0.869 | 9.0 | 9657 | 0.87 | 41.70 | 15.13 | 22.85 | 37.77 |
### Framework versions
- Transformers 4.25.1
- Pytorch 1.13.0+cu117
- Datasets 2.6.1
- Tokenizers 0.13.1 |