deberta-v3-large-AVeriTeC-nli
This model was finetuned from microsoft/deberta-v3-large on an AVeriTec dataset. It achieves the following results on the evaluation set:
Intended uses & limitations
This model is intended for usage in a pipeline for open-domain fact-checking task.
Training and evaluation data
See chenxwh/AVeriTeC
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- optimizer: adamw_torch
- training_precision: float16
- learning_rate: 1e-5
- per_device_train_batch_size: 32
- num_train_epochs: 10
- weight_decay: 0.01
- load_best_model_at_end: True #early stopping!
- warmup_ratio: 0.06
Training results
Framework versions
- Transformers 4.43.0
- TensorFlow 2.17.0
- Datasets 2.20.0
- Tokenizers 0.19.1
- Downloads last month
- 27
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for ctu-aic/deberta-v3-large-AVeriTeC-nli
Base model
microsoft/deberta-v3-largeEvaluation results
- dev macro F1 score on chenxwh/AVeriTeCself-reported0.710
- dev macro recall on chenxwh/AVeriTeCself-reported0.730
- dev macro precision on chenxwh/AVeriTeCself-reported0.710
- dev accuracy on chenxwh/AVeriTeCself-reported0.820