File size: 861 Bytes
1d22a03 336294d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 |
---
library_name: transformers
tags: []
---
# Time Saving Stated Aim Classifier
This is a roberta-base model that is trained to classify whether a set of explicit stated aims extracted from a British [historical patent](https://huggingface.co/datasets/matthewleechen/300YearsOfBritishPatents) includes a time-saving objective.
Labels were manually generated and then checked with Gemini 2.0 Flash with the attached [prompt](https://huggingface.co/matthewleechen/time-saving_stated_aim_classifier/blob/main/time-saving_prompt.txt).
Hyperparameters:
lr = 3e-5
batch size = 128
Test set results:
```text
{'eval_loss': 0.5350350737571716,
'eval_accuracy': 0.89,
'eval_precision': 0.8972633319553535,
'eval_recall': 0.89,
'eval_f1': 0.8891426257718393,
'eval_runtime': 0.4081,
'eval_samples_per_second': 245.053,
'eval_steps_per_second': 2.451}
```
|