--- library_name: transformers language: - en license: mit base_model: JeremiahZ/roberta-base-cola tags: - generated_from_trainer datasets: - glue model-index: - name: roberta-base-relu-cola results: [] --- # roberta-base-relu-cola This model is a fine-tuned version of [JeremiahZ/roberta-base-cola](https://huggingface.co/JeremiahZ/roberta-base-cola) on the GLUE COLA dataset. It achieves the following results on the evaluation set: - eval_loss: 1.2395 - eval_model_preparation_time: 0.0024 - eval_matthews_correlation: 0.5652 - eval_runtime: 9.4256 - eval_samples_per_second: 110.656 - eval_steps_per_second: 27.69 - step: 0 ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-5, 2e-5, 3e-5 - train_batch_size: 16 - eval_batch_size: 4 - seed: 42 - num_epochs: 10 The best model was selected based on the highest accuracy, which is the key evaluation metric for this task. ### Framework versions - Transformers 4.50.0.dev0 - Pytorch 2.6.0+cu124 - Datasets 3.3.2 - Tokenizers 0.21.0