|
{ |
|
"model_name_or_path": "bert-large-uncased", |
|
"task_name": "hatexplain", |
|
"output_dir": "./output/bert-large/hatexplain/0001/SEED0042/", |
|
"learning_rate": 5e-05, |
|
"num_train_epochs": 3, |
|
"per_device_eval_batch_size": 1, |
|
"per_device_train_batch_size": 1, |
|
"gradient_accumulation_steps": 32, |
|
"seed": 42, |
|
"warmup_steps": 150, |
|
"do_train": true, |
|
"do_eval": true, |
|
"do_predict": false, |
|
"pad_to_max_length": false, |
|
"max_seq_length": 128, |
|
"report_to": [], |
|
"save_strategy": "epoch", |
|
"evaluation_strategy": "epoch", |
|
"logging_steps": 2500, |
|
"use_fast_tokenizer": true, |
|
"group_by_length": true, |
|
"save_training_dynamics": true, |
|
"save_training_dynamics_after_epoch": true |
|
} |