File size: 458 Bytes
2f044c1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 |
model:
transformer_model: "microsoft/deberta-v3-large"
optimizer:
lr:
- 0.0001
- 0.00002
warmup_steps: 500
total_steps: ${training.trainer.max_steps}
total_reset: 1
weight_decay: 0.01
lr_decay: 0.9
no_decay_params:
- "bias"
- LayerNorm.weight
other_lr_params:
- "re_subject_projector"
- "re_object_projector"
- "re_relation_projector"
- "re_classifier"
relations_per_forward: 24
entities_per_forward:
|