Parameters
training_args = TrainingArguments(
output_dir='/kaggle/working/T5_Summarization',
num_train_epochs=1,
warmup_steps=500,
per_device_train_batch_size=1, # Reduce batch size if OOM persists
per_device_eval_batch_size=2,
weight_decay=0.01,
logging_steps=10,
evaluation_strategy='steps',
eval_steps=500,
save_steps=5000, # Save more frequently
gradient_accumulation_steps=16,
fp16=True # Enable mixed precision
)
- Downloads last month
- 3
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for ARSynopsis/T5_Full_FineTune_V0.1_40K
Base model
google-t5/t5-base