|
|
|
tokenizer: |
|
_component_: torchtune.models.llama3.llama3_tokenizer |
|
path: ../../slice_with_mergekit/merged/original/tokenizer.model |
|
|
|
|
|
dataset: |
|
_component_: custom_datasets.orpo_dpo_mix_40k_dataset |
|
max_seq_len: 4096 |
|
|
|
|
|
seed: 42 |
|
shuffle: True |
|
batch_size: 2 |
|
|
|
|
|
model: |
|
_component_: torchtune.models.llama3.lora_llama3 |
|
vocab_size: 128256 |
|
num_layers: 20 |
|
num_heads: 32 |
|
num_kv_heads: 8 |
|
embed_dim: 4096 |
|
max_seq_len: 4096 |
|
intermediate_dim: 14336 |
|
attn_dropout: 0.0 |
|
norm_eps: 1e-5 |
|
rope_base: 500000.0 |
|
lora_attn_modules: ["q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj"] |
|
apply_lora_to_mlp: True |
|
apply_lora_to_output: True |
|
lora_rank: 96 |
|
lora_alpha: 192 |
|
quantize_base: True |
|
|
|
|
|
checkpointer: |
|
_component_: torchtune.utils.FullModelHFCheckpointer |
|
checkpoint_dir: ../../slice_with_mergekit/merged/ |
|
checkpoint_files: [ |
|
model-00001-of-00003.safetensors, |
|
model-00002-of-00003.safetensors, |
|
model-00003-of-00003.safetensors |
|
] |
|
adapter_checkpoint: |
|
recipe_checkpoint: |
|
output_dir: /media/hailey/More/AI/PruneMe/train/torchtune/llama3-5b/ |
|
model_type: LLAMA3 |
|
resume_from_checkpoint: False |
|
|
|
|
|
epochs: 3 |
|
optimizer: |
|
_component_: torch.optim.AdamW |
|
lr: 3e-6 |
|
lr_scheduler: |
|
_component_: torchtune.modules.get_cosine_schedule_with_warmup |
|
num_warmup_steps: 1500 |
|
|
|
|
|
|
|
|
|
|
|
loss: |
|
_component_: torch.nn.CrossEntropyLoss |
|
|
|
max_steps_per_epoch: 500 |
|
gradient_accumulation_steps: 2 |
|
optimizer_in_bwd: False |
|
compile: False |
|
|
|
|
|
device: cuda |
|
|
|
|
|
enable_activation_checkpointing: True |
|
|
|
|
|
dtype: fp32 |
|
|
|
profiler: |
|
_component_: torchtune.utils.profiler |
|
enabled: False |
|
|
|
|
|
|
|
metric_logger: |
|
_component_: torchtune.utils.metric_logging.WandBLogger |
|
|
|
project: llama3-5b |
|
output_dir: /media/hailey/More/AI/PruneMe/train/torchtune/llama3-5b-dpo/ |
|
log_every_n_steps: 1 |
|
log_peak_memory_stats: False |
|
|