OSUM / train.yaml
xlgeng's picture
xx
2defe04
raw
history blame
3.04 kB
accum_grad: 4
adapter_type: gxl
cmvn: null
cmvn_conf:
cmvn_file: null
is_json_cmvn: null
ctc_conf:
ctc_blank_id: 50362
dataset: asr
dataset_conf:
batch_conf:
batch_size: 26
batch_type: dynamic
max_frames_in_batch: 2700
max_seq_in_batch: 2000
cycle: 100
emotion_en2zh_dict: conf/en2zh4emotion.json
eod_id: 151643
feats_type: log_mel_spectrogram
filter_conf:
filter_no_extra_info: true
max_length: 1000
max_seq_len: 950
min_length: 0
token_max_length: 360
token_min_length: 1
language_conf:
limited_langs:
- zh
log_mel_spectrogram_conf:
hop_length: 160
n_fft: 400
num_mel_bins: 80
padding: 0
resample_conf:
resample_rate: 16000
shuffle: true
shuffle_conf:
shuffle_size: 1500
sort: true
sort_conf:
sort_size: 500
spec_aug: true
spec_aug_conf:
max_f: 10
max_t: 50
num_f_mask: 2
num_t_mask: 2
spec_sub: true
spec_sub_conf:
max_t: 30
num_t_sub: 3
spec_trim: false
speed_perturb: false
split_num: 10
decoder: transformer
decoder_conf:
activation_type: gelu
attention_heads: 16
dropout_rate: 0.1
gradient_checkpointing: true
input_layer: embed_learnable_pe
key_bias: false
linear_units: 4096
normalize_before: true
num_blocks: 24
positional_dropout_rate: 0.1
self_attention_dropout_rate: 0.0
src_attention: true
src_attention_dropout_rate: 0.0
tie_word_embedding: true
use_output_layer: true
downsample_rate: 4
dtype: bf16
encoder: transformer
encoder_conf:
activation_type: gelu
attention_dropout_rate: 0.0
attention_heads: 16
dropout_rate: 0.1
gradient_checkpointing: true
input_layer: conv1d2
key_bias: false
linear_units: 4096
normalize_before: true
num_blocks: 24
output_size: 1024
pos_enc_layer_type: abs_pos_whisper
positional_dropout_rate: 0.1
static_chunk_size: -1
use_dynamic_chunk: false
use_dynamic_left_chunk: false
epoch: 11
fire_module: link_and_encoder_and_lora
grad_clip: 5
init_step: false
input_dim: 80
llm_path: Qwen/Qwen2-7B #/home/node54_tmpdata/xlgeng/ckpt/qwen-7B-instruct/qwen2_7b
log_interval: 10
lora_alpha: 32
lora_dropout: 0.1
lora_rank: 8
loss_dict:
acc: 0.0
loss: 1.4107781417203814e-05
lrs:
- 1.3727899783270902e-05
max_epoch: 100
model: llmasr
model_conf:
ctc_weight: 0
length_normalized_loss: false
lsm_weight: 0.1
model_dir: /home/node54_tmpdata/xlgeng/ckpt/wenet_whisper_finetune_xlgeng/examples/wenetspeech/whisper/exp/qwen2_multi_task_4_6gpus_gxl_adapter/update_data/epoch_1_with_token
optim: adamw
optim_conf:
betas:
- 0.9
- 0.99
eps: 1.0e-06
lr: 5.0e-05
weight_decay: 0.01
output_dim: 151646
save_interval: 5000
save_states: model+optimizer
save_time: 14/01/2025 20:35:10
scheduler: warmuplr
scheduler_conf:
warmup_steps: 8000
speech_token_num: 4097
step: 106124
tag: epoch_11
tokenizer: huggingface
tokenizer_conf:
llm_path: Qwen/Qwen2-7B #/home/node54_tmpdata/xlgeng/ckpt/qwen-7B-instruct/qwen2_7b
train_engine: deepspeed
use_amp: true
use_lora: true
vocab_size: 151646