xlgeng commited on
Commit
2defe04
·
1 Parent(s): a40a139
Files changed (1) hide show
  1. train.yaml +139 -0
train.yaml ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ accum_grad: 4
2
+ adapter_type: gxl
3
+ cmvn: null
4
+ cmvn_conf:
5
+ cmvn_file: null
6
+ is_json_cmvn: null
7
+ ctc_conf:
8
+ ctc_blank_id: 50362
9
+ dataset: asr
10
+ dataset_conf:
11
+ batch_conf:
12
+ batch_size: 26
13
+ batch_type: dynamic
14
+ max_frames_in_batch: 2700
15
+ max_seq_in_batch: 2000
16
+ cycle: 100
17
+ emotion_en2zh_dict: conf/en2zh4emotion.json
18
+ eod_id: 151643
19
+ feats_type: log_mel_spectrogram
20
+ filter_conf:
21
+ filter_no_extra_info: true
22
+ max_length: 1000
23
+ max_seq_len: 950
24
+ min_length: 0
25
+ token_max_length: 360
26
+ token_min_length: 1
27
+ language_conf:
28
+ limited_langs:
29
+ - zh
30
+ log_mel_spectrogram_conf:
31
+ hop_length: 160
32
+ n_fft: 400
33
+ num_mel_bins: 80
34
+ padding: 0
35
+ resample_conf:
36
+ resample_rate: 16000
37
+ shuffle: true
38
+ shuffle_conf:
39
+ shuffle_size: 1500
40
+ sort: true
41
+ sort_conf:
42
+ sort_size: 500
43
+ spec_aug: true
44
+ spec_aug_conf:
45
+ max_f: 10
46
+ max_t: 50
47
+ num_f_mask: 2
48
+ num_t_mask: 2
49
+ spec_sub: true
50
+ spec_sub_conf:
51
+ max_t: 30
52
+ num_t_sub: 3
53
+ spec_trim: false
54
+ speed_perturb: false
55
+ split_num: 10
56
+ decoder: transformer
57
+ decoder_conf:
58
+ activation_type: gelu
59
+ attention_heads: 16
60
+ dropout_rate: 0.1
61
+ gradient_checkpointing: true
62
+ input_layer: embed_learnable_pe
63
+ key_bias: false
64
+ linear_units: 4096
65
+ normalize_before: true
66
+ num_blocks: 24
67
+ positional_dropout_rate: 0.1
68
+ self_attention_dropout_rate: 0.0
69
+ src_attention: true
70
+ src_attention_dropout_rate: 0.0
71
+ tie_word_embedding: true
72
+ use_output_layer: true
73
+ downsample_rate: 4
74
+ dtype: bf16
75
+ encoder: transformer
76
+ encoder_conf:
77
+ activation_type: gelu
78
+ attention_dropout_rate: 0.0
79
+ attention_heads: 16
80
+ dropout_rate: 0.1
81
+ gradient_checkpointing: true
82
+ input_layer: conv1d2
83
+ key_bias: false
84
+ linear_units: 4096
85
+ normalize_before: true
86
+ num_blocks: 24
87
+ output_size: 1024
88
+ pos_enc_layer_type: abs_pos_whisper
89
+ positional_dropout_rate: 0.1
90
+ static_chunk_size: -1
91
+ use_dynamic_chunk: false
92
+ use_dynamic_left_chunk: false
93
+ epoch: 11
94
+ fire_module: link_and_encoder_and_lora
95
+ grad_clip: 5
96
+ init_step: false
97
+ input_dim: 80
98
+ llm_path: Qwen/Qwen2-7B #/home/node54_tmpdata/xlgeng/ckpt/qwen-7B-instruct/qwen2_7b
99
+ log_interval: 10
100
+ lora_alpha: 32
101
+ lora_dropout: 0.1
102
+ lora_rank: 8
103
+ loss_dict:
104
+ acc: 0.0
105
+ loss: 1.4107781417203814e-05
106
+ lrs:
107
+ - 1.3727899783270902e-05
108
+ max_epoch: 100
109
+ model: llmasr
110
+ model_conf:
111
+ ctc_weight: 0
112
+ length_normalized_loss: false
113
+ lsm_weight: 0.1
114
+ model_dir: /home/node54_tmpdata/xlgeng/ckpt/wenet_whisper_finetune_xlgeng/examples/wenetspeech/whisper/exp/qwen2_multi_task_4_6gpus_gxl_adapter/update_data/epoch_1_with_token
115
+ optim: adamw
116
+ optim_conf:
117
+ betas:
118
+ - 0.9
119
+ - 0.99
120
+ eps: 1.0e-06
121
+ lr: 5.0e-05
122
+ weight_decay: 0.01
123
+ output_dim: 151646
124
+ save_interval: 5000
125
+ save_states: model+optimizer
126
+ save_time: 14/01/2025 20:35:10
127
+ scheduler: warmuplr
128
+ scheduler_conf:
129
+ warmup_steps: 8000
130
+ speech_token_num: 4097
131
+ step: 106124
132
+ tag: epoch_11
133
+ tokenizer: huggingface
134
+ tokenizer_conf:
135
+ llm_path: Qwen/Qwen2-7B #/home/node54_tmpdata/xlgeng/ckpt/qwen-7B-instruct/qwen2_7b
136
+ train_engine: deepspeed
137
+ use_amp: true
138
+ use_lora: true
139
+ vocab_size: 151646