repo
stringlengths
2
99
file
stringlengths
13
225
code
stringlengths
0
18.3M
file_length
int64
0
18.3M
avg_line_length
float64
0
1.36M
max_line_length
int64
0
4.26M
extension_type
stringclasses
1 value
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-fixed-oc/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (self.spec_fb_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,473
34.490826
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-fixed-oc/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (self.spec_fb_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,473
34.490826
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-fixed-oc/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (self.spec_fb_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,473
34.490826
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-fixed-oc/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (self.spec_fb_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,473
34.490826
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-fixed-oc/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (self.spec_fb_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,473
34.490826
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-fixed-oc/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (self.spec_fb_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,473
34.490826
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-oc/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,902
33.951648
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-oc/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,902
33.951648
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-oc/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,902
33.951648
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-oc/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,902
33.951648
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-oc/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,902
33.951648
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-oc/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,902
33.951648
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-fixed-sig/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) ## a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfb_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,735
34.085714
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-fixed-sig/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) ## a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfb_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,735
34.085714
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-fixed-sig/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) ## a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfb_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,735
34.085714
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-fixed-sig/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) ## a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfb_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,735
34.085714
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-fixed-sig/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) ## a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfb_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,735
34.085714
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-fixed-sig/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) ## a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfb_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,735
34.085714
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-lstmsum-sig/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,593
33.747619
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-lstmsum-sig/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,593
33.747619
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-lstmsum-sig/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,593
33.747619
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-lstmsum-sig/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,593
33.747619
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-lstmsum-sig/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,593
33.747619
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-lstmsum-sig/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,593
33.747619
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-am/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.am_softmax as nii_amsoftmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output class self.v_out_class = 2 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_angle = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_angle = torch_nn.ModuleList(self.m_angle) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,596
34.207675
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-am/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.am_softmax as nii_amsoftmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output class self.v_out_class = 2 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_angle = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_angle = torch_nn.ModuleList(self.m_angle) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,596
34.207675
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-am/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.am_softmax as nii_amsoftmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output class self.v_out_class = 2 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_angle = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_angle = torch_nn.ModuleList(self.m_angle) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,596
34.207675
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-am/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.am_softmax as nii_amsoftmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output class self.v_out_class = 2 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_angle = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_angle = torch_nn.ModuleList(self.m_angle) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,596
34.207675
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-am/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.am_softmax as nii_amsoftmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output class self.v_out_class = 2 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_angle = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_angle = torch_nn.ModuleList(self.m_angle) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,596
34.207675
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-am/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.am_softmax as nii_amsoftmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output class self.v_out_class = 2 #### # create network #### self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_angle = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_angle = torch_nn.ModuleList(self.m_angle) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_emb = m_output(torch.flatten(hidden_features, 1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,596
34.207675
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-oc/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) a_softmax_act = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [a_softmax_act, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,372
34.753731
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-oc/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) a_softmax_act = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [a_softmax_act, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,372
34.753731
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-oc/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) a_softmax_act = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [a_softmax_act, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,372
34.753731
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-oc/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) a_softmax_act = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [a_softmax_act, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,372
34.753731
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-oc/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) a_softmax_act = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [a_softmax_act, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,372
34.753731
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-oc/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.oc_softmax as nii_oc_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 256 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 512), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(256, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_a_softmax.append( nii_oc_softmax.OCAngleLayer(self.v_emd_dim) ) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # compute a-softmax output for each feature configuration batch_size = feature_vec.shape[0] // self.v_submodels x_cos_val = torch.zeros( [feature_vec.shape[0], self.v_out_class], dtype=feature_vec.dtype, device=feature_vec.device) x_phi_val = torch.zeros_like(x_cos_val) for idx in range(self.v_submodels): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp1, tmp2 = self.m_a_softmax[idx](feature_vec[s_idx:e_idx], inference) x_cos_val[s_idx:e_idx] = tmp1 x_phi_val[s_idx:e_idx] = tmp2 if inference: return x_cos_val else: return [x_cos_val, x_phi_val] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) a_softmax_act = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device).long() target_vec = target_vec.repeat(self.v_submodels) return [a_softmax_act, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) score = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], score.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_oc_softmax.OCSoftmaxWithLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
14,372
34.753731
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-sig/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,026
33.865429
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-sig/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,026
33.865429
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-sig/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,026
33.865429
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-sig/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,025
33.944186
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-sig/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,026
33.865429
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-lstmsum-sig/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32), nii_nn.BLSTMLayer((self.spec_fb_dim//16) * 32, (self.spec_fb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim // 16) * 32, self.v_emd_dim) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,026
33.865429
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-sig/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
13,817
34.430769
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-sig/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
13,817
34.430769
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-sig/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
13,817
34.430769
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-sig/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
13,817
34.430769
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-sig/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
13,817
34.430769
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-fixed-sig/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_modules.a_softmax as nii_a_softmax import core_scripts.data_io.seq_info as nii_seq_tk import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## def protocol_parse(protocol_filepath): data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ## FOR MODEL class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) self.model_debug = False self.validation = False ##### # target data protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # working sampling rate, torchaudio is used to change sampling rate self.m_target_sr = 16000 # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) self.v_flag = 1 # frame shift (number of points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating self.amp_floor = 0.00001 # manual choose the first 600 frames in the data self.v_truncate_lens = [10 * 16 * 750 // x for x in self.frame_hops] # number of sub-models self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 1 # output class self.v_out_class = 1 self.m_transform = [] self.m_output_act = [] self.m_frontend = [] #self.m_a_softmax = [] for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]) ) ) self.m_output_act.append( torch_nn.Sequential( torch_nn.Dropout(0.7), torch_nn.Linear((trunc_len // 16) * (lfcc_dim // 16) * 32, 160), nii_nn.MaxFeatureMap2D(), torch_nn.Linear(80, self.v_emd_dim) ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) #self.m_a_softmax.append( # nii_a_softmax.AngleLayer(self.v_emd_dim, self.v_out_class) #) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_frontend = torch_nn.ModuleList(self.m_frontend) #self.m_a_softmax = torch_nn.ModuleList(self.m_a_softmax) # output # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.zeros([in_dim]) out_m = torch.ones([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features fs: frame shift fl: frame length fn: fft points trunc_len: number of frames per file (by truncating) datalength: original length of data """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # permute to (batch, fft_bin, frame_length) x_sp_amp = x_sp_amp.permute(0, 2, 1) # make sure the buffer is long enough x_sp_amp_buff = torch.zeros( [x_sp_amp.shape[0], x_sp_amp.shape[1], trunc_len], dtype=x_sp_amp.dtype, device=x_sp_amp.device) # for batch of data, handle the padding and trim independently fs = self.frame_hops[idx] for fileidx in range(x_sp_amp.shape[0]): # roughtly this is the number of frames true_frame_num = datalength[fileidx] // fs if true_frame_num > trunc_len: # trim randomly pos = torch.rand([1]) * (true_frame_num-trunc_len) pos = torch.floor(pos[0]).long() tmp = x_sp_amp[fileidx, :, pos:trunc_len+pos] x_sp_amp_buff[fileidx] = tmp else: rep = int(np.ceil(trunc_len / true_frame_num)) tmp = x_sp_amp[fileidx, :, 0:true_frame_num].repeat(1, rep) x_sp_amp_buff[fileidx] = tmp[:, 0:trunc_len] # permute to (batch, frame_length, fft_bin) x_sp_amp = x_sp_amp_buff.permute(0, 2, 1) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_score = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_output) in enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_output_act)): # extract feature (stft spectrogram) x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. flatten and transform through output function tmp_score = m_output(torch.flatten(hidden_features, 1)) output_score[idx * batch_size : (idx+1) * batch_size] = tmp_score return output_score def _compute_score(self, feature_vec, inference=False): """ """ # feature_vec is [batch * submodel, 1] if inference: return feature_vec.squeeze(1) else: return torch.sigmoid(feature_vec).squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): #with torch.no_grad(): # vad_waveform = self.m_vad(x.squeeze(-1)) # vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1])) # if vad_waveform.shape[-1] > 0: # x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1) # else: # pass filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = torch_nn.BCELoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
13,817
34.430769
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-attention-oc/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((lfcc_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32 * 2, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,153
33.519362
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-attention-oc/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((lfcc_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32 * 2, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,153
33.519362
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-attention-oc/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((lfcc_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32 * 2, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,153
33.519362
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-attention-oc/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((lfcc_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32 * 2, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,153
33.519362
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-attention-oc/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((lfcc_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32 * 2, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,153
33.519362
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfcc-lcnn-attention-oc/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.oc_softmax as nii_ocsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 1 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((lfcc_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32 * 2, self.v_emd_dim) ) self.m_angle.append( nii_ocsoftmax.OCAngleLayer(self.v_emd_dim) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return out_score_neg else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_ocsoftmax.OCSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,153
33.519362
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-lstmsum-p2s/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.p2sgrad as nii_p2sgrad import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32), nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_p2sgrad.P2SActivationLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] # compute score through p2sgrad layer out_score = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): tmp_score = m_score(x[idx * batch_size : (idx+1) * batch_size]) out_score[idx * batch_size : (idx+1) * batch_size] = tmp_score if inference: # output_score [:, 1] corresponds to the positive class return out_score[:, 1] else: return out_score def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_p2sgrad.P2SGradLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,081
33.751152
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-lstmsum-p2s/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.p2sgrad as nii_p2sgrad import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32), nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_p2sgrad.P2SActivationLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] # compute score through p2sgrad layer out_score = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): tmp_score = m_score(x[idx * batch_size : (idx+1) * batch_size]) out_score[idx * batch_size : (idx+1) * batch_size] = tmp_score if inference: # output_score [:, 1] corresponds to the positive class return out_score[:, 1] else: return out_score def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_p2sgrad.P2SGradLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,081
33.751152
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-lstmsum-p2s/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.p2sgrad as nii_p2sgrad import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32), nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_p2sgrad.P2SActivationLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] # compute score through p2sgrad layer out_score = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): tmp_score = m_score(x[idx * batch_size : (idx+1) * batch_size]) out_score[idx * batch_size : (idx+1) * batch_size] = tmp_score if inference: # output_score [:, 1] corresponds to the positive class return out_score[:, 1] else: return out_score def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_p2sgrad.P2SGradLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,081
33.751152
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-lstmsum-p2s/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.p2sgrad as nii_p2sgrad import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32), nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_p2sgrad.P2SActivationLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] # compute score through p2sgrad layer out_score = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): tmp_score = m_score(x[idx * batch_size : (idx+1) * batch_size]) out_score[idx * batch_size : (idx+1) * batch_size] = tmp_score if inference: # output_score [:, 1] corresponds to the positive class return out_score[:, 1] else: return out_score def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_p2sgrad.P2SGradLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,081
33.751152
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-lstmsum-p2s/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.p2sgrad as nii_p2sgrad import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32), nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_p2sgrad.P2SActivationLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] # compute score through p2sgrad layer out_score = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): tmp_score = m_score(x[idx * batch_size : (idx+1) * batch_size]) out_score[idx * batch_size : (idx+1) * batch_size] = tmp_score if inference: # output_score [:, 1] corresponds to the positive class return out_score[:, 1] else: return out_score def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_p2sgrad.P2SGradLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,081
33.751152
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/lfb-lcnn-lstmsum-p2s/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.p2sgrad as nii_p2sgrad import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFB dim (base component) self.lfb_dim = [60] self.lfb_with_delta = False # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part on training self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfb_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfb_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfb_with_delta: lfb_dim = lfb_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32), nii_nn.BLSTMLayer((lfb_dim//16) * 32, (lfb_dim//16) * 32) ) ) self.m_output_act.append( torch_nn.Linear((lfb_dim // 16) * 32, self.v_emd_dim) ) self.m_angle.append( nii_p2sgrad.P2SActivationLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFB(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfb_dim[idx], with_energy=False, with_emphasis=True, with_delta=self.lfb_with_delta) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_angle = torch_nn.ModuleList(self.m_angle) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output((hidden_features_lstm + hidden_features).sum(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models batch_size = x.shape[0] # compute score through p2sgrad layer out_score = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): tmp_score = m_score(x[idx * batch_size : (idx+1) * batch_size]) out_score[idx * batch_size : (idx+1) * batch_size] = tmp_score if inference: # output_score [:, 1] corresponds to the positive class return out_score[:, 1] else: return out_score def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=scores.dtype) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec, True] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_p2sgrad.P2SGradLoss() def compute(self, outputs, target): """ """ loss = self.m_loss(outputs[0], outputs[1]) return loss if __name__ == "__main__": print("Definition of model")
15,081
33.751152
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-attention-am/05/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # spectrogram dim (base component) self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((self.spec_fb_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim//16)*32*2, self.v_emd_dim) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,576
33.615556
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-attention-am/04/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # spectrogram dim (base component) self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((self.spec_fb_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim//16)*32*2, self.v_emd_dim) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,576
33.615556
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-attention-am/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # spectrogram dim (base component) self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((self.spec_fb_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim//16)*32*2, self.v_emd_dim) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,576
33.615556
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-attention-am/06/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # spectrogram dim (base component) self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((self.spec_fb_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim//16)*32*2, self.v_emd_dim) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,576
33.615556
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-attention-am/03/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # spectrogram dim (base component) self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((self.spec_fb_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim//16)*32*2, self.v_emd_dim) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,576
33.615556
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/03-asvspoof-mega/spec2-lcnn-attention-am/02/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## util ############## def protocol_parse(protocol_filepath): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial input: ----- protocol_filepath: string, path to the protocol file for convenience, I put train/dev/eval trials into a single protocol file output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} temp_buffer = np.loadtxt(protocol_filepath, dtype='str') for row in temp_buffer: if row[-1] == 'bonafide': data_buffer[row[1]] = 1 else: data_buffer[row[1]] = 0 return data_buffer ############## ## FOR MODEL ############## class TrainableLinearFb(nii_nn.LinearInitialized): """Linear layer initialized with linear filter bank """ def __init__(self, fn, sr, filter_num): super(TrainableLinearFb, self).__init__( nii_front_end.linear_fb(fn, sr, filter_num)) return def forward(self, x): return torch.log10( torch.pow(super(TrainableLinearFb, self).forward(x), 2) + torch.finfo(torch.float32).eps) class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) # self.model_debug = False # self.flag_validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_file = prj_conf.optional_argument[0] self.protocol_parser = protocol_parse(protocol_file) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # spectrogram dim (base component) self.spec_with_delta = False self.spec_fb_dim = 60 # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors self.v_emd_dim = 64 # output classes self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # pooling layer self.m_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # final part for output layer self.m_angle = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n) in enumerate(zip( self.v_truncate_lens, self.fft_n)): fft_n_bins = fft_n // 2 + 1 self.m_transform.append( torch_nn.Sequential( TrainableLinearFb(fft_n,self.m_target_sr,self.spec_fb_dim), torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_pooling.append( nii_nn.SelfWeightedPooling((self.spec_fb_dim // 16) * 32) ) self.m_output_act.append( torch_nn.Linear((self.spec_fb_dim//16)*32*2, self.v_emd_dim) ) self.m_angle.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.Spectrogram(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_pooling = torch_nn.ModuleList(self.m_pooling) self.m_angle = torch_nn.ModuleList(self.m_angle) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # batch_size batch_size = x.shape[0] // self.v_submodels # buffer to store output scores from sub-models output_emb = torch.zeros([x.shape[0], self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling hidden_features = m_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_output(hidden_features) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_score(self, x, inference=False): """ """ # number of sub models * batch_size batch_size = x.shape[0] // self.v_submodels # buffer to save the scores # for non-target classes out_score_neg = torch.zeros( [x.shape[0], self.v_out_class], device=x.device, dtype=x.dtype) # for target classes out_score_pos = torch.zeros_like(out_score_neg) # compute scores for each sub-models for idx, m_score in enumerate(self.m_angle): s_idx = idx * batch_size e_idx = idx * batch_size + batch_size tmp_score = m_score(x[s_idx:e_idx], inference) out_score_neg[s_idx:e_idx] = tmp_score[0] out_score_pos[s_idx:e_idx] = tmp_score[1] if inference: return torch_nn_func.softmax(out_score_neg, dim=1)[:, 1] else: return out_score_neg, out_score_pos def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] if self.training: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec) # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) return [scores, target_vec] else: feature_vec = self._compute_embedding(x, datalength) scores = self._compute_score(feature_vec, True) target = self._get_target(filenames) print("Output, %s, %d, %f" % (filenames[0], target[0], scores.mean())) # don't write output score as a single file return None class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ self.m_loss = nii_amsoftmax.AMSoftmaxWithLoss() def compute(self, input_data, target): """loss = compute(input_data, target_data) Note: 1. input_data will be the output from Model.forward() input_data will be a tuple of [scores, target_vec] 2. we will not use target given by the system script we will use the target_vec in input_data[1] """ loss = self.m_loss(input_data[0], input_data[1]) return loss if __name__ == "__main__": print("Definition of model")
15,576
33.615556
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/main.py
#!/usr/bin/env python """ main.py for project-NN-pytorch/projects The training/inference process wrapper. Dataset API is replaced with NII_MergeDataSetLoader. It is more convenient to train model on corpora stored in different directories. Requires model.py and config.py (config_merge_datasets.py) Usage: $: python main.py [options] """ from __future__ import absolute_import import os import sys import torch import importlib import core_scripts.other_tools.display as nii_warn import core_scripts.data_io.default_data_io as nii_default_dset import core_scripts.data_io.customize_dataset as nii_dset import core_scripts.data_io.conf as nii_dconf import core_scripts.other_tools.list_tools as nii_list_tool import core_scripts.config_parse.config_parse as nii_config_parse import core_scripts.config_parse.arg_parse as nii_arg_parse import core_scripts.op_manager.op_manager as nii_op_wrapper import core_scripts.nn_manager.nn_manager as nii_nn_wrapper import core_scripts.startup_config as nii_startup __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" def main(): """ main(): the default wrapper for training and inference process Please prepare config.py and model.py """ # arguments initialization args = nii_arg_parse.f_args_parsed() # nii_warn.f_print_w_date("Start program", level='h') nii_warn.f_print("Load module: %s" % (args.module_config)) nii_warn.f_print("Load module: %s" % (args.module_model)) prj_conf = importlib.import_module(args.module_config) prj_model = importlib.import_module(args.module_model) # initialization nii_startup.set_random_seed(args.seed, args) use_cuda = not args.no_cuda and torch.cuda.is_available() device = torch.device("cuda" if use_cuda else "cpu") # prepare data io if not args.inference: params = {'batch_size': args.batch_size, 'shuffle': args.shuffle, 'num_workers': args.num_workers, 'sampler': args.sampler} in_trans_fns = prj_conf.input_trans_fns \ if hasattr(prj_conf, 'input_trans_fns') else None out_trans_fns = prj_conf.output_trans_fns \ if hasattr(prj_conf, 'output_trans_fns') else None # Load file list and create data loader trn_lst = prj_conf.trn_list trn_set = nii_dset.NII_MergeDataSetLoader( prj_conf.trn_set_name, \ trn_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', params = params, truncate_seq = prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = True, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) if prj_conf.val_list is not None: val_lst = prj_conf.val_list val_set = nii_dset.NII_MergeDataSetLoader( prj_conf.val_set_name, val_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', \ params = params, truncate_seq= prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) else: val_set = None # initialize the model and loss function model = prj_model.Model(trn_set.get_in_dim(), \ trn_set.get_out_dim(), \ args, prj_conf, trn_set.get_data_mean_std()) loss_wrapper = prj_model.Loss(args) # initialize the optimizer optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args) # if necessary, resume training if args.trained_model == "": checkpoint = None else: checkpoint = torch.load(args.trained_model) # start training nii_nn_wrapper.f_train_wrapper(args, model, loss_wrapper, device, optimizer_wrapper, trn_set, val_set, checkpoint) # done for traing else: # for inference # default, no truncating, no shuffling params = {'batch_size': args.batch_size, 'shuffle': False, 'num_workers': args.num_workers, 'sampler': args.sampler} in_trans_fns = prj_conf.test_input_trans_fns \ if hasattr(prj_conf, 'test_input_trans_fns') else None out_trans_fns = prj_conf.test_output_trans_fns \ if hasattr(prj_conf, 'test_output_trans_fns') else None if type(prj_conf.test_list) is list: t_lst = prj_conf.test_list else: t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list) test_set = nii_dset.NII_MergeDataSetLoader( prj_conf.test_set_name, \ t_lst, \ prj_conf.test_input_dirs, prj_conf.input_exts, prj_conf.input_dims, prj_conf.input_reso, prj_conf.input_norm, prj_conf.test_output_dirs, prj_conf.output_exts, prj_conf.output_dims, prj_conf.output_reso, prj_conf.output_norm, './', params = params, truncate_seq= None, min_seq_len = None, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) # initialize model model = prj_model.Model(test_set.get_in_dim(), \ test_set.get_out_dim(), \ args, prj_conf) if args.trained_model == "": print("No model is loaded by ---trained-model for inference") print("By default, load %s%s" % (args.save_trained_name, args.save_model_ext)) checkpoint = torch.load("%s%s" % (args.save_trained_name, args.save_model_ext)) else: checkpoint = torch.load(args.trained_model) # do inference and output data nii_nn_wrapper.f_inference_wrapper(args, model, device, \ test_set, checkpoint) # done return if __name__ == "__main__": main()
7,862
35.915493
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/config_train_asvspoof2019_bc10.py
#!/usr/bin/env python """ config.py for project-NN-pytorch/projects Usage: For training, change Configuration for training stage For inference, change Configuration for inference stage """ import os __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2022, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['asvspoof2019_trn', 'bc2019_trn'] val_set_name = ['asvspoof2019_val', 'bc2019_val'] # for convenience tmp1 = os.path.dirname(__file__) + '/../../../DATA/asvspoof2019_LA' tmp2 = os.path.dirname(__file__) + '/../../../DATA/bc_release' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp1 + '/scp/train.lst', tmp2 + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp1 + '/scp/val.lst', tmp2 + '/scp/val.lst'] # Directories for input features # input_dirs = [[path_of_feature_1_trainset_1, path_of_feature_2_trainset_1.. ] # [path_of_feature_1_trainset_2, path_of_feature_2_trainset_2..]] # len(input_dirs) should = len(trn_list) = len(val_list) # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp1 + '/train_dev'], [tmp2 + '/train_dev']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] # len(input_dims) should be len(input_dirs[0]) input_dims = [1] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.wav'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [1] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [False] # Similar configurations for output features output_dirs = [[] for x in input_dirs] output_dims = [1] output_exts = ['.bin'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = 64000 # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = 8000 # Optional argument # Just a buffer for convenience # It can contain anything # Here we use it to hold the path to the protocol.txt # They will be loaded by model.py optional_argument = [tmp1 + '/protocol.txt', tmp2 + '/protocol.txt'] #import augment #input_trans_fns = [[augment.wav_aug]] #output_trans_fns = [[]] ######################################################### ## Configuration for inference stage (place holder) ######################################################### # Please use config_test_*.py inference # This part is just a place holder test_set_name = trn_set_name + val_set_name # List of test set data # for convenience, you may directly load test_set list here test_list = trn_list + val_list # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = input_dirs * 2 # Directories for output features, which are [] test_output_dirs = [[]] * 2
3,891
32.843478
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/config_train_asvspoof2019_esp.py
#!/usr/bin/env python """ config.py for project-NN-pytorch/projects Usage: For training, change Configuration for training stage For inference, change Configuration for inference stage """ import os __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2022, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['asvspoof2019_trn', 'espnet_trn'] val_set_name = ['asvspoof2019_val', 'espnet_val'] # for convenience tmp1 = os.path.dirname(__file__) + '/../../../DATA/asvspoof2019_LA' tmp2 = os.path.dirname(__file__) + '/../../../DATA/espnet_release' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp1 + '/scp/train.lst', tmp2 + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp1 + '/scp/val.lst', tmp2 + '/scp/val.lst'] # Directories for input features # input_dirs = [[path_of_feature_1_trainset_1, path_of_feature_2_trainset_1.. ] # [path_of_feature_1_trainset_2, path_of_feature_2_trainset_2..]] # len(input_dirs) should = len(trn_list) = len(val_list) # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp1 + '/train_dev'], [tmp2 + '/train_dev']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] # len(input_dims) should be len(input_dirs[0]) input_dims = [1] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.wav'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [1] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [False] # Similar configurations for output features output_dirs = [[] for x in input_dirs] output_dims = [1] output_exts = ['.bin'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = 64000 # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = 8000 # Optional argument # Just a buffer for convenience # It can contain anything # Here we use it to hold the path to the protocol.txt # They will be loaded by model.py optional_argument = [tmp1 + '/protocol.txt', tmp2 + '/protocol.txt'] #import augment #input_trans_fns = [[augment.wav_aug]] #output_trans_fns = [[]] ######################################################### ## Configuration for inference stage (place holder) ######################################################### # Please use config_test_*.py inference # This part is just a place holder test_set_name = trn_set_name + val_set_name # List of test set data # for convenience, you may directly load test_set list here test_list = trn_list + val_list # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = input_dirs * 2 # Directories for output features, which are [] test_output_dirs = [[]] * 2
3,895
32.878261
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/config_test_asvspoof2019.py
#!/usr/bin/env python """ config.py for project-NN-pytorch/projects Usage: For training, change Configuration for training stage For inference, change Configuration for inference stage """ import os __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2022, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Not necessary to change this part # Parameters like input_dims, input_exts will be used for inference too # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['asvspoof2019_trn'] val_set_name = ['asvspoof2019_val'] # for convenience tmp = '' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp + '/scp/val.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp + '/train_dev']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [1] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.wav'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [1] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [False] # Similar configurations for output features output_dirs = [[] for x in input_dirs] output_dims = [1] output_exts = ['.bin'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = None # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = None #import augment #input_trans_fns = [[augment.wav_aug]] #output_trans_fns = [[]] ######################################################### ## Configuration for inference stage ######################################################### # similar options to training stage # path to the test set directory tmp = os.path.dirname(__file__) + '/../../../DATA/asvspoof2019_LA' test_set_name = ['asvspoof2019_test'] # List of test set data # for convenience, you may directly load test_set list here test_list = [tmp + '/scp/test.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = [[tmp + '/eval']] # Directories for output features, which are [] test_output_dirs = [[]] # Optional argument # Just a buffer for convenience # It can contain anything # Here we use it to hold the path to the protocol.txt # They will be loaded by model.py optional_argument = [tmp + '/protocol.txt']
3,592
30.243478
75
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/config_train_asvspoof2019.py
#!/usr/bin/env python """ config.py for project-NN-pytorch/projects Usage: For training, change Configuration for training stage For inference, change Configuration for inference stage """ import os __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2022, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['asvspoof2019_trn'] val_set_name = ['asvspoof2019_val'] # for convenience tmp = os.path.dirname(__file__) + '/../../../DATA/asvspoof2019_LA' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp + '/scp/val.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp + '/train_dev']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [1] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.wav'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [1] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [False] # Similar configurations for output features output_dirs = [[] for x in input_dirs] output_dims = [1] output_exts = ['.bin'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = 64000 # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = 8000 # Optional argument # Just a buffer for convenience # It can contain anything # Here we use it to hold the path to the protocol.txt # They will be loaded by model.py optional_argument = [tmp + '/protocol.txt'] #import augment #input_trans_fns = [[augment.wav_aug]] #output_trans_fns = [[]] ######################################################### ## Configuration for inference stage (place holder) ######################################################### # Please use config_test_*.py inference # This part is just a place holder test_set_name = trn_set_name + val_set_name # List of test set data # for convenience, you may directly load test_set list here test_list = trn_list + val_list # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = input_dirs * 2 # Directories for output features, which are [] test_output_dirs = [[]] * 2
3,496
30.504505
75
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/config_test_vcc.py
#!/usr/bin/env python """ config.py for project-NN-pytorch/projects Usage: For training, change Configuration for training stage For inference, change Configuration for inference stage """ import os __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2022, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Not necessary to change this part # Parameters like input_dims, input_exts will be used for inference too # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['asvspoof2019_trn'] val_set_name = ['asvspoof2019_val'] # for convenience tmp = '' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp + '/scp/val.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp + '/train_dev']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [1] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.wav'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [1] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [False] # Similar configurations for output features output_dirs = [[] for x in input_dirs] output_dims = [1] output_exts = ['.bin'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = None # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = None #import augment #input_trans_fns = [[augment.wav_aug]] #output_trans_fns = [[]] ######################################################### ## Configuration for inference stage ######################################################### # similar options to training stage # path to the test set directory tmp2 = os.path.dirname(__file__) + '/../../../DATA/vcc_release' test_set_name = ['vcc_test'] # List of test set data # for convenience, you may directly load test_set list here test_list = [tmp2 + '/scp/test.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = [[tmp2 + '/eval']] # Directories for output features, which are [] test_output_dirs = [[]] # Optional argument # Just a buffer for convenience # It can contain anything # Here we use it to hold the path to the protocol.txt # They will be loaded by model.py optional_argument = [tmp2 + '/protocol.txt']
3,584
29.905172
75
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/Softmax-maxprob/config_train_asvspoof2019/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torchaudio import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import sandbox.eval_asvspoof as nii_asvspoof __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" g_attack_map = {'-': 1, 'A01': 1, 'A02': 1, 'A03': 1, 'A04': 1, 'A05': 1, 'A06': 1, 'A07': 1, 'A08': 1, 'A09': 1, 'A10': 1, 'A11': 1, 'A12': 1, 'A13': 1, 'A14': 1, 'A15': 1, 'A16': 1, 'A17': 1, 'A18': 1, 'A19': 1} def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial The format is: SPEAKER TRIAL_NAME - SPOOF_TYPE TAG LA_0031 LA_E_5932896 - A13 spoof LA_0030 LA_E_5849185 - - bonafide ... input: ----- protocol_filepath: string, path to the protocol file target_row: int, default -1, use line[-1] as the target label output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} if type(protocol_filepaths) is str: tmp = [protocol_filepaths] else: tmp = protocol_filepaths for protocol_filepath in tmp: with open(protocol_filepath, 'r') as file_ptr: for line in file_ptr: line = line.rstrip('\n') cols = line.split(sep) try: data_buffer[cols[1]] = g_attack_map[cols[target_row]] except KeyError: data_buffer[cols[1]] = False return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std( in_dim,out_dim, args, prj_conf, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_f = prj_conf.optional_argument self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f) self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map, ' ', -2) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = None self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # confidence predictor self.m_conf = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) if self.v_emd_dim is None: self.v_emd_dim = (lfcc_dim // 16) * 32 else: assert self.v_emd_dim == (lfcc_dim//16) * 32, "v_emd_dim error" self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output self.m_loss = torch_nn.CrossEntropyLoss() self.m_temp = 1 self.m_lambda = 0. self.m_e_m_in = -25.0 self.m_e_m_out = -7.0 # done return def prepare_mean_std(self, in_dim, out_dim, args, prj_conf, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = (hidden_features_lstm + hidden_features).mean(1) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_logit(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models output_act = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_output_act): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb) # feature_vec is [batch * submodel, output-class] return output_act def _compute_score(self, logits): """ """ # [batch * submodel, output-class], logits # [:, 1] denotes being bonafide if logits.shape[1] == 2: return logits[:, 1] - logits[:, 0] else: return logits[:, -1] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def _clamp_prob(self, input_prob, clamp_val=1e-12): return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val) def _get_in_out_indx(self, filenames): in_indx = [] out_indx = [] for x, y in enumerate(filenames): if self.in_out_parser[y]: in_indx.append(x) else: out_indx.append(x) return np.array(in_indx), np.array(out_indx) def _energy(self, logits): """ """ # - T \log \sum_y \exp (logits[x, y] / T) eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1) return eng def _loss(self, logits, targets, energy, in_indx, out_indx): """ """ # loss over cross-entropy on in-dist. data if len(in_indx): loss = self.m_loss(logits[in_indx], targets[in_indx]) else: loss = 0 # loss on energy of in-dist.data if len(in_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean() # loss on energy of out-dist. data if len(out_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean() return loss def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] # too short sentences if not self.training and x.shape[1] < 3000: targets = self._get_target(filenames) for filename, target in zip(filenames, targets): print("Output, %s, %d, %f, %f" % ( filename, target, 0.0, 0.0)) return None feature_vec = self._compute_embedding(x, datalength) logits = self._compute_logit(feature_vec) energy = self._energy(logits) in_indx, out_indx = self._get_in_out_indx(filenames) if self.training: # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) # loss loss = self._loss(logits, target_vec, energy, in_indx, out_indx) return loss else: scores = self._compute_score(logits) targets = self._get_target(filenames) # for baseline, get the conf_score through softmax conf_scores_new = torch_nn_func.softmax(logits, dim=1) conf_scores_new, _ = conf_scores_new.max(dim=1) for filename, target, score, conf in \ zip(filenames, targets, scores, conf_scores_new): print("Output, %s, %d, %f, %f" % ( filename, target, score.item(), conf.item())) # don't write output score as a single file return None def get_embedding(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] feature_vec = self._compute_embedding(x, datalength) return feature_vec class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ def compute(self, outputs, target): """ """ return outputs if __name__ == "__main__": print("Definition of model")
18,568
34.302281
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/AM-softmax-energy/config_train_asvspoof2019/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torchaudio import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import sandbox.eval_asvspoof as nii_asvspoof __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" g_attack_map = {'-': 1, 'A01': 1, 'A02': 1, 'A03': 1, 'A04': 1, 'A05': 1, 'A06': 1, 'A07': 1, 'A08': 1, 'A09': 1, 'A10': 1, 'A11': 1, 'A12': 1, 'A13': 1, 'A14': 1, 'A15': 1, 'A16': 1, 'A17': 1, 'A18': 1, 'A19': 1} def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial The format is: SPEAKER TRIAL_NAME - SPOOF_TYPE TAG LA_0031 LA_E_5932896 - A13 spoof LA_0030 LA_E_5849185 - - bonafide ... input: ----- protocol_filepath: string, path to the protocol file target_row: int, default -1, use line[-1] as the target label output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} if type(protocol_filepaths) is str: tmp = [protocol_filepaths] else: tmp = protocol_filepaths for protocol_filepath in tmp: with open(protocol_filepath, 'r') as file_ptr: for line in file_ptr: line = line.rstrip('\n') cols = line.split(sep) try: data_buffer[cols[1]] = g_attack_map[cols[target_row]] except KeyError: data_buffer[cols[1]] = False return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std( in_dim,out_dim, args, prj_conf, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_f = prj_conf.optional_argument self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f) self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map, ' ', -2) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 128 self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] self.m_after_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # confidence predictor self.m_conf = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_after_pooling.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_output_act.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_after_pooling = torch_nn.ModuleList(self.m_after_pooling) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output self.m_loss = torch_nn.CrossEntropyLoss() self.m_temp = 1 self.m_lambda = 0.0 self.m_e_m_in = -25.0 self.m_e_m_out = -7.0 # done return def prepare_mean_std(self, in_dim, out_dim, args, prj_conf, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_a_pool) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_after_pooling)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_a_pool((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_logit(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models output_act_pos = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) output_act_neg = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_output_act): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] tmp_neg, tmp_pos = m_output(tmp_emb) output_act_pos[idx*batch_size : (idx+1)*batch_size] = tmp_pos output_act_neg[idx*batch_size : (idx+1)*batch_size] = tmp_neg # feature_vec is [batch * submodel, output-class] return output_act_pos, output_act_neg def _compose_logit(self, logits, targets): """ """ logits_pos = logits[0] logits_neg = logits[1] with torch.no_grad(): index = torch.zeros_like(logits_pos).bool() index.scatter_(1, targets.data.view(-1, 1), 1) logits_new = logits_neg * 1.0 logits_new[index] = logits_pos[index] return logits_new def _compute_score(self, logits): """ """ return logits[1][:, -1] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def _clamp_prob(self, input_prob, clamp_val=1e-12): return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val) def _get_in_out_indx(self, filenames): in_indx = [] out_indx = [] for x, y in enumerate(filenames): if self.in_out_parser[y]: in_indx.append(x) else: out_indx.append(x) return np.array(in_indx), np.array(out_indx) def _energy(self, logits): """ """ # - T \log \sum_y \exp (logits[x, y] / T) eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1) return eng def _loss(self, logits, targets, energy, in_indx, out_indx): """ """ # loss over cross-entropy on in-dist. data if len(in_indx): loss = self.m_loss(logits[in_indx], targets[in_indx]) else: loss = 0 # loss on energy of in-dist.data if len(in_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean() # loss on energy of out-dist. data if len(out_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean() return loss def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] # too short sentences if not self.training and x.shape[1] < 3000: targets = self._get_target(filenames) for filename, target in zip(filenames, targets): print("Output, %s, %d, %f, %f" % ( filename, target, 0.0, 0.0)) return None feature_vec = self._compute_embedding(x, datalength) logits = self._compute_logit(feature_vec) in_indx, out_indx = self._get_in_out_indx(filenames) if self.training: # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) composed_logit = self._compose_logit(logits, target_vec) energy = self._energy(composed_logit) # loss loss = self._loss(composed_logit, target_vec, energy, in_indx, out_indx) return loss else: scores = self._compute_score(logits) targets = self._get_target(filenames) energy = self._energy(logits[1]) # for baseline, get the conf_score through softmax #conf_scores_new = torch_nn_func.softmax(logits[1], dim=1) #conf_scores_new, _ = conf_scores_new.max(dim=1) for filename, target, score, energytmp in \ zip(filenames, targets, scores, energy): print("Output, %s, %d, %f, %f" % ( filename, target, score.item(), -energytmp.item())) # don't write output score as a single file return None def get_embedding(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] feature_vec = self._compute_embedding(x, datalength) return feature_vec class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ def compute(self, outputs, target): """ """ return outputs if __name__ == "__main__": print("Definition of model")
19,396
34.460695
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/AM-softmax-conf/config_train_asvspoof2019/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torchaudio import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import sandbox.eval_asvspoof as nii_asvspoof __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std( in_dim,out_dim, args, prj_conf, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_f = prj_conf.optional_argument self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 128 self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] self.m_after_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # confidence predictor self.m_conf = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_after_pooling.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_output_act.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_conf.append( torch_nn.Sequential( torch_nn.Linear(self.v_emd_dim, 128), torch_nn.Tanh(), torch_nn.Linear(128, 1), torch_nn.Sigmoid() ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) self.m_after_pooling = torch_nn.ModuleList(self.m_after_pooling) self.m_conf = torch_nn.ModuleList(self.m_conf) # output self.m_loss = torch_nn.NLLLoss() self.m_lambda = torch_nn.Parameter(torch.tensor([0.1]), requires_grad=False) self.m_budget = 0.5 # done return def prepare_mean_std(self, in_dim, out_dim, args, prj_conf, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_a_pool) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_after_pooling)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_a_pool((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_logit(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models output_act_pos = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) output_act_neg = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_output_act): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] tmp_neg, tmp_pos = m_output(tmp_emb) output_act_pos[idx*batch_size : (idx+1)*batch_size] = tmp_pos output_act_neg[idx*batch_size : (idx+1)*batch_size] = tmp_neg # feature_vec is [batch * submodel, output-class] return output_act_pos, output_act_neg def _compose_logit(self, logits, targets): """ """ logits_pos = logits[0] logits_neg = logits[1] with torch.no_grad(): index = torch.zeros_like(logits_pos).bool() index.scatter_(1, targets.data.view(-1, 1), 1) logits_new = logits_neg * 1.0 logits_new[index] = logits_pos[index] return logits_new def _compute_score(self, logits): """ """ # bonafide binary from negative logits return logits[1][:, -1] def _compute_conf(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models conf_score = torch.zeros( [batch_size * self.v_submodels, 1], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_conf): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] tmp_emb = torch_nn_func.normalize(tmp_emb, p=2, dim=1) conf_score[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb) return conf_score.squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def _clamp_prob(self, input_prob, clamp_val=1e-12): return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val) def _loss(self, logits, targets, conf_scores): """ """ with torch.no_grad(): index = torch.zeros_like(logits) index.scatter_(1, targets.data.view(-1, 1), 1) # clamp the probablity and conf scores prob = self._clamp_prob(torch_nn_func.softmax(logits, dim=1)) conf_tmp = self._clamp_prob(conf_scores) # mixed log probablity log_prob = torch.log( prob * conf_tmp.view(-1, 1) + (1 - conf_tmp.view(-1, 1)) * index) loss = self.m_loss(log_prob, targets) loss2 = -torch.log(conf_scores).mean() loss = loss + self.m_lambda * loss2 if self.m_budget > loss2: self.m_lambda.data = self.m_lambda / 1.01 else: self.m_lambda.data = self.m_lambda / 0.99 return loss.mean() def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] # too short sentences if not self.training and x.shape[1] < 3000: targets = self._get_target(filenames) for filename, target in zip(filenames, targets): print("Output, %s, %d, %f, %f" % ( filename, target, 0.0, 0.0)) return None # for training, do balanced batch if self.training: target = self._get_target(filenames) bona_indx = np.argwhere(np.array(target) > 0)[:, 0] spoof_indx = np.argwhere(np.array(target) == 0)[:, 0] num = np.min([len(bona_indx), len(spoof_indx)]) trial_indx = np.concatenate([bona_indx[0:num], spoof_indx[0:num]]) if len(trial_indx) == 0: x_ = x flag_no_data = True else: filenames = [filenames[x] for x in trial_indx] datalength = [datalength[x] for x in trial_indx] x_ = x[trial_indx] flag_no_data = False else: x_ = x feature_vec = self._compute_embedding(x_, datalength) logits = self._compute_logit(feature_vec) conf_scores = self._compute_conf(feature_vec) if self.training: # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) # randomly set half of the conf data to 1 b = torch.bernoulli(torch.zeros_like(conf_scores).uniform_(0, 1)) conf_scores = conf_scores * b + (1 - b) composed_logit = self._compose_logit(logits, target_vec) loss = self._loss(composed_logit, target_vec, conf_scores) if flag_no_data: return loss * 0 else: return loss else: scores = self._compute_score(logits) targets = self._get_target(filenames) for filename, target, score, conf in \ zip(filenames, targets, scores, conf_scores): print("Output, %s, %d, %f, %f" % ( filename, target, score.item(), conf.item())) # don't write output score as a single file return None def get_embedding(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] feature_vec = self._compute_embedding(x, datalength) return feature_vec class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ def compute(self, outputs, target): """ """ return outputs if __name__ == "__main__": print("Definition of model")
19,307
34.821892
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/AM-softmax-maxprob/config_train_asvspoof2019/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torchaudio import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import core_modules.am_softmax as nii_amsoftmax import sandbox.eval_asvspoof as nii_asvspoof __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" g_attack_map = {'-': 1, 'A01': 1, 'A02': 1, 'A03': 1, 'A04': 1, 'A05': 1, 'A06': 1, 'A07': 1, 'A08': 1, 'A09': 1, 'A10': 1, 'A11': 1, 'A12': 1, 'A13': 1, 'A14': 1, 'A15': 1, 'A16': 1, 'A17': 1, 'A18': 1, 'A19': 1} def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial The format is: SPEAKER TRIAL_NAME - SPOOF_TYPE TAG LA_0031 LA_E_5932896 - A13 spoof LA_0030 LA_E_5849185 - - bonafide ... input: ----- protocol_filepath: string, path to the protocol file target_row: int, default -1, use line[-1] as the target label output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} if type(protocol_filepaths) is str: tmp = [protocol_filepaths] else: tmp = protocol_filepaths for protocol_filepath in tmp: with open(protocol_filepath, 'r') as file_ptr: for line in file_ptr: line = line.rstrip('\n') cols = line.split(sep) try: data_buffer[cols[1]] = g_attack_map[cols[target_row]] except KeyError: data_buffer[cols[1]] = False return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std( in_dim,out_dim, args, prj_conf, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_f = prj_conf.optional_argument self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f) self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map, ' ', -2) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = 128 self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] self.m_after_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # confidence predictor self.m_conf = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) self.m_after_pooling.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim) ) self.m_output_act.append( nii_amsoftmax.AMAngleLayer(self.v_emd_dim, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_after_pooling = torch_nn.ModuleList(self.m_after_pooling) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output self.m_loss = torch_nn.CrossEntropyLoss() self.m_temp = 1 self.m_lambda = 0.0 self.m_e_m_in = -25.0 self.m_e_m_out = -7.0 # done return def prepare_mean_std(self, in_dim, out_dim, args, prj_conf, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_a_pool) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_after_pooling)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = m_a_pool((hidden_features_lstm + hidden_features).mean(1)) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_logit(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models output_act_pos = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) output_act_neg = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_output_act): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] tmp_neg, tmp_pos = m_output(tmp_emb) output_act_pos[idx*batch_size : (idx+1)*batch_size] = tmp_pos output_act_neg[idx*batch_size : (idx+1)*batch_size] = tmp_neg # feature_vec is [batch * submodel, output-class] return output_act_pos, output_act_neg def _compose_logit(self, logits, targets): """ """ logits_pos = logits[0] logits_neg = logits[1] with torch.no_grad(): index = torch.zeros_like(logits_pos).bool() index.scatter_(1, targets.data.view(-1, 1), 1) logits_new = logits_neg * 1.0 logits_new[index] = logits_pos[index] return logits_new def _compute_score(self, logits): """ """ return logits[1][:, -1] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def _clamp_prob(self, input_prob, clamp_val=1e-12): return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val) def _get_in_out_indx(self, filenames): in_indx = [] out_indx = [] for x, y in enumerate(filenames): if self.in_out_parser[y]: in_indx.append(x) else: out_indx.append(x) return np.array(in_indx), np.array(out_indx) def _energy(self, logits): """ """ # - T \log \sum_y \exp (logits[x, y] / T) eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1) return eng def _loss(self, logits, targets, energy, in_indx, out_indx): """ """ # loss over cross-entropy on in-dist. data if len(in_indx): loss = self.m_loss(logits[in_indx], targets[in_indx]) else: loss = 0 # loss on energy of in-dist.data if len(in_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean() # loss on energy of out-dist. data if len(out_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean() return loss def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] # too short sentences if not self.training and x.shape[1] < 3000: targets = self._get_target(filenames) for filename, target in zip(filenames, targets): print("Output, %s, %d, %f, %f" % ( filename, target, 0.0, 0.0)) return None feature_vec = self._compute_embedding(x, datalength) logits = self._compute_logit(feature_vec) in_indx, out_indx = self._get_in_out_indx(filenames) if self.training: # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) composed_logit = self._compose_logit(logits, target_vec) energy = self._energy(composed_logit) # loss loss = self._loss(composed_logit, target_vec, energy, in_indx, out_indx) return loss else: scores = self._compute_score(logits) targets = self._get_target(filenames) energy = self._energy(logits[1]) # for baseline, get the conf_score through softmax conf_scores_new = torch_nn_func.softmax(logits[1], dim=1) conf_scores_new, _ = conf_scores_new.max(dim=1) for filename, target, score, conf in \ zip(filenames, targets, scores, conf_scores_new): print("Output, %s, %d, %f, %f" % ( filename, target, score.item(), conf.item())) # don't write output score as a single file return None def get_embedding(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] feature_vec = self._compute_embedding(x, datalength) return feature_vec class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ def compute(self, outputs, target): """ """ return outputs if __name__ == "__main__": print("Definition of model")
19,392
34.453382
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/Softmax-conf/config_train_asvspoof2019/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torchaudio import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import sandbox.eval_asvspoof as nii_asvspoof __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std( in_dim,out_dim, args, prj_conf, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_f = prj_conf.optional_argument self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = None self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # confidence predictor self.m_conf = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) if self.v_emd_dim is None: self.v_emd_dim = (lfcc_dim // 16) * 32 else: assert self.v_emd_dim == (lfcc_dim//16) * 32, "v_emd_dim error" self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_out_class) ) self.m_conf.append( torch_nn.Sequential( torch_nn.Linear((lfcc_dim // 16) * 32, 128), torch_nn.Tanh(), torch_nn.Linear(128, 1), torch_nn.Sigmoid() ) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) self.m_conf = torch_nn.ModuleList(self.m_conf) # output self.m_loss = torch_nn.NLLLoss() self.m_lambda = torch_nn.Parameter(torch.tensor([0.1]), requires_grad=False) self.m_budget = 0.5 # done return def prepare_mean_std(self, in_dim, out_dim, args, prj_conf, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = (hidden_features_lstm + hidden_features).mean(1) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_logit(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models output_act = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_output_act): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb) # feature_vec is [batch * submodel, output-class] return output_act def _compute_score(self, logits): """ """ # [batch * submodel, output-class], logits # [:, 1] denotes being bonafide if logits.shape[1] == 2: return logits[:, 1] - logits[:, 0] else: return logits[:, -1] def _compute_conf(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models conf_score = torch.zeros( [batch_size * self.v_submodels, 1], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_conf): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] conf_score[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb) return conf_score.squeeze(1) def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def _clamp_prob(self, input_prob, clamp_val=1e-12): return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val) def _loss(self, logits, targets, conf_scores): """ """ with torch.no_grad(): index = torch.zeros_like(logits) index.scatter_(1, targets.data.view(-1, 1), 1) # clamp the probablity and conf scores prob = self._clamp_prob(torch_nn_func.softmax(logits, dim=1)) conf_tmp = self._clamp_prob(conf_scores) # mixed log probablity log_prob = torch.log( prob * conf_tmp.view(-1, 1) + (1 - conf_tmp.view(-1, 1)) * index) loss = self.m_loss(log_prob, targets) loss2 = -torch.log(conf_scores).mean() loss = loss + self.m_lambda * loss2 if self.m_budget > loss2: self.m_lambda.data = self.m_lambda / 1.01 else: self.m_lambda.data = self.m_lambda / 0.99 return loss.mean() def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] # too short sentences if not self.training and x.shape[1] < 3000: targets = self._get_target(filenames) for filename, target in zip(filenames, targets): print("Output, %s, %d, %f, %f" % ( filename, target, 0.0, 0.0)) return None # for training, do balanced batch if self.training: target = self._get_target(filenames) bona_indx = np.argwhere(np.array(target) > 0)[:, 0] spoof_indx = np.argwhere(np.array(target) == 0)[:, 0] num = np.min([len(bona_indx), len(spoof_indx)]) trial_indx = np.concatenate([bona_indx[0:num], spoof_indx[0:num]]) if len(trial_indx) == 0: x_ = x flag_no_data = True else: filenames = [filenames[x] for x in trial_indx] datalength = [datalength[x] for x in trial_indx] x_ = x[trial_indx] flag_no_data = False else: x_ = x feature_vec = self._compute_embedding(x_, datalength) logits = self._compute_logit(feature_vec) conf_scores = self._compute_conf(feature_vec) if self.training: # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) # randomly set half of the conf data to 1 b = torch.bernoulli(torch.zeros_like(conf_scores).uniform_(0, 1)) conf_scores = conf_scores * b + (1 - b) loss = self._loss(logits, target_vec, conf_scores) if flag_no_data: return loss * 0 else: return loss else: scores = self._compute_score(logits) targets = self._get_target(filenames) for filename, target, score, conf in \ zip(filenames, targets, scores, conf_scores): print("Output, %s, %d, %f, %f" % ( filename, target, score.item(), conf.item())) # don't write output score as a single file return None def get_embedding(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] feature_vec = self._compute_embedding(x, datalength) return feature_vec class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ def compute(self, outputs, target): """ """ return outputs if __name__ == "__main__": print("Definition of model")
18,469
34.587669
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/Softmax-energy/config_train_asvspoof2019/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torchaudio import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import sandbox.eval_asvspoof as nii_asvspoof __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" g_attack_map = {'-': 1, 'A01': 1, 'A02': 1, 'A03': 1, 'A04': 1, 'A05': 1, 'A06': 1, 'A07': 1, 'A08': 1, 'A09': 1, 'A10': 1, 'A11': 1, 'A12': 1, 'A13': 1, 'A14': 1, 'A15': 1, 'A16': 1, 'A17': 1, 'A18': 1, 'A19': 1} def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial The format is: SPEAKER TRIAL_NAME - SPOOF_TYPE TAG LA_0031 LA_E_5932896 - A13 spoof LA_0030 LA_E_5849185 - - bonafide ... input: ----- protocol_filepath: string, path to the protocol file target_row: int, default -1, use line[-1] as the target label output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} if type(protocol_filepaths) is str: tmp = [protocol_filepaths] else: tmp = protocol_filepaths for protocol_filepath in tmp: with open(protocol_filepath, 'r') as file_ptr: for line in file_ptr: line = line.rstrip('\n') cols = line.split(sep) try: data_buffer[cols[1]] = g_attack_map[cols[target_row]] except KeyError: data_buffer[cols[1]] = False return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std( in_dim,out_dim, args, prj_conf, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_f = prj_conf.optional_argument self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f) self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map, ' ', -2) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = None self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # confidence predictor self.m_conf = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) if self.v_emd_dim is None: self.v_emd_dim = (lfcc_dim // 16) * 32 else: assert self.v_emd_dim == (lfcc_dim//16) * 32, "v_emd_dim error" self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output self.m_loss = torch_nn.CrossEntropyLoss() self.m_temp = 1 self.m_lambda = 0. self.m_e_m_in = -25.0 self.m_e_m_out = -7.0 # done return def prepare_mean_std(self, in_dim, out_dim, args, prj_conf, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = (hidden_features_lstm + hidden_features).mean(1) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_logit(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models output_act = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_output_act): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb) # feature_vec is [batch * submodel, output-class] return output_act def _compute_score(self, logits): """ """ # [batch * submodel, output-class], logits # [:, 1] denotes being bonafide if logits.shape[1] == 2: return logits[:, 1] - logits[:, 0] else: return logits[:, -1] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def _clamp_prob(self, input_prob, clamp_val=1e-12): return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val) def _get_in_out_indx(self, filenames): in_indx = [] out_indx = [] for x, y in enumerate(filenames): if self.in_out_parser[y]: in_indx.append(x) else: out_indx.append(x) return np.array(in_indx), np.array(out_indx) def _energy(self, logits): """ """ # - T \log \sum_y \exp (logits[x, y] / T) eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1) return eng def _loss(self, logits, targets, energy, in_indx, out_indx): """ """ # loss over cross-entropy on in-dist. data if len(in_indx): loss = self.m_loss(logits[in_indx], targets[in_indx]) else: loss = 0 # loss on energy of in-dist.data if len(in_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean() # loss on energy of out-dist. data if len(out_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean() return loss def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] # too short sentences if not self.training and x.shape[1] < 3000: targets = self._get_target(filenames) for filename, target in zip(filenames, targets): print("Output, %s, %d, %f, %f" % ( filename, target, 0.0, 0.0)) return None feature_vec = self._compute_embedding(x, datalength) logits = self._compute_logit(feature_vec) energy = self._energy(logits) in_indx, out_indx = self._get_in_out_indx(filenames) if self.training: # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) # loss loss = self._loss(logits, target_vec, energy, in_indx, out_indx) return loss else: scores = self._compute_score(logits) targets = self._get_target(filenames) for filename, target, score, energytmp in \ zip(filenames, targets, scores, energy): print("Output, %s, %d, %f, %f" % ( filename, target, score.item(), -energytmp.item())) # don't write output score as a single file return None def get_embedding(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] feature_vec = self._compute_embedding(x, datalength) return feature_vec class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ def compute(self, outputs, target): """ """ return outputs if __name__ == "__main__": print("Definition of model")
18,378
34.276392
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/Softmax-energy/config_train_asvspoof2019_bc10/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torchaudio import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import sandbox.eval_asvspoof as nii_asvspoof __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" g_attack_map = {'-': 1, 'A01': 1, 'A02': 1, 'A03': 1, 'A04': 1, 'A05': 1, 'A06': 1, 'A07': 1, 'A08': 1, 'A09': 1, 'A10': 1, 'A11': 1, 'A12': 1, 'A13': 1, 'A14': 1, 'A15': 1, 'A16': 1, 'A17': 1, 'A18': 1, 'A19': 1} def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial The format is: SPEAKER TRIAL_NAME - SPOOF_TYPE TAG LA_0031 LA_E_5932896 - A13 spoof LA_0030 LA_E_5849185 - - bonafide ... input: ----- protocol_filepath: string, path to the protocol file target_row: int, default -1, use line[-1] as the target label output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} if type(protocol_filepaths) is str: tmp = [protocol_filepaths] else: tmp = protocol_filepaths for protocol_filepath in tmp: with open(protocol_filepath, 'r') as file_ptr: for line in file_ptr: line = line.rstrip('\n') cols = line.split(sep) try: data_buffer[cols[1]] = g_attack_map[cols[target_row]] except KeyError: data_buffer[cols[1]] = False return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std( in_dim,out_dim, args, prj_conf, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_f = prj_conf.optional_argument self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f) self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map, ' ', -2) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = None self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # confidence predictor self.m_conf = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) if self.v_emd_dim is None: self.v_emd_dim = (lfcc_dim // 16) * 32 else: assert self.v_emd_dim == (lfcc_dim//16) * 32, "v_emd_dim error" self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output self.m_loss = torch_nn.CrossEntropyLoss() self.m_temp = 1 self.m_lambda = 0.1 self.m_e_m_in = -25.0 self.m_e_m_out = -7.0 # done return def prepare_mean_std(self, in_dim, out_dim, args, prj_conf, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = (hidden_features_lstm + hidden_features).mean(1) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_logit(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models output_act = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_output_act): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb) # feature_vec is [batch * submodel, output-class] return output_act def _compute_score(self, logits): """ """ # [batch * submodel, output-class], logits # [:, 1] denotes being bonafide if logits.shape[1] == 2: return logits[:, 1] - logits[:, 0] else: return logits[:, -1] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def _clamp_prob(self, input_prob, clamp_val=1e-12): return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val) def _get_in_out_indx(self, filenames): in_indx = [] out_indx = [] for x, y in enumerate(filenames): if self.in_out_parser[y]: in_indx.append(x) else: out_indx.append(x) return np.array(in_indx), np.array(out_indx) def _energy(self, logits): """ """ # - T \log \sum_y \exp (logits[x, y] / T) eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1) return eng def _loss(self, logits, targets, energy, in_indx, out_indx): """ """ # loss over cross-entropy on in-dist. data if len(in_indx): loss = self.m_loss(logits[in_indx], targets[in_indx]) else: loss = 0 # loss on energy of in-dist.data if len(in_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean() # loss on energy of out-dist. data if len(out_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean() return loss def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] # too short sentences if not self.training and x.shape[1] < 3000: targets = self._get_target(filenames) for filename, target in zip(filenames, targets): print("Output, %s, %d, %f, %f" % ( filename, target, 0.0, 0.0)) return None feature_vec = self._compute_embedding(x, datalength) logits = self._compute_logit(feature_vec) energy = self._energy(logits) in_indx, out_indx = self._get_in_out_indx(filenames) if self.training: # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) # loss loss = self._loss(logits, target_vec, energy, in_indx, out_indx) return loss else: scores = self._compute_score(logits) targets = self._get_target(filenames) for filename, target, score, conf in \ zip(filenames, targets, scores, energy): print("Output, %s, %d, %f, %f" % ( filename, target, score.item(), -energy.item())) # don't write output score as a single file return None def get_embedding(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] feature_vec = self._compute_embedding(x, datalength) return feature_vec class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ def compute(self, outputs, target): """ """ return outputs if __name__ == "__main__": print("Definition of model")
18,371
34.262956
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/06-asvspoof-ood/Softmax-energy/config_train_asvspoof2019_esp/01/model.py
#!/usr/bin/env python """ model.py Self defined model definition. Usage: """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torchaudio import torch.nn.functional as torch_nn_func import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_front_end import core_scripts.other_tools.debug as nii_debug import core_scripts.data_io.seq_info as nii_seq_tk import sandbox.eval_asvspoof as nii_asvspoof __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" g_attack_map = {'-': 1, 'A01': 1, 'A02': 1, 'A03': 1, 'A04': 1, 'A05': 1, 'A06': 1, 'A07': 1, 'A08': 1, 'A09': 1, 'A10': 1, 'A11': 1, 'A12': 1, 'A13': 1, 'A14': 1, 'A15': 1, 'A16': 1, 'A17': 1, 'A18': 1, 'A19': 1} def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1): """ Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial The format is: SPEAKER TRIAL_NAME - SPOOF_TYPE TAG LA_0031 LA_E_5932896 - A13 spoof LA_0030 LA_E_5849185 - - bonafide ... input: ----- protocol_filepath: string, path to the protocol file target_row: int, default -1, use line[-1] as the target label output: ------- data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof) """ data_buffer = {} if type(protocol_filepaths) is str: tmp = [protocol_filepaths] else: tmp = protocol_filepaths for protocol_filepath in tmp: with open(protocol_filepath, 'r') as file_ptr: for line in file_ptr: line = line.rstrip('\n') cols = line.split(sep) try: data_buffer[cols[1]] = g_attack_map[cols[target_row]] except KeyError: data_buffer[cols[1]] = False return data_buffer ############## ## FOR MODEL ############## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ##### required part, no need to change ##### # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std( in_dim,out_dim, args, prj_conf, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) # a flag for debugging (by default False) #self.model_debug = False #self.validation = False ##### #### # on input waveform and output target #### # Load protocol and prepare the target data for network training protocol_f = prj_conf.optional_argument self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f) self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map, ' ', -2) # Working sampling rate # torchaudio may be used to change sampling rate self.m_target_sr = 16000 #### # optional configs (not used) #### # re-sampling (optional) #self.m_resampler = torchaudio.transforms.Resample( # prj_conf.wav_samp_rate, self.m_target_sr) # vad (optional) #self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr) # flag for balanced class (temporary use) #self.v_flag = 1 #### # front-end configuration # multiple front-end configurations may be used # by default, use a single front-end #### # frame shift (number of waveform points) self.frame_hops = [160] # frame length self.frame_lens = [320] # FFT length self.fft_n = [512] # LFCC dim (base component) self.lfcc_dim = [20] self.lfcc_with_delta = True # window type self.win = torch.hann_window # floor in log-spectrum-amplitude calculating (not used) self.amp_floor = 0.00001 # number of frames to be kept for each trial # no truncation self.v_truncate_lens = [None for x in self.frame_hops] # number of sub-models (by default, a single model) self.v_submodels = len(self.frame_lens) # dimension of embedding vectors # here, the embedding is just the activation before sigmoid() self.v_emd_dim = None self.v_out_class = 2 #### # create network #### # 1st part of the classifier self.m_transform = [] # self.m_before_pooling = [] # 2nd part of the classifier self.m_output_act = [] # front-end self.m_frontend = [] # confidence predictor self.m_conf = [] # it can handle models with multiple front-end configuration # by default, only a single front-end for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip( self.v_truncate_lens, self.fft_n, self.lfcc_dim)): fft_n_bins = fft_n // 2 + 1 if self.lfcc_with_delta: lfcc_dim = lfcc_dim * 3 self.m_transform.append( torch_nn.Sequential( torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(48, affine=False), torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch.nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(64, affine=False), torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]), nii_nn.MaxFeatureMap2D(), torch_nn.BatchNorm2d(32, affine=False), torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]), nii_nn.MaxFeatureMap2D(), torch_nn.MaxPool2d([2, 2], [2, 2]), torch_nn.Dropout(0.7) ) ) self.m_before_pooling.append( torch_nn.Sequential( nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32), nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32) ) ) if self.v_emd_dim is None: self.v_emd_dim = (lfcc_dim // 16) * 32 else: assert self.v_emd_dim == (lfcc_dim//16) * 32, "v_emd_dim error" self.m_output_act.append( torch_nn.Linear((lfcc_dim // 16) * 32, self.v_out_class) ) self.m_frontend.append( nii_front_end.LFCC(self.frame_lens[idx], self.frame_hops[idx], self.fft_n[idx], self.m_target_sr, self.lfcc_dim[idx], with_energy=True) ) self.m_frontend = torch_nn.ModuleList(self.m_frontend) self.m_transform = torch_nn.ModuleList(self.m_transform) self.m_output_act = torch_nn.ModuleList(self.m_output_act) self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling) # output self.m_loss = torch_nn.CrossEntropyLoss() self.m_temp = 1 self.m_lambda = 0.1 self.m_e_m_in = -25.0 self.m_e_m_out = -7.0 # done return def prepare_mean_std(self, in_dim, out_dim, args, prj_conf, data_mean_std=None): """ prepare mean and std for data processing This is required for the Pytorch project, but not relevant to this code """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data This is required for the Pytorch project, but not relevant to this code """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data This is required for the Pytorch project, but not relevant to this code """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network This is required for the Pytorch project, but not relevant to this code """ return y * self.output_std + self.output_mean def _front_end(self, wav, idx, trunc_len, datalength): """ simple fixed front-end to extract features input: ------ wav: waveform idx: idx of the trial in mini-batch trunc_len: number of frames to be kept after truncation datalength: list of data length in mini-batch output: ------- x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim) """ with torch.no_grad(): x_sp_amp = self.m_frontend[idx](wav.squeeze(-1)) # return return x_sp_amp def _compute_embedding(self, x, datalength): """ definition of forward method Assume x (batchsize, length, dim) Output x (batchsize * number_filter, output_dim) """ # resample if necessary #x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1) # number of sub models batch_size = x.shape[0] # buffer to store output scores from sub-models output_emb = torch.zeros([batch_size * self.v_submodels, self.v_emd_dim], device=x.device, dtype=x.dtype) # compute scores for each sub-models for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \ enumerate( zip(self.frame_hops, self.frame_lens, self.fft_n, self.v_truncate_lens, self.m_transform, self.m_before_pooling, self.m_output_act)): # extract front-end feature x_sp_amp = self._front_end(x, idx, trunc_len, datalength) # compute scores # 1. unsqueeze to (batch, 1, frame_length, fft_bin) # 2. compute hidden features hidden_features = m_trans(x_sp_amp.unsqueeze(1)) # 3. (batch, channel, frame//N, feat_dim//N) -> # (batch, frame//N, channel * feat_dim//N) # where N is caused by conv with stride hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous() frame_num = hidden_features.shape[1] hidden_features = hidden_features.view(batch_size, frame_num, -1) # 4. pooling # 4. pass through LSTM then summing hidden_features_lstm = m_be_pool(hidden_features) # 5. pass through the output layer tmp_emb = (hidden_features_lstm + hidden_features).mean(1) output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb return output_emb def _compute_logit(self, feature_vec, inference=False): """ """ # number of sub models batch_size = feature_vec.shape[0] # buffer to store output scores from sub-models output_act = torch.zeros( [batch_size * self.v_submodels, self.v_out_class], device=feature_vec.device, dtype=feature_vec.dtype) # compute scores for each sub-models for idx, m_output in enumerate(self.m_output_act): tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size] output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb) # feature_vec is [batch * submodel, output-class] return output_act def _compute_score(self, logits): """ """ # [batch * submodel, output-class], logits # [:, 1] denotes being bonafide if logits.shape[1] == 2: return logits[:, 1] - logits[:, 0] else: return logits[:, -1] def _get_target(self, filenames): try: return [self.protocol_parser[x] for x in filenames] except KeyError: print("Cannot find target data for %s" % (str(filenames))) sys.exit(1) def _clamp_prob(self, input_prob, clamp_val=1e-12): return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val) def _get_in_out_indx(self, filenames): in_indx = [] out_indx = [] for x, y in enumerate(filenames): if self.in_out_parser[y]: in_indx.append(x) else: out_indx.append(x) return np.array(in_indx), np.array(out_indx) def _energy(self, logits): """ """ # - T \log \sum_y \exp (logits[x, y] / T) eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1) return eng def _loss(self, logits, targets, energy, in_indx, out_indx): """ """ # loss over cross-entropy on in-dist. data if len(in_indx): loss = self.m_loss(logits[in_indx], targets[in_indx]) else: loss = 0 # loss on energy of in-dist.data if len(in_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean() # loss on energy of out-dist. data if len(out_indx): loss += self.m_lambda * torch.pow( torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean() return loss def forward(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] # too short sentences if not self.training and x.shape[1] < 3000: targets = self._get_target(filenames) for filename, target in zip(filenames, targets): print("Output, %s, %d, %f, %f" % ( filename, target, 0.0, 0.0)) return None feature_vec = self._compute_embedding(x, datalength) logits = self._compute_logit(feature_vec) energy = self._energy(logits) in_indx, out_indx = self._get_in_out_indx(filenames) if self.training: # target target = self._get_target(filenames) target_vec = torch.tensor(target, device=x.device, dtype=torch.long) target_vec = target_vec.repeat(self.v_submodels) # loss loss = self._loss(logits, target_vec, energy, in_indx, out_indx) return loss else: scores = self._compute_score(logits) targets = self._get_target(filenames) for filename, target, score, conf in \ zip(filenames, targets, scores, energy): print("Output, %s, %d, %f, %f" % ( filename, target, score.item(), -energy.item())) # don't write output score as a single file return None def get_embedding(self, x, fileinfo): filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo] datalength = [nii_seq_tk.parse_length(y) for y in fileinfo] feature_vec = self._compute_embedding(x, datalength) return feature_vec class Loss(): """ Wrapper to define loss function """ def __init__(self, args): """ """ def compute(self, outputs, target): """ """ return outputs if __name__ == "__main__": print("Definition of model")
18,371
34.262956
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/blow/main.py
#!/usr/bin/env python """ main.py for project-NN-pytorch/projects The training/inference process wrapper. Dataset API is replaced with NII_MergeDataSetLoader. It is more convenient to train model on corpora stored in different directories. Requires model.py and config.py (config_merge_datasets.py) Usage: $: python main.py [options] """ from __future__ import absolute_import import os import sys import torch import importlib import core_scripts.other_tools.display as nii_warn import core_scripts.data_io.default_data_io as nii_default_dset import core_scripts.data_io.customize_dataset as nii_dset import core_scripts.data_io.conf as nii_dconf import core_scripts.other_tools.list_tools as nii_list_tool import core_scripts.config_parse.config_parse as nii_config_parse import core_scripts.config_parse.arg_parse as nii_arg_parse import core_scripts.op_manager.op_manager as nii_op_wrapper import core_scripts.nn_manager.nn_manager as nii_nn_wrapper import core_scripts.startup_config as nii_startup __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" def main(): """ main(): the default wrapper for training and inference process Please prepare config.py and model.py """ # arguments initialization args = nii_arg_parse.f_args_parsed() # nii_warn.f_print_w_date("Start program", level='h') nii_warn.f_print("Load module: %s" % (args.module_config)) nii_warn.f_print("Load module: %s" % (args.module_model)) prj_conf = importlib.import_module(args.module_config) prj_model = importlib.import_module(args.module_model) # initialization nii_startup.set_random_seed(args.seed, args) use_cuda = not args.no_cuda and torch.cuda.is_available() device = torch.device("cuda" if use_cuda else "cpu") # prepare data io if not args.inference: params = {'batch_size': args.batch_size, 'shuffle': args.shuffle, 'num_workers': args.num_workers, 'sampler': args.sampler} in_trans_fns = prj_conf.input_trans_fns \ if hasattr(prj_conf, 'input_trans_fns') else None out_trans_fns = prj_conf.output_trans_fns \ if hasattr(prj_conf, 'output_trans_fns') else None # Load file list and create data loader trn_lst = prj_conf.trn_list trn_set = nii_dset.NII_MergeDataSetLoader( prj_conf.trn_set_name, \ trn_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', params = params, truncate_seq = prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = True, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) if prj_conf.val_list is not None: val_lst = prj_conf.val_list val_set = nii_dset.NII_MergeDataSetLoader( prj_conf.val_set_name, val_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', \ params = params, truncate_seq= prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) else: val_set = None # initialize the model and loss function model = prj_model.Model(trn_set.get_in_dim(), \ trn_set.get_out_dim(), \ args, prj_conf, trn_set.get_data_mean_std()) loss_wrapper = prj_model.Loss(args) # initialize the optimizer optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args) # if necessary, resume training if args.trained_model == "": checkpoint = None else: checkpoint = torch.load(args.trained_model) # start training nii_nn_wrapper.f_train_wrapper(args, model, loss_wrapper, device, optimizer_wrapper, trn_set, val_set, checkpoint) # done for traing else: # for inference # default, no truncating, no shuffling params = {'batch_size': args.batch_size, 'shuffle': False, 'num_workers': args.num_workers} in_trans_fns = prj_conf.test_input_trans_fns \ if hasattr(prj_conf, 'test_input_trans_fns') else None out_trans_fns = prj_conf.test_output_trans_fns \ if hasattr(prj_conf, 'test_output_trans_fns') else None if type(prj_conf.test_list) is list: t_lst = prj_conf.test_list else: t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list) test_set = nii_dset.NII_MergeDataSetLoader( prj_conf.test_set_name, \ t_lst, \ prj_conf.test_input_dirs, prj_conf.input_exts, prj_conf.input_dims, prj_conf.input_reso, prj_conf.input_norm, prj_conf.test_output_dirs, prj_conf.output_exts, prj_conf.output_dims, prj_conf.output_reso, prj_conf.output_norm, './', params = params, truncate_seq= None, min_seq_len = None, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) # initialize model model = prj_model.Model(test_set.get_in_dim(), \ test_set.get_out_dim(), \ args, prj_conf) if args.trained_model == "": print("No model is loaded by ---trained-model for inference") print("By default, load %s%s" % (args.save_trained_name, args.save_model_ext)) checkpoint = torch.load("%s%s" % (args.save_trained_name, args.save_model_ext)) else: checkpoint = torch.load(args.trained_model) # do inference and output data nii_nn_wrapper.f_inference_wrapper(args, model, device, \ test_set, checkpoint) # done return if __name__ == "__main__": main()
7,819
35.886792
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/blow/model.py
#!/usr/bin/env python """ model.py for Blow version: 1 """ from __future__ import absolute_import from __future__ import print_function import os import sys import time import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import core_scripts.other_tools.debug as nii_debug import core_scripts.other_tools.display as nii_warn import sandbox.block_nn as nii_nn import sandbox.block_blow as nii_blow __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2021, Xin Wang" ######## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ################# ## must-have ################# # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) self.input_dim = in_dim self.output_dim = out_dim # a flag for debugging (by default False) self.model_debug = False ################# ## model config ################# # waveform sampling rate self.sample_rate = prj_conf.wav_samp_rate # load speaker map self.speaker_map = prj_conf.options['speaker_map'] self.speaker_num = self.speaker_map.num() if 'conversion_map' in prj_conf.options: self.conversion_map = prj_conf.options['conversion_map'] else: self.conversion_map = None self.cond_dim = 128 self.num_block = 8 self.num_flow_steps_perblock = 12 self.num_conv_channel_size = 512 self.num_conv_conv_kernel = 3 self.m_spk_emd = torch.nn.Embedding(self.speaker_num, self.cond_dim) self.m_blow = nii_blow.Blow( self.cond_dim, self.num_block, self.num_flow_steps_perblock, self.num_conv_channel_size, self.num_conv_conv_kernel) # only used for synthesis self.m_overlap = nii_blow.OverlapAdder(4096, 4096//4, False) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def forward(self, wav, fileinfo): """loss = forward(self, input_feat, wav) input ----- wav: tensor, target waveform (batchsize, length2, 1) it should be raw waveform, flot valued, between (-1, 1) the code will do mu-law conversion fileinfo: list, file information for each data in the batch output ------ loss: tensor / scalar, Note: returned loss can be directly used as the loss value no need to write Loss() """ # prepare speaker IDs # (batch, ) speaker_ids = torch.tensor( [self.speaker_map.parse(x) for x in fileinfo], dtype=torch.long, device=wav.device) # convert to embeddings # (batch, 1, cond_dim) speaker_emd = self.m_spk_emd(speaker_ids).unsqueeze(1) # normalize conditiona feature #input_feat = self.normalize_input(input_feat) # compute z, neg_logp, logp_z, log_detjac = self.m_blow(wav, speaker_emd) return [[-logp_z, -log_detjac], [True, True]] def inference(self, wav, fileinfo): """wav = inference(mels) input ----- wav: tensor, target waveform (batchsize, length2, 1) output ------ wav_new: tensor, same shape """ # framing the input waveform into frames # # framed_wav (batchsize, frame_num, frame_length) framed_wav = self.m_overlap(wav) batch, frame_num, frame_len = framed_wav.shape # framed_Wav (batchsize * frame_num, frame_length, 1) framed_wav = framed_wav.view(-1, frame_len).unsqueeze(-1) # get the speaker IDs # (batch, ) speaker_ids = torch.tensor( [self.speaker_map.parse(x) for x in fileinfo], dtype=torch.long, device=wav.device) # (batch * frame_num) speaker_ids = speaker_ids.repeat_interleave(frame_num) # if conversion map is defined, change the speaker identity following # the conversion map. Otherwise, specify target speaker ID # through environment variable TEMP_TARGET_SPEAKER if self.conversion_map: target_speaker = torch.tensor( [self.speaker_map.get_idx( self.conversion_map[self.speaker_map.parse(x, False)]) \ for x in fileinfo], device=wav.device, dtype=torch.long) target_speaker = target_speaker.repeat_interleave(frame_num) else: # target speaker ID target_speaker = torch.tensor( [int(os.getenv('TEMP_TARGET_SPEAKER'))] * speaker_ids.shape[0], device=wav.device, dtype=torch.long) # if it is intended to swap the source / target speaker ID # this can be used to recover the original waveform given converted # speech flag_reverse = os.getenv('TEMP_FLAG_REVERSE') if flag_reverse and int(flag_reverse): # if this is for reverse conversion # swap the IDs print("revert IDs") tmp = speaker_ids speaker_ids = target_speaker target_speaker = tmp # print some information for idx in range(len(speaker_ids)): if idx % frame_num == 0: print("From ID {:3d} to {:3d}, ".format( speaker_ids[idx], target_speaker[idx]), end=' ') # get embeddings (batch * frame_num, 1, cond_dim) speaker_emd = self.m_spk_emd(speaker_ids).unsqueeze(1) target_speaker_emb = self.m_spk_emd(target_speaker).unsqueeze(1) # compute z, neg_logp, logp_z, log_detjac = self.m_blow(framed_wav, speaker_emd) # output_framed (batch * frame, frame_length, 1) output_framed = self.m_blow.reverse(z, target_speaker_emb) # overlap and add # view -> (batch, frame_num, frame_length) return self.m_overlap.reverse(output_framed.view(batch, -1, frame_len), True) def convert(self, wav, src_id, tar_id): """wav = inference(mels) input ----- wav: tensor, target waveform (batchsize, length2, 1) src_id: int, ID of source speaker tar_id: int, ID of target speaker output ------ wav_new: tensor, same shape """ # framing the input waveform into frames # m_overlap.forward does framing # framed_wav (batchsize, frame_num, frame_length) framed_wav = self.m_overlap(wav) batch, frame_num, frame_len = framed_wav.shape # change frames into batch # framed_Wav (batchsize * frame_num, frame_length, 1) framed_wav = framed_wav.view(-1, frame_len).unsqueeze(-1) # source speaker IDs # (batch, ) speaker_ids = torch.tensor([src_id for x in wav], dtype=torch.long, device=wav.device) # (batch * frame_num) speaker_ids = speaker_ids.repeat_interleave(frame_num) # get embeddings (batch * frame_num, 1, cond_dim) speaker_emd = self.m_spk_emd(speaker_ids).unsqueeze(1) # target speaker IDs tar_speaker_ids = torch.tensor([tar_id for x in wav], dtype=torch.long, device=wav.device) # (batch * frame_num) tar_speaker_ids = tar_speaker_ids.repeat_interleave(frame_num) target_speaker_emb = self.m_spk_emd(tar_speaker_ids).unsqueeze(1) # analysis z, _, _, _ = self.m_blow(framed_wav, speaker_emd) # synthesis # output_framed (batch * frame, frame_length, 1) output_framed = self.m_blow.reverse(z, target_speaker_emb) # overlap and add # view -> (batch, frame_num, frame_length) return self.m_overlap.reverse( output_framed.view(batch, -1, frame_len), True) # Loss is returned by model.forward(), no need to specify # just a place holder so that the output of model.forward() can be # sent to the optimizer class Loss(): def __init__(self, args): return def compute(self, output, target): return output if __name__ == "__main__": print("Definition of model")
10,716
33.795455
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/blow/config.py
#!/usr/bin/env python """ config.py This configuration file specifiess the input and output data """ __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['vctk_blow_trn'] val_set_name = ['vctk_blow_val'] # for convenience tmp1 = '../DATA/vctk-blow' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp1 + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp1 + '/scp/val.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp1 + '/vctk_wav']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [1] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.wav'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [1] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [False] # Similar configurations for output features # for convenience, simply load the data as target too output_dirs = input_dirs output_dims = [1] output_exts = ['.wav'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = None # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = None # Other optional arguments, the definition of which depends on a specific model # here we define a speaker ID manager class VCTKSpeakerMap: def __init__(self): speaker_list = tmp1 + '/scp/spk.lst' self.m_speaker_map = {} with open(speaker_list, 'r') as file_ptr: for idx, line in enumerate(file_ptr): line = line.rstrip('\n') self.m_speaker_map[line] = idx return def num(self): # leave one for unseen return len(self.m_speaker_map) + 1 def parse(self, filename, return_idx=True): # # filename will be in this format: '8758,p***_***,1,4096,16000' # we need to get the p*** part spk = filename.split('_')[0].split(',')[-1] if return_idx: return self.get_idx(spk) else: return spk def get_idx(self, spk): if spk in self.m_speaker_map: return self.m_speaker_map[spk] else: return len(self.m_speaker_map) # conversion_map defines the conversion pairs: p361 -> p245, p278 -> p287 ... options = {'speaker_map': VCTKSpeakerMap(), 'conversion_map': {'p361': 'p245', 'p278': 'p287', 'p302': 'p298', 'p361': 'p345', 'p260': 'p267', 'p273': 'p351', 'p245': 'p273', 'p304': 'p238', 'p297': 'p283', 'p246': 'p362'}} # data pre-processing functions # input_trans_fns must have the same shape as input_dirs # output_trans_fns must have the same shape as output_dirs # Thus, input_trans_fns[x][y] defines the transformation function # for the y-th feature of the x-th sub-database. # # This function is called when DataLoader loads the data from the disk # (see f_post_data_process in core_scripts.data_io.default_io.py) # # 4096 denotes frame length # 0.2 is the reference coefficient for waveform emphasis from sandbox import block_blow input_trans_fns = [[lambda x: block_blow.wav_aug(x, 4096, 0.2, wav_samp_rate)]] output_trans_fns = [[lambda x: block_blow.wav_aug(x, 4096, 0.2, wav_samp_rate)]] ######################################################### ## Configuration for inference stage ######################################################### # similar options to training stage test_set_name = ['vctk_blow_test'] # List of test set data # for convenience, you may directly load test_set list here test_list = [tmp1 + '/scp/test_tiny.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = [[tmp1 + '/vctk_wav_test_tiny']] # Directories for output features, which are [] test_output_dirs = [[]]
5,379
32.416149
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/waveglow/main.py
#!/usr/bin/env python """ main.py for project-NN-pytorch/projects The training/inference process wrapper. Dataset API is replaced with NII_MergeDataSetLoader. It is more convenient to train model on corpora stored in different directories. Requires model.py and config.py (config_merge_datasets.py) Usage: $: python main.py [options] """ from __future__ import absolute_import import os import sys import torch import importlib import core_scripts.other_tools.display as nii_warn import core_scripts.data_io.default_data_io as nii_default_dset import core_scripts.data_io.customize_dataset as nii_dset import core_scripts.data_io.conf as nii_dconf import core_scripts.other_tools.list_tools as nii_list_tool import core_scripts.config_parse.config_parse as nii_config_parse import core_scripts.config_parse.arg_parse as nii_arg_parse import core_scripts.op_manager.op_manager as nii_op_wrapper import core_scripts.nn_manager.nn_manager as nii_nn_wrapper import core_scripts.startup_config as nii_startup __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" def main(): """ main(): the default wrapper for training and inference process Please prepare config.py and model.py """ # arguments initialization args = nii_arg_parse.f_args_parsed() # nii_warn.f_print_w_date("Start program", level='h') nii_warn.f_print("Load module: %s" % (args.module_config)) nii_warn.f_print("Load module: %s" % (args.module_model)) prj_conf = importlib.import_module(args.module_config) prj_model = importlib.import_module(args.module_model) # initialization nii_startup.set_random_seed(args.seed, args) use_cuda = not args.no_cuda and torch.cuda.is_available() device = torch.device("cuda" if use_cuda else "cpu") # prepare data io if not args.inference: params = {'batch_size': args.batch_size, 'shuffle': args.shuffle, 'num_workers': args.num_workers, 'sampler': args.sampler} # Load file list and create data loader trn_lst = prj_conf.trn_list trn_set = nii_dset.NII_MergeDataSetLoader( prj_conf.trn_set_name, \ trn_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', params = params, truncate_seq = prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = True, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets) if prj_conf.val_list is not None: val_lst = prj_conf.val_list val_set = nii_dset.NII_MergeDataSetLoader( prj_conf.val_set_name, val_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', \ params = params, truncate_seq= prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets) else: val_set = None # initialize the model and loss function model = prj_model.Model(trn_set.get_in_dim(), \ trn_set.get_out_dim(), \ args, prj_conf, trn_set.get_data_mean_std()) loss_wrapper = prj_model.Loss(args) # initialize the optimizer optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args) # if necessary, resume training if args.trained_model == "": checkpoint = None else: checkpoint = torch.load(args.trained_model) # start training nii_nn_wrapper.f_train_wrapper(args, model, loss_wrapper, device, optimizer_wrapper, trn_set, val_set, checkpoint) # done for traing else: # for inference # default, no truncating, no shuffling params = {'batch_size': args.batch_size, 'shuffle': False, 'num_workers': args.num_workers} if type(prj_conf.test_list) is list: t_lst = prj_conf.test_list else: t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list) test_set = nii_dset.NII_MergeDataSetLoader( prj_conf.test_set_name, \ t_lst, \ prj_conf.test_input_dirs, prj_conf.input_exts, prj_conf.input_dims, prj_conf.input_reso, prj_conf.input_norm, prj_conf.test_output_dirs, prj_conf.output_exts, prj_conf.output_dims, prj_conf.output_reso, prj_conf.output_norm, './', params = params, truncate_seq= None, min_seq_len = None, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets) # initialize model model = prj_model.Model(test_set.get_in_dim(), \ test_set.get_out_dim(), \ args, prj_conf) if args.trained_model == "": print("No model is loaded by ---trained-model for inference") print("By default, load %s%s" % (args.save_trained_name, args.save_model_ext)) checkpoint = torch.load("%s%s" % (args.save_trained_name, args.save_model_ext)) else: checkpoint = torch.load(args.trained_model) # do inference and output data nii_nn_wrapper.f_inference_wrapper(args, model, device, \ test_set, checkpoint) # done return if __name__ == "__main__": main()
6,776
35.047872
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/waveglow/model.py
#!/usr/bin/env python """ model.py waveglow is defined in sandbox.block_waveglow. This file wrapps waveglow inside model() so that it can be used by the script. """ from __future__ import absolute_import from __future__ import print_function import sys import time import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import core_scripts.other_tools.debug as nii_debug import core_scripts.other_tools.display as nii_warn import sandbox.block_nn as nii_nn import sandbox.util_frontend as nii_nn_frontend import sandbox.block_waveglow as nii_waveglow __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ################# ## must-have ################# # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) self.input_dim = in_dim self.output_dim = out_dim # a flag for debugging (by default False) self.model_debug = False ################# ## model config ################# # whether z-normalize the input features? # the pre-trained model was trained without normalizing feat self.flag_normalize_input = False # waveform sampling rate self.sample_rate = prj_conf.wav_samp_rate # up-sample rate self.up_sample = prj_conf.input_reso[0] # configuration for WaveGlow # number of waveglow blocks self.num_waveglow_blocks = 3 # number of flow steps in each block self.num_flow_steps_perblock = 4 # number of wavenet layers in one flow step self.num_wn_blocks_perflow = 8 # channel size of the dilated conv in wavenet layer self.num_wn_channel_size = 256 # kernel size of the dilated conv in wavenet layer self.num_wn_conv_kernel = 3 # use affine transformation? (if False, use a + b) self.flag_affine = True # dimension of the early extracted latent z self.early_z_feature_dim = 2 # use legacy implementation of affine (default False) # the difference is the wavenet core to compute the affine # parameter. For more details, check sandbox.block_waveglow self.flag_affine_legacy_implementation = False self.m_waveglow = nii_waveglow.WaveGlow( in_dim, self.up_sample, self.num_waveglow_blocks, self.num_flow_steps_perblock, self.num_wn_blocks_perflow, self.num_wn_channel_size, self.num_wn_conv_kernel, self.flag_affine, self.early_z_feature_dim, self.flag_affine_legacy_implementation) # buffer for noise in inference self.bg_noise = None # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def forward(self, input_feat, wav): """loss = forward(self, input_feat, wav) input ----- input_feat: tensor, input features (batchsize, length1, input_dim) wav: tensor, target waveform (batchsize, length2, 1) it should be raw waveform, flot valued, between (-1, 1) the code will do mu-law conversion output ------ loss: tensor / scalar Note: returned loss can be directly used as the loss value no need to write Loss() """ # normalize conditiona feature if self.flag_normalize_input: input_feat = self.normalize_input(input_feat) # compute z_bags, neg_logp, logp_z, log_detjac = self.m_waveglow(wav, input_feat) return [[-logp_z, -log_detjac], [True, True]] def inference(self, input_feat): """wav = inference(mels) input ----- input_feat: tensor, input features (batchsize, length1, input_dim) output ------ wav: tensor, target waveform (batchsize, length2, 1) Note: length2 will be = length1 * self.up_sample """ #normalize input if self.flag_normalize_input: input_feat = self.normalize_input(input_feat) length = input_feat.shape[1] * self.up_sample noise = self.m_waveglow.get_z_noises(length, noise_std=0.6, batchsize=input_feat.shape[0]) output = self.m_waveglow.reverse(noise, input_feat) # use a randomized if self.bg_noise is None: self.bg_noise = self.m_waveglow.reverse(noise, input_feat * 0) # do post-filtering (spectra substraction) output = nii_nn_frontend.spectral_substraction(output,self.bg_noise,1.0) return output # Loss is returned by model.forward(), no need to specify # just a place holder so that the output of model.forward() can be # sent to the optimizer class Loss(): def __init__(self, args): return def compute(self, output, target): return output if __name__ == "__main__": print("Definition of model")
7,385
33.353488
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/waveglow/config.py
#!/usr/bin/env python """ config.py This configuration file specifiess the input and output data """ __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2021, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['cmu_all_trn'] val_set_name = ['cmu_all_val'] # for convenience tmp1 = '../DATA/cmu-arctic-data-set' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp1 + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp1 + '/scp/val.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp1 + '/5ms/melspec']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [80] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.mfbsp'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [80] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [True] # Similar configurations for output features output_dirs = [[tmp1 + '/wav_16k_norm']] output_dims = [1] output_exts = ['.wav'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = 16000 # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = 16000 # Other optional arguments, the definition of which depends on a specific model options = {} ######################################################### ## Configuration for inference stage ######################################################### # similar options to training stage test_set_name = ['cmu_all_test_tiny'] # List of test set data # for convenience, you may directly load test_set list here test_list = [['slt_arctic_b0474']] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = [[tmp1 + '/5ms/melspec/']] # Directories for output features, which are [] test_output_dirs = [[]]
3,121
29.910891
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/waveglow-2/main.py
#!/usr/bin/env python """ main.py for project-NN-pytorch/projects The training/inference process wrapper. Dataset API is replaced with NII_MergeDataSetLoader. It is more convenient to train model on corpora stored in different directories. Requires model.py and config.py (config_merge_datasets.py) Usage: $: python main.py [options] """ from __future__ import absolute_import import os import sys import torch import importlib import core_scripts.other_tools.display as nii_warn import core_scripts.data_io.default_data_io as nii_default_dset import core_scripts.data_io.customize_dataset as nii_dset import core_scripts.data_io.conf as nii_dconf import core_scripts.other_tools.list_tools as nii_list_tool import core_scripts.config_parse.config_parse as nii_config_parse import core_scripts.config_parse.arg_parse as nii_arg_parse import core_scripts.op_manager.op_manager as nii_op_wrapper import core_scripts.nn_manager.nn_manager as nii_nn_wrapper import core_scripts.startup_config as nii_startup __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" def main(): """ main(): the default wrapper for training and inference process Please prepare config.py and model.py """ # arguments initialization args = nii_arg_parse.f_args_parsed() # nii_warn.f_print_w_date("Start program", level='h') nii_warn.f_print("Load module: %s" % (args.module_config)) nii_warn.f_print("Load module: %s" % (args.module_model)) prj_conf = importlib.import_module(args.module_config) prj_model = importlib.import_module(args.module_model) # initialization nii_startup.set_random_seed(args.seed, args) use_cuda = not args.no_cuda and torch.cuda.is_available() device = torch.device("cuda" if use_cuda else "cpu") # prepare data io if not args.inference: params = {'batch_size': args.batch_size, 'shuffle': args.shuffle, 'num_workers': args.num_workers, 'sampler': args.sampler} # Load file list and create data loader trn_lst = prj_conf.trn_list trn_set = nii_dset.NII_MergeDataSetLoader( prj_conf.trn_set_name, \ trn_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', params = params, truncate_seq = prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = True, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets) if prj_conf.val_list is not None: val_lst = prj_conf.val_list val_set = nii_dset.NII_MergeDataSetLoader( prj_conf.val_set_name, val_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', \ params = params, truncate_seq= prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets) else: val_set = None # initialize the model and loss function model = prj_model.Model(trn_set.get_in_dim(), \ trn_set.get_out_dim(), \ args, prj_conf, trn_set.get_data_mean_std()) loss_wrapper = prj_model.Loss(args) # initialize the optimizer optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args) # if necessary, resume training if args.trained_model == "": checkpoint = None else: checkpoint = torch.load(args.trained_model) # start training nii_nn_wrapper.f_train_wrapper(args, model, loss_wrapper, device, optimizer_wrapper, trn_set, val_set, checkpoint) # done for traing else: # for inference # default, no truncating, no shuffling params = {'batch_size': args.batch_size, 'shuffle': False, 'num_workers': args.num_workers} if type(prj_conf.test_list) is list: t_lst = prj_conf.test_list else: t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list) test_set = nii_dset.NII_MergeDataSetLoader( prj_conf.test_set_name, \ t_lst, \ prj_conf.test_input_dirs, prj_conf.input_exts, prj_conf.input_dims, prj_conf.input_reso, prj_conf.input_norm, prj_conf.test_output_dirs, prj_conf.output_exts, prj_conf.output_dims, prj_conf.output_reso, prj_conf.output_norm, './', params = params, truncate_seq= None, min_seq_len = None, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets) # initialize model model = prj_model.Model(test_set.get_in_dim(), \ test_set.get_out_dim(), \ args, prj_conf) if args.trained_model == "": print("No model is loaded by ---trained-model for inference") print("By default, load %s%s" % (args.save_trained_name, args.save_model_ext)) checkpoint = torch.load("%s%s" % (args.save_trained_name, args.save_model_ext)) else: checkpoint = torch.load(args.trained_model) # do inference and output data nii_nn_wrapper.f_inference_wrapper(args, model, device, \ test_set, checkpoint) # done return if __name__ == "__main__": main()
6,776
35.047872
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/waveglow-2/model.py
#!/usr/bin/env python """ model.py waveglow is defined in sandbox.block_waveglow. This file wrapps waveglow inside model() so that it can be used by the script. """ from __future__ import absolute_import from __future__ import print_function import sys import time import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import core_scripts.other_tools.debug as nii_debug import core_scripts.other_tools.display as nii_warn import sandbox.block_nn as nii_nn import sandbox.block_waveglow as nii_waveglow import sandbox.util_frontend as nii_nn_frontend __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ################# ## must-have ################# # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) self.input_dim = in_dim self.output_dim = out_dim # a flag for debugging (by default False) self.model_debug = False ################# ## model config ################# # waveform sampling rate self.sample_rate = prj_conf.wav_samp_rate # up-sample rate self.up_sample = prj_conf.input_reso[0] # configuration for WaveGlow # number of waveglow blocks self.num_waveglow_blocks = 3 # number of flow steps in each block self.num_flow_steps_perblock = 4 # number of wavenet layers in one flow step self.num_wn_blocks_perflow = 8 # channel size of the dilated conv in wavenet layer self.num_wn_channel_size = 256 # kernel size of the dilated conv in wavenet layer self.num_wn_conv_kernel = 3 # use affine transformation? (if False, use a + b) self.flag_affine = True # dimension of the early extracted latent z self.early_z_feature_dim = 2 # use legacy implementation of affine (default False) # the difference is the wavenet core to compute the affine # parameter. For more details, check sandbox.block_waveglow self.flag_affine_legacy_implementation = True self.m_waveglow = nii_waveglow.WaveGlow( in_dim, self.up_sample, self.num_waveglow_blocks, self.num_flow_steps_perblock, self.num_wn_blocks_perflow, self.num_wn_channel_size, self.num_wn_conv_kernel, self.flag_affine, self.early_z_feature_dim, self.flag_affine_legacy_implementation) # buffer for noise in inference self.bg_noise = None # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def forward(self, input_feat, wav): """loss = forward(self, input_feat, wav) input ----- input_feat: tensor, input features (batchsize, length1, input_dim) wav: tensor, target waveform (batchsize, length2, 1) it should be raw waveform, flot valued, between (-1, 1) the code will do mu-law conversion output ------ loss: tensor / scalar Note: returned loss can be directly used as the loss value no need to write Loss() """ # normalize conditiona feature #input_feat = self.normalize_input(input_feat) # compute z_bags, neg_logp, logp_z, log_detjac = self.m_waveglow(wav, input_feat) return [[-logp_z, -log_detjac], [True, True]] def inference(self, input_feat): """wav = inference(mels) input ----- input_feat: tensor, input features (batchsize, length1, input_dim) output ------ wav: tensor, target waveform (batchsize, length2, 1) Note: length2 will be = length1 * self.up_sample """ #normalize input #input_feat = self.normalize_input(input_feat) length = input_feat.shape[1] * self.up_sample noise = self.m_waveglow.get_z_noises(length, noise_std=0.6, batchsize=input_feat.shape[0]) output = self.m_waveglow.reverse(noise, input_feat) # use a randomized if self.bg_noise is None: self.bg_noise = self.m_waveglow.reverse(noise, input_feat * 0) # do post filtering output = nii_nn_frontend.spectral_substraction(output,self.bg_noise,1.0) return output # Loss is returned by model.forward(), no need to specify # just a place holder so that the output of model.forward() can be # sent to the optimizer class Loss(): def __init__(self, args): return def compute(self, output, target): return output if __name__ == "__main__": print("Definition of model")
7,095
33.280193
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/waveglow-2/config.py
#!/usr/bin/env python """ config.py This configuration file specifiess the input and output data """ __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2021, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['cmu_all_trn'] val_set_name = ['cmu_all_val'] # for convenience tmp1 = '../DATA/cmu-arctic-data-set' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp1 + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp1 + '/scp/val.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp1 + '/5ms/melspec']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [80] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.mfbsp'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [80] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [True] # Similar configurations for output features output_dirs = [[tmp1 + '/wav_16k_norm']] output_dims = [1] output_exts = ['.wav'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = 16000 # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = 16000 # Other optional arguments, the definition of which depends on a specific model options = {} ######################################################### ## Configuration for inference stage ######################################################### # similar options to training stage test_set_name = ['cmu_all_test_tiny'] # List of test set data # for convenience, you may directly load test_set list here test_list = [['slt_arctic_b0474']] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = [[tmp1 + '/5ms/melspec/']] # Directories for output features, which are [] test_output_dirs = [[]]
3,121
29.910891
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/wavenet/main.py
#!/usr/bin/env python """ main.py for project-NN-pytorch/projects The training/inference process wrapper. Dataset API is replaced with NII_MergeDataSetLoader. It is more convenient to train model on corpora stored in different directories. Requires model.py and config.py (config_merge_datasets.py) Usage: $: python main.py [options] """ from __future__ import absolute_import import os import sys import torch import importlib import core_scripts.other_tools.display as nii_warn import core_scripts.data_io.default_data_io as nii_default_dset import core_scripts.data_io.customize_dataset as nii_dset import core_scripts.data_io.conf as nii_dconf import core_scripts.other_tools.list_tools as nii_list_tool import core_scripts.config_parse.config_parse as nii_config_parse import core_scripts.config_parse.arg_parse as nii_arg_parse import core_scripts.op_manager.op_manager as nii_op_wrapper import core_scripts.nn_manager.nn_manager as nii_nn_wrapper import core_scripts.startup_config as nii_startup __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" def main(): """ main(): the default wrapper for training and inference process Please prepare config.py and model.py """ # arguments initialization args = nii_arg_parse.f_args_parsed() # nii_warn.f_print_w_date("Start program", level='h') nii_warn.f_print("Load module: %s" % (args.module_config)) nii_warn.f_print("Load module: %s" % (args.module_model)) prj_conf = importlib.import_module(args.module_config) prj_model = importlib.import_module(args.module_model) # initialization nii_startup.set_random_seed(args.seed, args) use_cuda = not args.no_cuda and torch.cuda.is_available() device = torch.device("cuda" if use_cuda else "cpu") # prepare data io if not args.inference: params = {'batch_size': args.batch_size, 'shuffle': args.shuffle, 'num_workers': args.num_workers, 'sampler': args.sampler} in_trans_fns = prj_conf.input_trans_fns \ if hasattr(prj_conf, 'input_trans_fns') else None out_trans_fns = prj_conf.output_trans_fns \ if hasattr(prj_conf, 'output_trans_fns') else None # Load file list and create data loader trn_lst = prj_conf.trn_list trn_set = nii_dset.NII_MergeDataSetLoader( prj_conf.trn_set_name, \ trn_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', params = params, truncate_seq = prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = True, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) if prj_conf.val_list is not None: val_lst = prj_conf.val_list val_set = nii_dset.NII_MergeDataSetLoader( prj_conf.val_set_name, val_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', \ params = params, truncate_seq= prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) else: val_set = None # initialize the model and loss function model = prj_model.Model(trn_set.get_in_dim(), \ trn_set.get_out_dim(), \ args, prj_conf, trn_set.get_data_mean_std()) loss_wrapper = prj_model.Loss(args) # initialize the optimizer optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args) # if necessary, resume training if args.trained_model == "": checkpoint = None else: checkpoint = torch.load(args.trained_model) # start training nii_nn_wrapper.f_train_wrapper(args, model, loss_wrapper, device, optimizer_wrapper, trn_set, val_set, checkpoint) # done for traing else: # for inference # default, no truncating, no shuffling params = {'batch_size': args.batch_size, 'shuffle': False, 'num_workers': args.num_workers} in_trans_fns = prj_conf.test_input_trans_fns \ if hasattr(prj_conf, 'test_input_trans_fns') else None out_trans_fns = prj_conf.test_output_trans_fns \ if hasattr(prj_conf, 'test_output_trans_fns') else None if type(prj_conf.test_list) is list: t_lst = prj_conf.test_list else: t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list) test_set = nii_dset.NII_MergeDataSetLoader( prj_conf.test_set_name, \ t_lst, \ prj_conf.test_input_dirs, prj_conf.input_exts, prj_conf.input_dims, prj_conf.input_reso, prj_conf.input_norm, prj_conf.test_output_dirs, prj_conf.output_exts, prj_conf.output_dims, prj_conf.output_reso, prj_conf.output_norm, './', params = params, truncate_seq= None, min_seq_len = None, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) # initialize model model = prj_model.Model(test_set.get_in_dim(), \ test_set.get_out_dim(), \ args, prj_conf) if args.trained_model == "": print("No model is loaded by ---trained-model for inference") print("By default, load %s%s" % (args.save_trained_name, args.save_model_ext)) checkpoint = torch.load("%s%s" % (args.save_trained_name, args.save_model_ext)) else: checkpoint = torch.load(args.trained_model) # do inference and output data nii_nn_wrapper.f_inference_wrapper(args, model, device, \ test_set, checkpoint) # done return if __name__ == "__main__": main()
7,819
35.886792
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/wavenet/model.py
#!/usr/bin/env python """ model.py wavenet is defined in sandbox.block_wavenet, This file wrapps wavenet inside model() so that it can be used by the script. """ from __future__ import absolute_import from __future__ import print_function import sys import time import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import core_scripts.other_tools.debug as nii_debug import core_scripts.other_tools.display as nii_warn import sandbox.block_nn as nii_nn import sandbox.block_dist as nii_dist import sandbox.util_dsp as nii_dsp import sandbox.block_wavenet as nii_wavenet import config as prj_conf __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ############## # class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ################# ## must-have ################# # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) self.input_dim = in_dim self.output_dim = out_dim # a flag for debugging (by default False) self.model_debug = False ################# ## model config ################# # number of bits for mu-law self.num_bits = 10 self.up_sample = prj_conf.input_reso[0] # model # Most of the configurations are fixed in sandbox.block_wavenet.py self.m_wavenet = nii_wavenet.WaveNet_v1(in_dim, self.up_sample, self.num_bits) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def _waveform_encode_target(self, target_wav): return nii_dsp.mulaw_encode(target_wav, self.num_classes) def _waveform_decode_target(self, gen_wav): return nii_dsp.mulaw_decode(gen_wav, self.num_classes) def forward(self, input_feat, wav): """loss = forward(self, input_feat, wav) input ----- input_feat: tensor, input features (batchsize, length1, input_dim) wav: tensor, target waveform (batchsize, length2, 1) it should be raw waveform, flot valued, between (-1, 1) the code will do mu-law conversion output ------ loss: tensor / scalar Note: returned loss can be directly used as the loss value no need to write Loss() """ # normalize input features input_feat = self.normalize_input(input_feat) # compute cross-entropy using wavenet return self.m_wavenet.forward(input_feat, wav) def inference(self, input_feat): """wav = inference(mels) input ----- input_feat: tensor, input features (batchsize, length1, input_dim) output ------ wav: tensor, target waveform (batchsize, length2, 1) Note: length2 will be = length1 * self.up_sample """ # normalize the input input_feat = self.normalize_input(input_feat) # get the output waveform wave = self.m_wavenet.inference(input_feat) return wave # Loss is returned by model.forward(), no need to specify # just a place holder so that the output of model.forward() can be # sent to the optimizer class Loss(): def __init__(self, args): return def compute(self, output, target): return output if __name__ == "__main__": print("Definition of model")
5,750
31.491525
77
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/wavenet/config.py
#!/usr/bin/env python """ config.py This configuration file specifiess the input and output data """ __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2021, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['cmu_all_trn'] val_set_name = ['cmu_all_val'] # for convenience tmp1 = '../DATA/cmu-arctic-data-set' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp1 + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp1 + '/scp/val.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp1 + '/5ms/melspec']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [80] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.mfbsp'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [80] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [True] # Similar configurations for output features output_dirs = [[tmp1 + '/wav_16k_norm']] output_dims = [1] output_exts = ['.wav'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = 16000 * 3 # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = None # Other optional arguments, the definition of which depends on a specific model options = {} ######################################################### ## Configuration for inference stage ######################################################### # similar options to training stage test_set_name = ['cmu_all_test_tiny'] # List of test set data # for convenience, you may directly load test_set list here test_list = [['slt_arctic_b0474']] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = [[tmp1 + '/5ms/melspec/']] # Directories for output features, which are [] test_output_dirs = [[]]
3,124
29.940594
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/ilpcnet/main.py
#!/usr/bin/env python """ main.py for project-NN-pytorch/projects The training/inference process wrapper. Dataset API is replaced with NII_MergeDataSetLoader. It is more convenient to train model on corpora stored in different directories. Requires model.py and config.py (config_merge_datasets.py) Usage: $: python main.py [options] """ from __future__ import absolute_import import os import sys import torch import importlib import core_scripts.other_tools.display as nii_warn import core_scripts.data_io.default_data_io as nii_default_dset import core_scripts.data_io.customize_dataset as nii_dset import core_scripts.data_io.conf as nii_dconf import core_scripts.other_tools.list_tools as nii_list_tool import core_scripts.config_parse.config_parse as nii_config_parse import core_scripts.config_parse.arg_parse as nii_arg_parse import core_scripts.op_manager.op_manager as nii_op_wrapper import core_scripts.nn_manager.nn_manager as nii_nn_wrapper import core_scripts.startup_config as nii_startup __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" def main(): """ main(): the default wrapper for training and inference process Please prepare config.py and model.py """ # arguments initialization args = nii_arg_parse.f_args_parsed() # nii_warn.f_print_w_date("Start program", level='h') nii_warn.f_print("Load module: %s" % (args.module_config)) nii_warn.f_print("Load module: %s" % (args.module_model)) prj_conf = importlib.import_module(args.module_config) prj_model = importlib.import_module(args.module_model) # initialization nii_startup.set_random_seed(args.seed, args) use_cuda = not args.no_cuda and torch.cuda.is_available() device = torch.device("cuda" if use_cuda else "cpu") # prepare data io if not args.inference: params = {'batch_size': args.batch_size, 'shuffle': args.shuffle, 'num_workers': args.num_workers, 'sampler': args.sampler} in_trans_fns = prj_conf.input_trans_fns \ if hasattr(prj_conf, 'input_trans_fns') else None out_trans_fns = prj_conf.output_trans_fns \ if hasattr(prj_conf, 'output_trans_fns') else None # Load file list and create data loader trn_lst = prj_conf.trn_list trn_set = nii_dset.NII_MergeDataSetLoader( prj_conf.trn_set_name, \ trn_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', params = params, truncate_seq = prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = True, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) if prj_conf.val_list is not None: val_lst = prj_conf.val_list val_set = nii_dset.NII_MergeDataSetLoader( prj_conf.val_set_name, val_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', \ params = params, truncate_seq= prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) else: val_set = None # initialize the model and loss function model = prj_model.Model(trn_set.get_in_dim(), \ trn_set.get_out_dim(), \ args, prj_conf, trn_set.get_data_mean_std()) loss_wrapper = prj_model.Loss(args) # initialize the optimizer optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args) # if necessary, resume training if args.trained_model == "": checkpoint = None else: checkpoint = torch.load(args.trained_model) # start training nii_nn_wrapper.f_train_wrapper(args, model, loss_wrapper, device, optimizer_wrapper, trn_set, val_set, checkpoint) # done for traing else: # for inference # default, no truncating, no shuffling params = {'batch_size': args.batch_size, 'shuffle': False, 'num_workers': args.num_workers} in_trans_fns = prj_conf.test_input_trans_fns \ if hasattr(prj_conf, 'test_input_trans_fns') else None out_trans_fns = prj_conf.test_output_trans_fns \ if hasattr(prj_conf, 'test_output_trans_fns') else None if type(prj_conf.test_list) is list: t_lst = prj_conf.test_list else: t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list) test_set = nii_dset.NII_MergeDataSetLoader( prj_conf.test_set_name, \ t_lst, \ prj_conf.test_input_dirs, prj_conf.input_exts, prj_conf.input_dims, prj_conf.input_reso, prj_conf.input_norm, prj_conf.test_output_dirs, prj_conf.output_exts, prj_conf.output_dims, prj_conf.output_reso, prj_conf.output_norm, './', params = params, truncate_seq= None, min_seq_len = None, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns) # initialize model model = prj_model.Model(test_set.get_in_dim(), \ test_set.get_out_dim(), \ args, prj_conf) if args.trained_model == "": print("No model is loaded by ---trained-model for inference") print("By default, load %s%s" % (args.save_trained_name, args.save_model_ext)) checkpoint = torch.load("%s%s" % (args.save_trained_name, args.save_model_ext)) else: checkpoint = torch.load(args.trained_model) # do inference and output data nii_nn_wrapper.f_inference_wrapper(args, model, device, \ test_set, checkpoint) # done return if __name__ == "__main__": main()
7,819
35.886792
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/ilpcnet/block_lpcnet.py
#!/usr/bin/env python """ """ from __future__ import absolute_import import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import torch.nn.init as torch_init import core_scripts.data_io.dsp_tools as nii_dsp_np import sandbox.block_nn as nii_nn import sandbox.block_dist as nii_dist import sandbox.util_dsp as nii_dsp import core_scripts.other_tools.debug as nii_debug __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2021, Xin Wang" ####################################### # Numpy utilities for feature extraction ####################################### def get_excit(wav, frame_length, frame_shift, lpc_order): lpc_handler = nii_dsp_np.LPClite(frame_length, frame_shift, lpc_order) lpc_coef, _, rc, gain, _, _ = lpc_handler.analysis(wav) return [wav, lpc_coef, rc, gain] ####################################### # Network component ####################################### class LPCLitePytorch(torch_nn.Module): """LPCLitePytorch This API is only used to do AR and MA operation, not LPC analysis Example: length = 1000 frame_l = 320 frame_s = 80 wav = torch.randn([1, length, 1]) frame_num = (length - frame_l) // frame_s + 1 lpc_coef = torch.tensor([[1, -0.97]] * frame_num).unsqueeze(0) m_lpc = LPCLitePytorch(320, 80, 2) with torch.no_grad(): # analysis wavp, excit = m_lpc.LPCAnalsysisPytorch(wav, lpc_coef) # synthesis output = torch.zeros_like(wavp) for idx in range(output.shape[1]): lpc_coef_tmp = lpc_coef[:, idx // frame_s, :].unsqueeze(1) output[:, idx:idx+1, :] = m_lpc.LPCSynthesisPytorch( excit[:, idx:idx+1, :], lpc_coef_tmp, idx) print(torch.mean(output[0] - wav[0][0:720])) """ def __init__(self, framelen, frameshift, lpc_order, flag_emph=True, emph_coef=0.97, noise_std=0, noise_bit=8, flag_emph_ignore_start=True, flag_win_begining=True): super(LPCLitePytorch, self).__init__() self.fl = framelen self.fs = frameshift self.lpc_order = lpc_order self.emph_coef = emph_coef self.noise_std = noise_std self.noise_mulaw_q = 2 ** noise_bit self.flag_emph = flag_emph self.flag_emph_ignore_start = flag_emph_ignore_start self.flag_win_begining = flag_win_begining self.m_winbuf = None self.m_buf = None return def _preemphasis(self, wav): """ input ----- wav: tensor, (batch, length, 1) output ------ wav: tensor, (batch, length, 1) """ wav_tmp = torch.zeros_like(wav) + wav wav_tmp[:, 1:] = wav_tmp[:, 1:] - self.emph_coef * wav_tmp[:, :-1] if self.flag_emph_ignore_start: wav_tmp[:, 0] = wav_tmp[:, 1] return wav_tmp def _deemphasis(self, wav): """ input ----- wav: tensor, (batch, length, 1) output ------ wav: tensor, (batch, length, 1) """ out = torch.zeros_like(wav) + wav for idx in range(1, wav.shape[1]): out[:, idx] = out[:, idx] + self.emph_coef * out[:, idx-1] return out def deemphasis(self, wav): """ input ----- wav: tensor, (batch, length, 1) output ------ wav: tensor, (batch, length, 1) """ if self.flag_emph: return self._deemphasis(wav) else: return wav def preemphasis(self, wav): """ input ----- wav: tensor, (batch, length, 1) output ------ wav: tensor, (batch, length, 1) """ if self.flag_emph: return self._preemphasis(wav) else: return wav def LPCAnalysisPytorch(self, wav, lpc_coef, gain=None): """ input ----- wav: tensor, (batch, length, 1) lpc_coef, tensor, (batch, frame, order) gain: tensor, (batch, frame, 1), or None output ------ wavp: tensor, (batch, length2, 1) excit: tensor, (batch, length2, 1) wav_input: tensor, (batch, length2, 1), pre-processed wav Note that length2 depends on the frame length, shift length2 = min( ((length - frame_length) // frame_shift + 1) * frame_shift, lpc_coef.shape[1] * frame_shift) """ if self.flag_emph: wav_input = self._preemphasis(wav) else: wav_input = torch.zeros_like(wav) + wav if self.flag_win_begining: if self.m_winbuf is None: self.m_winbuf = torch.hann_window( self.fl, dtype=wav.dtype, device=wav.device) wav_input[:, 0:self.fl//2, 0] *= self.m_winbuf[0:self.fl//2] if self.noise_std > 0: wav_mulaw_noised = nii_dsp.mulaw_encode( wav, self.noise_mulaw_q, scale_to_int=True).to(dtype=wav.dtype) wav_mulaw_noised += (torch.rand_like(wav) - 0.5) * self.noise_std wav_mulaw_noised = wav_mulaw_noised.clamp(0, self.noise_mulaw_q-1) wav_feedback = nii_dsp.mulaw_decode( wav_mulaw_noised, self.noise_mulaw_q, input_int=True) wav_mulaw_noised_quan = wav_mulaw_noised.to(torch.int64) else: wav_mulaw_noised_quan = None wav_feedback = wav_input # batch = lpc_coef.shape[0] # LPC oroder + 1 poly_order = lpc_coef.shape[2] # take the minimum length wavlen = np.min([((wav.shape[1] - self.fl) // self.fs + 1) * self.fs, lpc_coef.shape[1] * self.fs]) # to pad pad_zero = torch.zeros( [batch, poly_order-1], dtype=wav.dtype, device=wav.device) # (batch, length + poly_order - 1) # flip wavf = [x[n], x[n-1], ..., x[0], 0, 0, 0] wavf = torch.flip( torch.cat([pad_zero, wav_feedback.squeeze(-1)], dim=1), dims=[1]) # (batch, length, poly_order) # unfold [[x[n], x[n-1], x[n-2], ], [x[n-1], x[n-2], x[n-3], ]] # flip back [[x[0], 0, 0...], ..., [x[n], x[n-1], x[n-2], ], ] # wavf[i, n, :] = [wav[n], wav[n-1], wav[n-2], ...m wav[n-order+1]] # this can be used for LPC analysis for the n-th time step wavf = torch.flip(wavf.unfold(dimension=1, size=poly_order, step=1), dims=[1])[:, 0:wavlen, :] # duplicate lpc coefficients for each time step # (batch, length, poly_order) # lpcf[i, n, :] = [1, a_1, ..., a_order] for n-th step in i-th # we need to pad lpcf = lpc_coef.repeat_interleave(self.fs, dim=1)[:, 0:wavlen, :] # excitation # wavf[i, n, :] * lpcf[i, n, :] = wav[n] * 1 + wav[n-1] * a_1 ... excit = torch.sum(wavf * lpcf, dim=-1).unsqueeze(-1) # predicted_wav = wav - excit wavp = wav_input[:, 0:wavlen, :] - excit if gain is not None: gain_tmp = gain.repeat_interleave(self.fs, dim=1)[:, 0:wavlen, :] excit = excit / gain_tmp return wavp, excit, wav_mulaw_noised_quan, wav_input def LPCSynthesisPytorchCore(self, lpc_coef, buf): """ predicted = LPCSynthesisPytorchCore(self, lpc_coef, buf): input ----- lpc_coef: tensor, (batch, 1, poly_order) buf: tensor, (batch, order - 1, 1) output ------ output: tensor, (batch, 1, 1) Compute (x[n-1] * a[2] + ...) for LP synthesis This should be combined with excitation excit[n] - (x[n-1] * a[2] + ...) as the final output waveform """ batch = lpc_coef.shape[0] poly_order = lpc_coef.shape[-1] # (batch, poly_order - 1) # flip so that data in order [x[n-1], ..., x[n-order+1]] pre_sample = torch.flip(buf, dims=[1]).squeeze(-1) # (batch, ) # (x[n-1] * a[2] + ...) predicted = torch.sum(lpc_coef[:, 0, 1:] * pre_sample, dim=1) # (batch, 1, 1) # -(x[n-1] * a[2] + ...) return -1 * predicted.unsqueeze(-1).unsqueeze(-1) def LPCSynthesisPytorchStep(self, excit, lpc_coef, stepidx, gain=None): """ input ----- excit: tensor, (batch, 1, 1) lpc_coef: tensor, (batch, 1, poly_order) stepidx: int, which time step is this gain: tensor, (batch, 1, 1) output ------ output: tensor, (batch, 1, 1) """ batch = lpc_coef.shape[0] poly_order = lpc_coef.shape[-1] # if no buffer is provided. save the output here if stepidx == 0: # save as [n-order-1, ..., n-1] self.m_buf = torch.zeros( [batch, poly_order - 1, 1], dtype=lpc_coef.dtype, device=lpc_coef.device) pre = self.LPCSynthesisPytorchCore(lpc_coef, self.m_buf) # excit[n] + [- (x[n-1] * a[2] + ...)] if gain is None: output = excit + pre else: output = excit * gain + pre # save the previous value # roll and save as [x[n-order+2], ..., x[n-1], x[n]] self.m_buf = torch.roll(self.m_buf, -1, dims=1) self.m_buf[:, -1, :] = output[:, 0, :] return output def LPCSynthesisPytorch(self, excit, lpc_coef, gain=None): """ input ----- wav: tensor, (batch, length, 1) lpc_coef, tensor, (batch, frame, order) gain: tensor, (batch, frame, 1), or None output ------ wavp: tensor, (batch, length2, 1) excit: tensor, (batch, length2, 1) Note that length2 depends on the frame length, shift length2 = min( ((length - frame_length) // frame_shift + 1) * frame_shift, lpc_coef.shape[1] * frame_shift) """ # synthesis output = torch.zeros_like(excit) for idx in range(output.shape[1]): if gain is None: gain_tmp = None else: gain_tmp = gain[:, idx // self.fs, :].unsqueeze(1) output[:, idx:idx+1, :] = self.LPCSynthesisPytorchStep( excit[:, idx:idx+1, :], lpc_coef[:, idx // self.fs, :].unsqueeze(1), idx, gain_tmp) if self.flag_emph: output = self._deemphasis(output) return output def LPC_rc2lpc(self, rc_input): """from reflection coefficients to LPC coefficients Based on numpy version in core_scripts/data_io/dsp_tools.py input ----- rc: tensor, (batch, length, poly_order - 1) output ------ lpc: tensor, (batch, length, poly_order) """ # from (batch, length, poly_order - 1) to (batch * length, poly_order-1) batch, frame_num, order = rc_input.shape rc = rc_input.view(-1, order) # (frame_num, order) frame_num, order = rc.shape polyOrder = order + 1 lpc_coef = torch.zeros([frame_num, 2, polyOrder], dtype=rc_input.dtype, device=rc_input.device) lpc_coef[:, 0, 0] = 1.0 for index in np.arange(1, polyOrder): lpc_coef[:, 1, index] = 1.0 gamma = rc[:, index-1] lpc_coef[:, 1, 0] = -1.0 * gamma if index > 1: lpc_coef[:, 1, 1:index] = lpc_coef[:, 0, 0:index-1] \ + lpc_coef[:, 1, 0:1] * torch.flip(lpc_coef[:, 0, 0:index-1], dims=[1]) lpc_coef[:, 0, :] = lpc_coef[:, 1,:] lpc_coef = torch.flip(lpc_coef[:, 0, :], dims=[1]) return lpc_coef.view(batch, -1, order+1) class CondNetwork(torch_nn.Module): """CondNetwork(input_dim, output_dim) Predict reflection coefficients and gain factor from the input Args ---- input_dim: int, input feature dimension, output_dim: int, output feature dimension, = dimension of reflection coeffcients + 1 for gain """ def __init__(self, input_dim, output_dim): super(CondNetwork, self).__init__() self.m_layer1 = nii_nn.Conv1dKeepLength(input_dim, input_dim, 1, 5) self.m_layer2 = nii_nn.Conv1dKeepLength(input_dim, input_dim, 1, 5) self.m_layer3 = torch_nn.Sequential( nii_nn.GRULayer(input_dim, 64, True), torch_nn.Linear(128, 128), torch_nn.Tanh(), torch_nn.Linear(128, output_dim) ) self.m_act_1 = torch_nn.Tanh() return def forward(self, x): """ rc_coef, gain = CondNetwork(x) input ----- x: tensor, (batch, length, input_dim) output ------ rc_coef: tensor, reflection coefficients, (batch, length, out_dim - 1) gain: tensor, gain, (batch, length, 1) Length denotes the number of frames """ x_tmp = self.m_layer3(self.m_layer2(self.m_layer1(x)) + x) part1, part2 = x_tmp.split([x_tmp.shape[-1]-1, 1], dim=-1) # self.m_act_1(part1) is the predicted reflection coefficients # torch.exp(part2) is the gain. return self.m_act_1(part1), torch.exp(part2) class FrameRateNet(torch_nn.Module): """FrameRateNet(input_dim, output_dim) Args ---- input_dim: int, input feature dimension, output_dim: int, output feature dimension, """ def __init__(self, input_dim, output_dim): super(FrameRateNet, self).__init__() self.m_layer1 = nii_nn.Conv1dKeepLength(input_dim, 128, 1, 3) self.m_layer2 = nii_nn.Conv1dKeepLength(128, 128, 1, 3) self.m_layer3 = torch_nn.Sequential( torch_nn.Linear(128, 128), torch_nn.Tanh(), torch_nn.Linear(128, output_dim), torch_nn.Tanh()) return def forward(self, x): """y = FrameRateNet(x) input ----- x: tensor, (batch, length, input_dim) output ------ y: tensor, (batch, length, out_dim) Length denotes the number of frames """ return self.m_layer3(self.m_layer2(self.m_layer1(x))) class OutputNet(torch_nn.Module): """OutputNet(cond_dim, feedback_dim, out_dim) Args ---- cond_dim: int, dimension of input condition feature feedback_dim: int, dimension of feedback feature out_dim: int, output dimension """ def __init__(self, cond_dim, feedback_dim, out_dim): super(OutputNet, self).__init__() # Use the wrapper of GRULayer in nii_nn self.m_gru1 = nii_nn.GRULayer(cond_dim + feedback_dim, 256) self.m_gru2 = nii_nn.GRULayer(256 + feedback_dim, 16) self.m_outfc = torch_nn.Sequential( torch_nn.utils.weight_norm( torch_nn.Linear(16, 2, bias=False), name='weight') ) self.m_stdact = torch_nn.ReLU() return def _forward(self, cond, feedback): """Forward for training stage """ tmp = self.m_gru1(torch.cat([cond, feedback], dim=-1)) tmp = self.m_gru2(torch.cat([tmp, feedback], dim=-1)) return self.m_outfc(tmp) def _forward_step(self, cond, feedback, stepidx): """Forward for inference stage """ tmp = self.m_gru1(torch.cat([cond, feedback], dim=-1), stepidx) tmp = self.m_gru2(torch.cat([tmp, feedback], dim=-1), stepidx) return self.m_outfc(tmp) def forward(self, cond, feedback, stepidx=None): """mean, std = forward(cond, feedback, stepidx=Non) input ----- cond: tensor, input condition feature, (batch, length, input_dim) feedback: tensor, feedback feature, (batch, length, feedback_dim) stepidx: int or None, the index of the time step, starting from 0 output ------ mean: tensor, mean of the Gaussian dist., (batch, length, out_dim//2) std: tensor, std of the Gaussian dist., (batch, length, out_dim//2) If stepidx is None, length is equal to the waveform sequence length, and the forward() method is in the training mode (self._forward) If stepidx is from 0 to T, length is equal to 1, and forward() only computes output for the t=stepidx. (self._forward_step) The nii_nn.GRULayer will save the hidden states of GRULayer. Note that self._forward_step must be called from stepidx=0 to stepidx=T. """ if stepidx is None: output = self._forward(cond, feedback) else: output = self._forward_step(cond, feedback, stepidx) mean, log_std = output.chunk(2, dim=-1) return mean, torch.exp(self.m_stdact(log_std + 9) - 9) class LPCNetV1(torch_nn.Module): """LPCNetV1(framelen, frameshift, lpc_order, cond_dim, flag_fix_cond) Args ---- framelen: int, frame length for LP ana/syn, in number of waveform points frameshift: int, frame shift for LP ana/syn, in number of waveform points lpc_order: int, LP order cond_dim: input condition feature dimension flag_fix_cond: bool, whether fix the condition network or not In training, we can use a two-staged training approach: train the condition network. first, then fix the condition network and train the iLPCNet. flag_fix_cond is used to indicate the stage """ def __init__(self, framelen, frameshift, lpc_order, cond_dim, flag_fix_cond): super(LPCNetV1, self).__init__() # ===== # options # ===== self.m_fl = framelen self.m_fs = frameshift self.m_lpc_order = lpc_order # dimension of input conditional feature self.m_cond_dim = cond_dim # ===== # hyper-parameters # ===== # dimension of output of frame-rate network self.m_hid_dim = 256 # dimension of waveform self.m_wav_dim = 1 # number of discrete pitch classes # We can directly fed pitch to the network, but in LPCNet the pitch # is quantized and embedded self.m_pitch_cat = 256 self.m_pitch_emb = 64 self.m_emb_f0 = torch_nn.Embedding(self.m_pitch_cat, self.m_pitch_emb) # ===== # network definition # ===== # lpc analyszer self.m_lpc = LPCLitePytorch(framelen, frameshift, lpc_order) # frame rate network self.m_net_framerate = FrameRateNet(cond_dim + self.m_pitch_emb, self.m_hid_dim) # output network self.m_net_out = OutputNet(self.m_hid_dim, self.m_wav_dim, self.m_wav_dim * 2) # condition network to convert conditional feature to rc coef and gain # (although we will not use gain) self.m_cond_net = CondNetwork(cond_dim, lpc_order + 1) # loss for condition network self.m_cond_loss = torch_nn.MSELoss() # fix self.flag_fix_cond = flag_fix_cond return def _negloglikelihood(self, data, mean, std): """ nll = self._negloglikelihood(data, mean, std) neg log likelihood of normal distribution on the data input ----- data: tensor, data, (batch, length, dim) mean: tensor, mean of dist., (batch, length, dim) std: tensor, std of dist., (batch, length, dim) output ------ nll: scalar, neg log likelihood """ return (0.5 * np.log(2 * np.pi) \ + torch.log(std) + 0.5 * (((data - mean)/std) ** 2)).mean() def _convert_pitch(self, pitch_value): """ output = self._convert_pitch(pitch_value) input ----- pitch_value: tensor, any shape output ------ output: tensor in int64, quantized pitch """ return torch.clamp((pitch_value - 33) // 2, 0, self.m_pitch_cat-1).to(torch.int64) def forward(self, cond_feat, cond_feat_normed, lpc_coef, rc_coef, gain, wav): """ loss_wav, loss_cond = forward(cond_feat, cond_feat_normed, lpc_coef, rc_coef, gain, wav) input ----- cond_feat: tensor, condition feature, unnormed (batch, frame_num1, cond_dim) cond_feat_normed: tensor, condition feature, normed (batch, frame_num1, cond_dim) lpc_coef: tensor, LP coefficients, (batch, frame_num2, lpc_order + 1) rc_coef: tensor, reflection coeffs (batch, frame_num2, lpc_order) gain: tensor, gain (batch, frame_num2, 1) wav: tensor, target waveform, (batch, length, 1) output ------ loss_wav: scalar, loss for the waveform modeling (neg log likelihood) loss_cond: scalar, loss for the condition network """ # == step 1 == # network to convert cond_feat predict # cond -> rc_coef # we can compute the loss for condition network if necessary cond_len = np.min([rc_coef.shape[1], cond_feat.shape[1]]) if self.flag_fix_cond: with torch.no_grad(): rc_coef_pre, gain_pre = self.m_cond_net(cond_feat[:, :cond_len]) loss_cond = (self.m_cond_loss(rc_coef, rc_coef_pre) + self.m_cond_loss(gain_pre, gain)) * 0 else: rc_coef_pre, gain_pre = self.m_cond_net(cond_feat[:, :cond_len]) loss_cond = (self.m_cond_loss(rc_coef, rc_coef_pre) + self.m_cond_loss(gain_pre, gain)) # == step 2 == # do LP analysis given the LP coeffs (from predicted reflection coefs) with torch.no_grad(): # convert predicted rc_coef to LP coef lpc_coef_pre = self.m_lpc.LPC_rc2lpc(rc_coef_pre) # LP analysis given LPC coefficients # wav_pre is the LP prediction # wav_err is the LP residual # wav is the pre-processed waveform by the LP analyzer. # here wav_err + wav_pre = wav wav_pre, wav_err, _, wav = self.m_lpc.LPCAnalysisPytorch( wav, lpc_coef_pre) # quantize pitch, we put this line here just for convenience pitch_quantized = self._convert_pitch(cond_feat[:, :cond_len, -1]) # == step 3 == # frame-rate network # embedding F0 pitch_emb = self.m_emb_f0(pitch_quantized) # cond -> feature lpccond_feat = self.m_net_framerate( torch.cat([cond_feat_normed[:, :cond_len, :], pitch_emb], dim=-1)) # duplicate (upsampling) to waveform level lpccond_feat = lpccond_feat.repeat_interleave(self.m_fs, dim=1) # == step 4 == # waveform generation network # # get the minimum length wavlen = np.min([wav.shape[1],wav_pre.shape[1],lpccond_feat.shape[1]]) # feedback waveform (add noise and dropout) 16384 = np.power(2, 14) noise1 = torch.randn_like(wav[:, :wavlen, :]) / 16384 feedback_wav = torch.roll(wav[:, :wavlen, :], 1, dims=1) feedback_wav[:, 0, :] *= 0 feedback_wav += noise1 # compute LP residual mean and std mean, std = self.m_net_out(lpccond_feat[:, :wavlen, :], feedback_wav) # compute wav dist. mean and std mean_wav = mean + wav_pre[:, :wavlen] std_wav = std # get likelihood loss_wav = self._negloglikelihood(wav[:, :wavlen], mean_wav, std_wav) if self.flag_fix_cond: loss_cond = loss_cond * 0 else: loss_wav = loss_wav * 0 return loss_cond, loss_wav def inference(self, cond_feat, cond_feat_normed): """ wav = inference(cond_feat) input ----- cond_feat: tensor, condition feature, unnormed (batch, frame_num1, cond_dim) cond_feat_normed, tensor, condition feature, unnormed (batch, frame_num1, cond_dim) output ------ wav, tensor, (batch, frame_num1 * frame_shift, 1) """ # prepare batch, frame_num, _ = cond_feat.shape xtyp = cond_feat.dtype xdev = cond_feat.device # quantize F0 and F0 embedding pitch_quantized = self._convert_pitch(cond_feat[:, :, -1]) pitch_emb = self.m_emb_f0(pitch_quantized) # == step1. == # predict reflection coeff from cond_feat # rc_coef_pre (batch, frame_num, lpc_order) rc_coef_pre, gain_pre = self.m_cond_net(cond_feat) # == step2. == # from reflection coeff to LPC coef # (batch, frame_num, lpc_order + 1) lpc_coef_pre = self.m_lpc.LPC_rc2lpc(rc_coef_pre) # == step3. == # frame rate network lpccond_feat = self.m_net_framerate( torch.cat([cond_feat_normed, pitch_emb], dim=-1)) # == step4. == # step-by-step generation # waveform length = frame num * up-sampling rate wavlen = frame_num * self.m_fs # many buffers # buf to store output waveofmr wavbuf = torch.zeros([batch, wavlen, 1], dtype=xtyp, device=xdev) # buf to store excitation signal exibuf = torch.zeros_like(wavbuf) # buf to store LPC predicted wave prebuf = torch.zeros_like(wavbuf) # buf to store LPC input x[n-1], ... x[n - poly_order+1] lpcbuf = torch.zeros([batch, self.m_lpc_order, 1], dtype=xtyp, device=xdev) # mean and std buf meanbuf = torch.zeros_like(wavbuf) stdbuf = torch.zeros_like(wavbuf) # loop for idx in range(0, wavlen): if idx % 1000 == 0: print(idx, end=' ', flush=True) frame_idx = idx // self.m_fs # 4.1 LP predicted wav # [- (x[n-1] * a_1 + x[n-2] * a_2 ... )] pre_raw = self.m_lpc.LPCSynthesisPytorchCore( lpc_coef_pre[:, frame_idx : frame_idx+1, :], lpcbuf) # save it (for debugging) prebuf[:, idx:idx+1, :] = pre_raw # 4.2 predict excitation if idx == 0: wav_raw = torch.zeros_like(pre_raw) else: pass # mean, std mean, std = self.m_net_out(lpccond_feat[:, frame_idx:frame_idx+1], wav_raw, idx) meanbuf[:, idx:idx+1, :]= mean stdbuf[:, idx:idx+1, :] = std # sampling exi_raw = torch.randn_like(mean) * std * 0.7 + mean # save excit (for debugging) exibuf[:, idx:idx+1,:] = exi_raw # 4.3 waveform output # excit[n] + [-(x[n-1] * a_1 + x[n-2] * a_2 ... )] wav_raw = exi_raw + pre_raw # save waveform wavbuf[:, idx:idx+1,:] = wav_raw # save it to the LPC buffer. # It will be used by LPCSynthesisPytorchCore lpcbuf = torch.roll(lpcbuf, -1, dims=1) lpcbuf[:, -1, :] = wav_raw[:, 0, :] return self.m_lpc.deemphasis(wavbuf) if __name__ == "__main__": print("Components of LPCNet")
28,731
33.575211
80
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/ilpcnet/model.py
#!/usr/bin/env python """ model.py for iLPCNet version: 1 """ from __future__ import absolute_import from __future__ import print_function import os import sys import time import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import core_scripts.other_tools.debug as nii_debug import core_scripts.other_tools.display as nii_warn import sandbox.block_nn as nii_nn import block_lpcnet as nii_lpcnet __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ######## class Model(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(Model, self).__init__() ################# ## must-have ################# # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) self.input_dim = in_dim self.output_dim = out_dim # a flag for debugging (by default False) self.model_debug = False ################# ## model config ################# # waveform sampling rate self.sample_rate = prj_conf.wav_samp_rate self.fl = prj_conf.options['framelength'] self.fs = prj_conf.options['frameshift'] self.lpc_order = prj_conf.options['lpc_order'] self.feat_reso = prj_conf.input_reso[0] if self.fs % self.feat_reso == 0: if self.fs >= self.feat_reso: self.feat_upsamp = self.fs // self.feat_reso else: print("Adjust condition feature resolution not supported ") sys.exit(1) self.flag_fix_cond = args.temp_flag == 'stage2' self.m_lpcnet = nii_lpcnet.LPCNetV1( self.fl, self.fs, self.lpc_order, in_dim, self.flag_fix_cond) # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ #return (y - self.output_mean) / self.output_std # now the target features will be a list of features return y def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def forward(self, cond_feat, target): """ input ----- output ------ loss: tensor / scalar Note: returned loss can be directly used as the loss value no need to write Loss() """ # wav = target[0] # (batch, frame_num, lpc_order) lpc_coef = target[1] # (batch, framge_num, lpc_order - 1) rc = target[2] # (batch, framge_num, 1) gain = target[3] # condition feature adjust cond_feat_tmp = cond_feat[:, ::self.feat_upsamp] loss_cond, loss_lpcnet = self.m_lpcnet( cond_feat_tmp, self.normalize_input(cond_feat_tmp), lpc_coef, rc, gain, wav) return [[loss_cond, loss_lpcnet], [True, True]] def inference(self, cond_feat): """wav = inference(mels) input ----- output ------ wav_new: tensor, same shape """ # condition feature adjust cond_feat_tmp = cond_feat[:, ::self.feat_upsamp] return self.m_lpcnet.inference( cond_feat_tmp, self.normalize_input(cond_feat_tmp)) # Loss is returned by Model.forward(), no need to specify # just a place holder class Loss(): def __init__(self, args): return def compute(self, output, target): return output if __name__ == "__main__": print("Definition of model")
5,573
28.807487
75
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/ilpcnet/config.py
#!/usr/bin/env python """ config.py To merge different corpora (or just one corpus), *_set_name are lists *_list are lists of lists *_dirs are lists of lists """ __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2020, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['cmu_all_trn'] val_set_name = ['cmu_all_val'] # for convenience tmp1 = '../DATA/cmu-arctic-data-set' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp1 + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp1 + '/scp/val.lst'] # Directories for input features # input_dirs = [[path_of_feature_1, path_of_feature_2, ..., ], # [path_of_feature_1, path_of_feature_2, ..., ], ...] # len(input_dirs) should be equal to len(trn_set_name) and len(val_set_name) # iput_dirs[N] is the path for trn_set_name[N] and val_set_name[N] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp1 + '/5ms/melspec', tmp1 + '/5ms/f0']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [80, 1] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.mfbsp', '.f0'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [80, 80] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [True, True] # Similar configurations for output features output_dirs = [[tmp1 + '/wav_16k_norm']] output_dims = [1] output_exts = ['.wav'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = 2400 # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = 2400 # Other optional arguments, the definition of which depends on a specific model # This may be imported by model.py # # Here, we specify the framelenth, frameshift, and lpc_order for LPC analysis # Note that these configurations can be different from what we used to extract # the input acoustic features, but they should be compatible. # # For example, input acoustic features are extracted with a frameshift of 80 # then, we can use frameshift 160 here. Since frameshift for LPC features # is x2 of input features, the model will down-sample the input features. # See line 148 and 167 in model.py options = {'framelength': 160, 'frameshift': 160, 'lpc_order': 15} # input_trans_fns and output_trans_fns are used by the DataSet. # When loading the data from the disk, the input data will be transformed # by input_trans_fns, output will be transformed by output_trans_fns. # This is done before converting the data into pytorch tensor, which # is more convient. # # They are used by ../../../core_scripts/data_io/default_data_io.py: # f_post_data_process() # If you want to debug into these functions, remember turn off multi workers # $: python -m pdb main.py --num-workers 0 # # These are the rules: # len(input_trans_fns) == len(output_trans_fns) == len(trn_set_name) # input_trans_fns[N] is for trn_set_name[N] and val_set_name[N] # len(input_trans_fns[N]) == len(input_exts) # len(output_trans_fns[N]) == len(output_exts) # input_trans_fns[N][M] is for the input_exts[M] of trn_set_name[N] # ... import block_lpcnet input_trans_fns = [[]] output_trans_fns = [[lambda x: block_lpcnet.get_excit(x, 160, 160, 15)]] ######################################################### ## Configuration for inference stage ######################################################### # similar options to training stage test_set_name = ['cmu_all_test_tiny'] # List of test set data (in the same way as trn_list or val_list) # Or, for convenience, you may directly load test_set list test_list = [['slt_arctic_b0474']] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = [[tmp1 + '/5ms/melspec', tmp1 + '/5ms/f0']] # Directories for output features, which are [[]] test_output_dirs = [[]] # #test_output_trans_fns = output_trans_fns
5,067
34.690141
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/hifigan/main.py
#!/usr/bin/env python """ main.py for project-NN-pytorch/projects The default training/inference process wrapper Requires model.py and config.py Usage: $: python main.py [options] """ from __future__ import absolute_import import os import sys import torch import importlib import core_scripts.data_io.default_data_io as nii_default_dset import core_scripts.data_io.customize_dataset as nii_dset import core_scripts.other_tools.display as nii_warn import core_scripts.data_io.conf as nii_dconf import core_scripts.other_tools.list_tools as nii_list_tool import core_scripts.config_parse.config_parse as nii_config_parse import core_scripts.config_parse.arg_parse as nii_arg_parse import core_scripts.op_manager.op_manager as nii_op_wrapper import core_scripts.nn_manager.nn_manager as nii_nn_wrapper import core_scripts.nn_manager.nn_manager_GAN as nii_nn_wrapper_GAN import core_scripts.startup_config as nii_startup __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2021, Xin Wang" def main(): """ main(): the default wrapper for training and inference process Please prepare config.py and model.py """ # arguments initialization args = nii_arg_parse.f_args_parsed() # nii_warn.f_print_w_date("Start program", level='h') nii_warn.f_print("Load module: %s" % (args.module_config)) nii_warn.f_print("Load module: %s" % (args.module_model)) prj_conf = importlib.import_module(args.module_config) prj_model = importlib.import_module(args.module_model) # initialization nii_startup.set_random_seed(args.seed, args) use_cuda = not args.no_cuda and torch.cuda.is_available() device = torch.device("cuda" if use_cuda else "cpu") # prepare data io if not args.inference: params = {'batch_size': args.batch_size, 'shuffle': args.shuffle, 'num_workers': args.num_workers, 'sampler': args.sampler, 'pin_memory': True} in_trans_fns = prj_conf.input_trans_fns \ if hasattr(prj_conf, 'input_trans_fns') else None out_trans_fns = prj_conf.output_trans_fns \ if hasattr(prj_conf, 'output_trans_fns') else None inout_trans_fns = prj_conf.input_output_trans_fn \ if hasattr(prj_conf, 'input_output_trans_fn') else None # Load file list and create data loader trn_lst = prj_conf.trn_list trn_set = nii_dset.NII_MergeDataSetLoader( prj_conf.trn_set_name, \ trn_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', params = params, truncate_seq = prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = True, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns, inoutput_augment_func = inout_trans_fns) if prj_conf.val_list is not None: val_lst = prj_conf.val_list val_set = nii_dset.NII_MergeDataSetLoader( prj_conf.val_set_name, val_lst, prj_conf.input_dirs, \ prj_conf.input_exts, \ prj_conf.input_dims, \ prj_conf.input_reso, \ prj_conf.input_norm, \ prj_conf.output_dirs, \ prj_conf.output_exts, \ prj_conf.output_dims, \ prj_conf.output_reso, \ prj_conf.output_norm, \ './', \ params = params, truncate_seq= prj_conf.truncate_seq, min_seq_len = prj_conf.minimum_len, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns, inoutput_augment_func = inout_trans_fns) else: val_set = None # initialize the model and loss function model_G = prj_model.ModelGenerator( trn_set.get_in_dim(), trn_set.get_out_dim(), \ args, prj_conf, trn_set.get_data_mean_std()) model_D = prj_model.ModelDiscriminator( trn_set.get_in_dim(), trn_set.get_out_dim(), args, prj_conf, trn_set.get_data_mean_std()) loss_wrapper = None # initialize the optimizer optimizer_G_wrap = nii_op_wrapper.OptimizerWrapper(model_G, args) optimizer_D_wrap = nii_op_wrapper.OptimizerWrapper(model_D, args) # if necessary, resume training if args.trained_model == "": checkpoint_G = None checkpoint_D = None else: tmp_str = args.trained_model.split(",") checkpoint_G = torch.load(tmp_str[0]) if len(tmp_str) > 1: checkpoint_D = torch.load(tmp_str[1]) else: checkpoint_D = None # start training nii_nn_wrapper_GAN.f_train_wrapper_GAN( args, model_G, model_D, loss_wrapper, device, optimizer_G_wrap, optimizer_D_wrap, trn_set, val_set, checkpoint_G, checkpoint_D) # done for traing else: # for inference # default, no truncating, no shuffling params = {'batch_size': args.batch_size, 'shuffle': False, 'num_workers': args.num_workers} in_trans_fns = prj_conf.input_trans_fns \ if hasattr(prj_conf, 'test_input_trans_fns') else None out_trans_fns = prj_conf.output_trans_fns \ if hasattr(prj_conf, 'test_output_trans_fns') else None inout_trans_fns = prj_conf.output_trans_fns \ if hasattr(prj_conf, 'test_input_output_trans_fn') \ else None if type(prj_conf.test_list) is list: t_lst = prj_conf.test_list else: t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list) test_set = nii_dset.NII_MergeDataSetLoader( prj_conf.test_set_name, \ t_lst, \ prj_conf.test_input_dirs, prj_conf.input_exts, prj_conf.input_dims, prj_conf.input_reso, prj_conf.input_norm, prj_conf.test_output_dirs, prj_conf.output_exts, prj_conf.output_dims, prj_conf.output_reso, prj_conf.output_norm, './', params = params, truncate_seq = None, min_seq_len = None, save_mean_std = False, wav_samp_rate = prj_conf.wav_samp_rate, way_to_merge = args.way_to_merge_datasets, global_arg = args, dset_config = prj_conf, input_augment_funcs = in_trans_fns, output_augment_funcs = out_trans_fns, inoutput_augment_func = inout_trans_fns) # initialize model model = prj_model.ModelGenerator( test_set.get_in_dim(), test_set.get_out_dim(), args, prj_conf) if args.trained_model == "": print("Please provide ---trained-model") sys.exit(1) else: checkpoint = torch.load(args.trained_model) # do inference and output data nii_nn_wrapper.f_inference_wrapper( args, model, device, test_set, checkpoint) # done return if __name__ == "__main__": main()
8,338
35.574561
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/hifigan/model.py
#!/usr/bin/env python """ model.py for HiFiGAN HifiGAN is based on code in from https://github.com/jik876/hifi-gan HiFi-GAN: Generative Adversarial Networks for Efficient and High Fidelity Speech Synthesis By Jungil Kong, Jaehyeon Kim, Jaekyoung Bae MIT License Copyright (c) 2020 Jungil Kong Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ from __future__ import absolute_import from __future__ import print_function import sys import numpy as np import torch import torch.nn as torch_nn import torch.nn.functional as torch_nn_func import core_scripts.other_tools.debug as nii_debug import sandbox.util_frontend as nii_frontend __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2021, Xin Wang" ######### ## Loss definition ######### class LossMel(torch_nn.Module): """ Wrapper to define loss function """ def __init__(self, sr): super(LossMel, self).__init__() """ """ # extractir fl = 1024 fs = 256 fn = 1024 num_mel = 80 self.m_frontend = nii_frontend.MFCC( fl, fs, fn, sr, num_mel, with_emphasis=False, with_delta=False, flag_for_MelSpec=True) # loss function self.loss_weight = 45 self.loss = torch_nn.L1Loss() return def forward(self, outputs, target): with torch.no_grad(): # (batch, length, 1) -> (batch, length) -> (batch, length, dim) target_mel = self.m_frontend(target.squeeze(-1)) output_mel = self.m_frontend(outputs.squeeze(-1)) # done return self.loss(output_mel, target_mel) * self.loss_weight ##### ## Model Generator definition ##### from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm def get_padding(kernel_size, dilation=1): # L_out = (L_in + 2*pad - dila * (ker - 1) - 1) // stride + 1 # stride -> 1 # L_out = L_in + 2*pad - dila * (ker - 1) # L_out == L_in -> # 2 * pad = dila * (ker - 1) return int((kernel_size*dilation - dilation)/2) def init_weights(m, mean=0.0, std=0.01): classname = m.__class__.__name__ if classname.find("Conv") != -1: m.weight.data.normal_(mean, std) class ResBlock1(torch_nn.Module): """ """ def __init__(self, channels, kernel_size=3, dilation=(1, 3, 5)): super(ResBlock1, self).__init__() self.leaky_relu_slope = 0.1 self.convs1 = torch_nn.ModuleList([ weight_norm( torch_nn.Conv1d( channels, channels, kernel_size, 1, dilation=dilation[0], padding=get_padding(kernel_size, dilation[0]))), weight_norm( torch_nn.Conv1d( channels, channels, kernel_size, 1, dilation=dilation[1], padding=get_padding(kernel_size, dilation[1]))), weight_norm( torch_nn.Conv1d( channels, channels, kernel_size, 1, dilation=dilation[2], padding=get_padding(kernel_size, dilation[2]))) ]) # initialize the weight self.convs1.apply(init_weights) self.convs2 = torch_nn.ModuleList([ weight_norm( torch_nn.Conv1d( channels, channels, kernel_size, 1, dilation=1, padding=get_padding(kernel_size, 1))), weight_norm( torch_nn.Conv1d( channels, channels, kernel_size, 1, dilation=1, padding=get_padding(kernel_size, 1))), weight_norm( torch_nn.Conv1d( channels, channels, kernel_size, 1, dilation=1, padding=get_padding(kernel_size, 1))) ]) self.convs2.apply(init_weights) return def forward(self, x): for c1, c2 in zip(self.convs1, self.convs2): xt = torch_nn_func.leaky_relu(x, self.leaky_relu_slope) xt = c1(xt) xt = torch_nn_func.leaky_relu(xt, self.leaky_relu_slope) xt = c2(xt) x = xt + x return x def remove_weight_norm(self): for l in self.convs1: remove_weight_norm(l) for l in self.convs2: remove_weight_norm(l) class Generator(torch_nn.Module): """ """ def __init__(self, in_dim, resblock_kernel_sizes, resblock_dilation_sizes, upsample_rates, upsample_kernel_sizes, upsample_init_channel): super(Generator, self).__init__() self.leaky_relu_slope = 0.1 self.resblock_kernel_sizes = resblock_kernel_sizes self.resblock_dilation_sizes = resblock_dilation_sizes self.num_kernels = len(resblock_kernel_sizes) self.upsample_rates = upsample_rates self.upsample_kernel_sizes = upsample_kernel_sizes self.num_upsamples = len(upsample_rates) self.upsample_init_channel = upsample_init_channel self.conv_pre = weight_norm( torch_nn.Conv1d(in_dim, upsample_init_channel, 7, 1, padding=3)) self.ups = torch_nn.ModuleList() for i, (u, k) in enumerate(zip(upsample_rates, upsample_kernel_sizes)): # L_out = (L_in - 1) * stride - 2 * pad + dila * (ker - 1) + 1 # dilation = 1 -> # L_out = (L_in - 1) * stride - 2 * pad + ker # L_out = L_in * stride - 2 * pad + (ker - stride) self.ups.append(weight_norm( torch_nn.ConvTranspose1d( upsample_init_channel//(2**i), upsample_init_channel//(2**(i+1)), k, u, padding=(k-u)//2))) # resblock = ResBlock1 self.resblocks = torch_nn.ModuleList() for i in range(len(self.ups)): ch = upsample_init_channel//(2**(i+1)) for j, (k, d) in enumerate( zip(resblock_kernel_sizes, resblock_dilation_sizes)): self.resblocks.append(resblock(ch, k, d)) self.conv_post = weight_norm(torch_nn.Conv1d(ch, 1, 7, 1, padding=3)) self.ups.apply(init_weights) self.conv_post.apply(init_weights) return def forward(self, x): x = self.conv_pre(x) for i in range(self.num_upsamples): x = torch_nn_func.leaky_relu(x, self.leaky_relu_slope) x = self.ups[i](x) xs = None for j in range(self.num_kernels): if xs is None: xs = self.resblocks[i*self.num_kernels+j](x) else: xs += self.resblocks[i*self.num_kernels+j](x) x = xs / self.num_kernels x = torch_nn_func.leaky_relu(x) x = self.conv_post(x) x = torch.tanh(x) return x def remove_weight_norm(self): print('Removing weight norm...') for l in self.ups: remove_weight_norm(l) for l in self.resblocks: l.remove_weight_norm() remove_weight_norm(self.conv_pre) remove_weight_norm(self.conv_post) class ModelGenerator(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(ModelGenerator, self).__init__() ########## basic config ######## # mean std of input and output in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\ args, mean_std) self.input_mean = torch_nn.Parameter(in_m, requires_grad=False) self.input_std = torch_nn.Parameter(in_s, requires_grad=False) self.output_mean = torch_nn.Parameter(out_m, requires_grad=False) self.output_std = torch_nn.Parameter(out_s, requires_grad=False) self.input_dim = in_dim self.output_dim = out_dim ############################### ###### ## model definition ###### h = prj_conf.options['hifigan_config'] self.m_gen = Generator(in_dim, h['resblock_kernel_sizes'], h['resblock_dilation_sizes'], h['upsample_rates'], h['upsample_kernel_sizes'], h['upsample_initial_channel']) self.m_mel_loss = LossMel(prj_conf.wav_samp_rate) self.flag_removed_weight_norm = False # done return def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None): """ """ if data_mean_std is not None: in_m = torch.from_numpy(data_mean_std[0]) in_s = torch.from_numpy(data_mean_std[1]) out_m = torch.from_numpy(data_mean_std[2]) out_s = torch.from_numpy(data_mean_std[3]) if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim: print("Input dim: {:d}".format(in_dim)) print("Mean dim: {:d}".format(in_m.shape[0])) print("Std dim: {:d}".format(in_s.shape[0])) print("Input dimension incompatible") sys.exit(1) if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim: print("Output dim: {:d}".format(out_dim)) print("Mean dim: {:d}".format(out_m.shape[0])) print("Std dim: {:d}".format(out_s.shape[0])) print("Output dimension incompatible") sys.exit(1) else: in_m = torch.zeros([in_dim]) in_s = torch.ones([in_dim]) out_m = torch.zeros([out_dim]) out_s = torch.ones([out_dim]) return in_m, in_s, out_m, out_s def normalize_input(self, x): """ normalizing the input data """ return (x - self.input_mean) / self.input_std def normalize_target(self, y): """ normalizing the target data """ return (y - self.output_mean) / self.output_std def denormalize_output(self, y): """ denormalizing the generated output from network """ return y * self.output_std + self.output_mean def forward(self, x): if not self.training and not self.flag_removed_weight_norm: self.m_gen.remove_weight_norm() self.flag_removed_weight_norm = True x = self.normalize_input(x) gen_output = self.m_gen(x.permute(0, 2, 1)).permute(0, 2, 1) return gen_output def loss_aux(self, nat_wav, gen_wav, data_in): return self.m_mel_loss(gen_wav, nat_wav) ######### ## Model Discriminator definition ######### class DiscriminatorP(torch_nn.Module): def __init__(self, period, kernel_size=5, stride=3, use_spectral_norm=False): super(DiscriminatorP, self).__init__() self.leaky_relu_slope = 0.1 self.period = period norm_f = weight_norm if use_spectral_norm == False else spectral_norm self.convs = torch_nn.ModuleList([ norm_f( torch_nn.Conv2d(1, 32, (kernel_size, 1), (stride, 1), padding=(get_padding(5, 1), 0))), norm_f( torch_nn.Conv2d(32, 128, (kernel_size, 1), (stride, 1), padding=(get_padding(5, 1), 0))), norm_f( torch_nn.Conv2d(128, 512, (kernel_size, 1), (stride, 1), padding=(get_padding(5, 1), 0))), norm_f( torch_nn.Conv2d(512, 1024, (kernel_size, 1), (stride, 1), padding=(get_padding(5, 1), 0))), norm_f( torch_nn.Conv2d(1024, 1024, (kernel_size, 1), 1, padding=(2, 0))), ]) self.conv_post = norm_f( torch_nn.Conv2d(1024, 1, (3, 1), 1, padding=(1, 0))) return def forward(self, x): fmap = [] # 1d to 2d b, c, t = x.shape if t % self.period != 0: # pad first n_pad = self.period - (t % self.period) x = torch_nn_func.pad(x, (0, n_pad), "reflect") t = t + n_pad x = x.view(b, c, t // self.period, self.period) for l in self.convs: x = l(x) x = torch_nn_func.leaky_relu(x, self.leaky_relu_slope) fmap.append(x) x = self.conv_post(x) fmap.append(x) x = torch.flatten(x, 1, -1) return x, fmap class MultiPeriodDiscriminator(torch_nn.Module): def __init__(self): super(MultiPeriodDiscriminator, self).__init__() self.discriminators = torch_nn.ModuleList([ DiscriminatorP(2), DiscriminatorP(3), DiscriminatorP(5), DiscriminatorP(7), DiscriminatorP(11), ]) def forward(self, y, y_hat): y_d_rs = [] y_d_gs = [] fmap_rs = [] fmap_gs = [] for i, d in enumerate(self.discriminators): y_d_r, fmap_r = d(y) y_d_g, fmap_g = d(y_hat) y_d_rs.append(y_d_r) fmap_rs.append(fmap_r) y_d_gs.append(y_d_g) fmap_gs.append(fmap_g) return y_d_rs, y_d_gs, fmap_rs, fmap_gs class DiscriminatorS(torch_nn.Module): def __init__(self, use_spectral_norm=False): super(DiscriminatorS, self).__init__() self.leaky_relu_slope = 0.1 norm_f = weight_norm if use_spectral_norm == False else spectral_norm self.convs = torch_nn.ModuleList([ norm_f( torch_nn.Conv1d(1, 128, 15, 1, padding=7)), norm_f( torch_nn.Conv1d(128, 128, 41, 2, groups=4, padding=20)), norm_f( torch_nn.Conv1d(128, 256, 41, 2, groups=16, padding=20)), norm_f( torch_nn.Conv1d(256, 512, 41, 4, groups=16, padding=20)), norm_f( torch_nn.Conv1d(512, 1024, 41, 4, groups=16, padding=20)), norm_f( torch_nn.Conv1d(1024, 1024, 41, 1, groups=16, padding=20)), norm_f( torch_nn.Conv1d(1024, 1024, 5, 1, padding=2)), ]) self.conv_post = norm_f(torch_nn.Conv1d(1024, 1, 3, 1, padding=1)) return def forward(self, x): fmap = [] for l in self.convs: x = l(x) x = torch_nn_func.leaky_relu(x, self.leaky_relu_slope) fmap.append(x) x = self.conv_post(x) fmap.append(x) x = torch.flatten(x, 1, -1) return x, fmap class MultiScaleDiscriminator(torch_nn.Module): def __init__(self): super(MultiScaleDiscriminator, self).__init__() self.discriminators = torch_nn.ModuleList([ DiscriminatorS(use_spectral_norm=True), DiscriminatorS(), DiscriminatorS(), ]) self.meanpools = torch_nn.ModuleList([ torch_nn.AvgPool1d(4, 2, padding=2), torch_nn.AvgPool1d(4, 2, padding=2) ]) def forward(self, y, y_hat): y_d_rs = [] y_d_gs = [] fmap_rs = [] fmap_gs = [] for i, d in enumerate(self.discriminators): if i != 0: y = self.meanpools[i-1](y) y_hat = self.meanpools[i-1](y_hat) y_d_r, fmap_r = d(y) y_d_g, fmap_g = d(y_hat) y_d_rs.append(y_d_r) fmap_rs.append(fmap_r) y_d_gs.append(y_d_g) fmap_gs.append(fmap_g) return y_d_rs, y_d_gs, fmap_rs, fmap_gs class ModelDiscriminator(torch_nn.Module): """ Model definition """ def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None): super(ModelDiscriminator, self).__init__() self.m_mpd = MultiPeriodDiscriminator() self.m_msd = MultiScaleDiscriminator() # done return def _feature_loss(self, fmap_r, fmap_g): loss = 0 for dr, dg in zip(fmap_r, fmap_g): for rl, gl in zip(dr, dg): loss += torch.mean(torch.abs(rl - gl)) return loss*2 def _discriminator_loss(self, disc_real_outputs, disc_generated_outputs): loss = 0 r_losses = [] g_losses = [] for dr, dg in zip(disc_real_outputs, disc_generated_outputs): r_loss = torch.mean((1-dr)**2) g_loss = torch.mean(dg**2) loss += (r_loss + g_loss) r_losses.append(r_loss.item()) g_losses.append(g_loss.item()) return loss, r_losses, g_losses def _generator_loss(self, disc_outputs): loss = 0 gen_losses = [] for dg in disc_outputs: l = torch.mean((1-dg)**2) gen_losses.append(l) loss += l return loss, gen_losses def loss_for_D(self, nat_wav, gen_wav_detached, input_feat): # gen_wav has been detached nat_wav_tmp = nat_wav.permute(0, 2, 1) gen_wav_tmp = gen_wav_detached.permute(0, 2, 1) # MPD y_df_hat_r, y_df_hat_g, _, _ = self.m_mpd(nat_wav_tmp, gen_wav_tmp) loss_disc_f, _, _ = self._discriminator_loss(y_df_hat_r, y_df_hat_g) # MSD y_ds_hat_r, y_ds_hat_g, _, _ = self.m_msd(nat_wav_tmp, gen_wav_tmp) loss_disc_s, _, _ = self._discriminator_loss(y_ds_hat_r, y_ds_hat_g) return loss_disc_f + loss_disc_s def loss_for_G(self, nat_wav, gen_wav, input_feat): nat_wav_tmp = nat_wav.permute(0, 2, 1) gen_wav_tmp = gen_wav.permute(0, 2, 1) # MPD y_df_hat_r, y_df_hat_g, fmap_f_r, fmap_f_g = self.m_mpd(nat_wav_tmp, gen_wav_tmp) # MSD y_ds_hat_r, y_ds_hat_g, fmap_s_r, fmap_s_g = self.m_msd(nat_wav_tmp, gen_wav_tmp) loss_fm_f = self._feature_loss(fmap_f_r, fmap_f_g) loss_fm_s = self._feature_loss(fmap_s_r, fmap_s_g) loss_gen_f, _ = self._generator_loss(y_df_hat_g) loss_gen_s, _ = self._generator_loss(y_ds_hat_g) return loss_fm_f + loss_fm_s + loss_gen_f + loss_gen_s if __name__ == "__main__": print("Definition of model")
19,494
33.565603
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/05-nn-vocoders/hifigan/config.py
#!/usr/bin/env python """ config.py This configuration file specifiess the input and output data """ __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2021, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # after data preparation, trn/val_set_name are used to save statistics # about the data sets trn_set_name = ['cmu_all_trn'] val_set_name = ['cmu_all_val'] # for convenience tmp1 = '../DATA/cmu-arctic-data-set' # File lists (text file, one data name per line, without name extension) # trin_file_list: list of files for training set trn_list = [tmp1 + '/scp/train.lst'] # val_file_list: list of files for validation set. It can be None val_list = [tmp1 + '/scp/val.lst'] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory input_dirs = [[tmp1 + '/5ms/melspec']] # Dimensions of input features # input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...] input_dims = [80] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # Please put ".f0" as the last feature input_exts = ['.mfbsp'] # Temporal resolution for input features # input_reso = [reso_feature_1, reso_feature_2, ...] # for waveform modeling, temporal resolution of input acoustic features # may be = waveform_sampling_rate * frame_shift_of_acoustic_features # for example, 80 = 16000 Hz * 5 ms input_reso = [80] # Whether input features should be z-normalized # input_norm = [normalize_feature_1, normalize_feature_2] input_norm = [True] # Similar configurations for output features output_dirs = [[tmp1 + '/wav_16k_norm']] output_dims = [1] output_exts = ['.wav'] output_reso = [1] output_norm = [False] # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None truncate_seq = 16000 # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = 16000 # Other optional arguments, the definition of which depends on a specific model # Here we include the configuration hifigan here # Note that upsample_rates must match with input_reso options = {'hifigan_config': { 'upsample_rates': [5, 4, 2, 2], 'upsample_kernel_sizes': [15, 8, 4, 4], 'upsample_initial_channel': 512, 'resblock_kernel_sizes': [3, 7, 11], 'resblock_dilation_sizes': [[1, 3, 5], [1, 3, 5], [1, 3, 5]] }} ######################################################### ## Configuration for inference stage ######################################################### # similar options to training stage test_set_name = ['cmu_all_test_tiny'] # List of test set data # for convenience, you may directly load test_set list here test_list = [['slt_arctic_b0474']] # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = [[tmp1 + '/5ms/melspec/']] # Directories for output features, which are [] test_output_dirs = [[]]
3,551
31.290909
79
py
project-NN-Pytorch-scripts
project-NN-Pytorch-scripts-master/project/09-asvspoof-vocoded-trn/config_train_toyset_ID_2.py
#!/usr/bin/env python """ config.py for project-NN-pytorch/projects Usage: For training, change Configuration for training stage For inference, change Configuration for inference stage """ import os import pandas as pd __author__ = "Xin Wang" __email__ = "[email protected]" __copyright__ = "Copyright 2022, Xin Wang" ######################################################### ## Configuration for training stage ######################################################### # Name of datasets # this will be used as the name of cache files created for each set # # Name for the seed training set, in case you merge multiple data sets as # a single training set, just specify the name for each subset. # Here we only have 1 training subset trn_set_name = ['asvspoof2019_toyset_vocoded_trn'] val_set_name = ['asvspoof2019_toyset_vocoded_val'] # For convenience, specify a path to the toy data set # because config*.py will be copied into model-*/config_AL_train_toyset/NN # we need to use ../../../ tmp = os.path.dirname(__file__) + '/../../../DATA/toy_example_vocoded' # File list for training and development sets # (text file, one file name per line, without name extension) # we need to provide one lst for each subset # trn_list[n] will correspond to trn_set_name[n] # for training set, baseline method directly load all bonafide and spoofed data trn_list = [tmp + '/scp/train.lst'] # for development set val_list = [tmp + '/scp/dev.lst'] # Directories for input data # We need to provide the path to the directory that saves the input data. # We assume waveforms for training and development of one subset # are stored in the same directory. # Hence, input_dirs[n] is for trn_set_name[n] and val_set_name[n] # # If you need to specify a separate val_input_dirs # val_input_dirs = [[PATH_TO_DEVELOPMENT_SET]] # # Each input_dirs[n] is a list, # for example, input_dirs[n] = [wav, speaker_label, augmented_wav, ...] # # Here, input for each file is a single waveform input_dirs = [[tmp + '/train_dev']] val_input_dirs = [[tmp + '/train_dev']] # Dimensions of input features # What is the dimension of the input feature # len(input_dims) should be equal to len(input_dirs[n]) # # Here, input for each file is a single waveform, dimension is 1 input_dims = [1] # File name extension for input features # input_exts = [name_extention_of_feature_1, ...] # len(input_exts) should be equal to len(input_dirs[n]) # # Here, input file extension is .wav # We use .wav not .flac input_exts = ['.wav'] # Temporal resolution for input features # This is not relevant for CM but for other projects # len(input_reso) should be equal to len(input_dirs[n]) # Here, it is 1 for waveform input_reso = [1] # Whether input features should be z-normalized # This is not relevant for CM but for other projects # len(input_norm) should be equal to len(input_dirs[n]) # Here, it is False for waveform # We don't normalize the waveform input_norm = [False] # Similar configurations for output features # Here, we set output to empty because we will load # the target labels from protocol rather than output feature # '.bin' is also a place holder output_dirs = [[] for x in input_dirs] val_output_dirs = [[]] output_dims = [1] output_exts = ['.bin'] output_reso = [1] output_norm = [False] # === # Waveform configuration # === # Waveform sampling rate # wav_samp_rate can be None if no waveform data is used wav_samp_rate = 16000 # Truncating input sequences so that the maximum length = truncate_seq # When truncate_seq is larger, more GPU mem required # If you don't want truncating, please truncate_seq = None # Here, we don't use truncate_seq included in data_io, but we will do # truncation in data_augmentation functions truncate_seq = None # Minimum sequence length # If sequence length < minimum_len, this sequence is not used for training # minimum_len can be None minimum_len = None # Optional argument # This used to load protocol(s) # Multiple protocol files can be specified in the list # # Note that these protocols should cover all the # training, development, and pool set data. # Otherwise, the code will raise an error # # Here, this protocol will cover all the data in the toy set optional_argument = [tmp + '/protocol.txt'] # === # pre-trained SSL model # === # We will load this pre-trained SSL model as the front-end # # path to the SSL model (it is downloaded by 01_download.sh) ssl_front_end_path = os.path.dirname(__file__) \ + '/../../../SSL_pretrained/xlsr_53_56k.pt' # dimension of the SSL model output # this must be provided. ssl_front_end_out_dim = 1024 # === # data augmentation option # === # for training with aligned bonafide-spoofed mini-batches, # we have to use this customized function to make sure that # we can load the aligned files # We will use function in data_augment.py to process the loaded mini-batch import data_augment # path to the waveform directory (the same as input_dirs above) wav_path = None # path to the protocol of spoofed data ( the same as optional_argument) protocol_path = None # configuration to use Pandas to parse the protocol protocol_cols = None # length to truncate the waveform trim_len = 64000 #### # wrapper of data augmentation functions # the wrapper calls the data_augmentation function defined in # data_augment.py. # these wrapper will be called in data_io when loading the data # from disk #### # wrapper for training set input_trans_fns = [ [lambda x, y: data_augment.wav_aug_wrapper( x, y, wav_samp_rate, trim_len)], [lambda x, y: data_augment.wav_aug_wrapper( x, y, wav_samp_rate, trim_len)]] output_trans_fns = [[], []] # wrapper for development set # development does nothing but simply truncate the waveforms val_input_trans_fns = [ [lambda x, y: data_augment.wav_aug_wrapper_val( x, y, wav_samp_rate, trim_len)], [lambda x, y: data_augment.wav_aug_wrapper_val( x, y, wav_samp_rate, trim_len)]] val_output_trans_fns = [[], []] ######################################################### ## Configuration for inference stage ######################################################### # This part is not used in this project # They are place holders test_set_name = trn_set_name + val_set_name # List of test set data # for convenience, you may directly load test_set list here test_list = trn_list + val_list # Directories for input features # input_dirs = [path_of_feature_1, path_of_feature_2, ..., ] # we assume train and validation data are put in the same sub-directory test_input_dirs = input_dirs * 2 # Directories for output features, which are [] test_output_dirs = [[]] * 2
6,753
31.009479
79
py