def test_random_stretch_squeeze(self): test_utils.set_seed(self.seed) audio = np.zeros(self.input_shape) audio[:, 2:5,] = 1 inputs = tf.keras.layers.Input( shape=self.input_shape[1:], batch_size=self.input_shape[0], dtype=tf.float32) outputs = random_stretch_squeeze.RandomStretchSqueeze( resample_offset=0.5, seed=self.seed)( inputs, training=True) model = tf.keras.models.Model(inputs, outputs) prediction = model.predict(audio) # confirm that data are squeezed target0 = np.array([0., 0., 1., 1., 0., 0., 0.]) self.assertAllClose(prediction[0, :], target0) # confirm that data are stretched target1 = np.array([0., 0.44444, 1., 1., 1., 0.44444, 0.]) self.assertAllClose(prediction[1, :], target1, atol=1e-4)
def build(self, input_shape): super(SpeechFeatures, self).build(input_shape) if self.params[ 'sp_time_shift_samples'] != 0.0 and self.mode == modes.Modes.TRAINING: self.rand_shift = random_shift.RandomShift( self.params['sp_time_shift_samples']) else: self.rand_shift = tf.keras.layers.Lambda(lambda x: x) if self.params[ 'sp_resample'] != 0.0 and self.mode == modes.Modes.TRAINING: self.rand_stretch_squeeze = random_stretch_squeeze.RandomStretchSqueeze( self.params['sp_resample']) else: self.rand_stretch_squeeze = tf.keras.layers.Lambda(lambda x: x) self.data_frame = data_frame.DataFrame( mode=self.mode, inference_batch_size=self.inference_batch_size, frame_size=self.frame_size, frame_step=self.frame_step) if self.noise_scale != 0.0 and self.mode == modes.Modes.TRAINING: self.add_noise = tf.keras.layers.GaussianNoise( stddev=self.noise_scale) else: self.add_noise = tf.keras.layers.Lambda(lambda x: x) if self.params['preemph'] != 0.0: self.preemphasis = preemphasis.Preemphasis( preemph=self.params['preemph']) else: self.preemphasis = tf.keras.layers.Lambda(lambda x: x) if self.params['window_type'] is not None: self.windowing = windowing.Windowing( window_size=self.frame_size, window_type=self.params['window_type']) else: self.windowing = tf.keras.layers.Lambda(lambda x: x) # If use_tf_fft is False, we will use # Real Discrete Fourier Transformation(RDFT), which is slower than RFFT # To increase RDFT efficiency we use properties of mel spectrum. # We find a range of non zero values in mel spectrum # and use it to compute RDFT: it will speed up computations. # If use_tf_fft is True, then we use TF RFFT which require # signal length alignment, so we disable mel_non_zero_only. self.mag_rdft_mel = magnitude_rdft_mel.MagnitudeRDFTmel( use_tf_fft=self.params['use_tf_fft'], magnitude_squared=self.params['fft_magnitude_squared'], num_mel_bins=self.params['mel_num_bins'], lower_edge_hertz=self.params['mel_lower_edge_hertz'], upper_edge_hertz=self.params['mel_upper_edge_hertz'], sample_rate=self.params['sample_rate'], mel_non_zero_only=self.params['mel_non_zero_only']) self.log_max = tf.keras.layers.Lambda(lambda x: tf.math.log( tf.math.maximum(x, self.params['log_epsilon']))) if self.params['dct_num_features'] != 0: self.dct = dct.DCT(num_features=self.params['dct_num_features']) else: self.dct = tf.keras.layers.Lambda(lambda x: x) self.normalizer = normalizer.Normalizer(mean=self.mean, stddev=self.stddev) # in any inference mode there is no need to add dynamic logic in tf graph if self.params[ 'use_spec_augment'] and self.mode == modes.Modes.TRAINING: self.spec_augment = spectrogram_augment.SpecAugment( time_masks_number=self.params['time_masks_number'], time_mask_max_size=self.params['time_mask_max_size'], frequency_masks_number=self.params['frequency_masks_number'], frequency_mask_max_size=self.params['frequency_mask_max_size']) else: self.spec_augment = tf.keras.layers.Lambda(lambda x: x) if self.params['use_spec_cutout'] and self.mode == modes.Modes.TRAINING: self.spec_cutout = spectrogram_cutout.SpecCutout( masks_number=self.params['spec_cutout_masks_number'], time_mask_size=self.params['spec_cutout_time_mask_size'], frequency_mask_size=self. params['spec_cutout_frequency_mask_size']) else: self.spec_cutout = tf.keras.layers.Lambda(lambda x: x)