def __init__(self, config): self.generator = Generator() self.discriminator = Discriminator() self.flow_encoder = FlowEncoder(flow_type='flownet') self.audio_deri_encoder = AudioDeriEncoder() self.encoder = Encoder() self.encoder.load_state_dict( torch.load( '/mnt/disk0/dat/lchen63/grid/model/model_embedding/encoder_6.pth' )) for param in self.encoder.parameters(): param.requires_grad = False if config.load_model: flownet = flownets() else: flownet = flownets(config.flownet_pth) self.flows_gen = FlowsGen(flownet) print(self.generator) print(self.discriminator) self.bce_loss_fn = nn.BCELoss() self.mse_loss_fn = nn.MSELoss() self.l1_loss_fn = nn.L1Loss() # self.cosine_loss_fn = CosineSimiLoss(config.batch_size) self.cosine_loss_fn = nn.CosineEmbeddingLoss() self.opt_g = torch.optim.Adam( [p for p in self.generator.parameters() if p.requires_grad], lr=config.lr, betas=(config.beta1, config.beta2)) self.opt_d = torch.optim.Adam(self.discriminator.parameters(), lr=config.lr, betas=(config.beta1, config.beta2)) self.opt_corr = torch.optim.Adam(itertools.chain( self.flow_encoder.parameters(), self.audio_deri_encoder.parameters()), lr=config.lr_corr, betas=(0.9, 0.999), weight_decay=0.0005) self.opt_flownet = torch.optim.Adam(self.flows_gen.parameters(), lr=config.lr_flownet, betas=(0.9, 0.999), weight_decay=4e-4) if config.dataset == 'grid': self.dataset = VaganDataset(config.dataset_dir, train=config.is_train) elif config.dataset == 'lrw': self.dataset = LRWdataset(config.dataset_dir, train=config.is_train) self.data_loader = DataLoader(self.dataset, batch_size=config.batch_size, num_workers=config.num_thread, shuffle=True, drop_last=True) self.ones = Variable(torch.ones(config.batch_size)) self.zeros = Variable(torch.zeros(config.batch_size)) self.one_corr = Variable(torch.ones(config.batch_size)) if config.cuda: device_ids = [int(i) for i in config.device_ids.split(',')] self.generator = nn.DataParallel(self.generator.cuda(), device_ids=device_ids) self.discriminator = nn.DataParallel(self.discriminator.cuda(), device_ids=device_ids) self.flow_encoder = nn.DataParallel(self.flow_encoder.cuda(), device_ids=device_ids) self.audio_deri_encoder = nn.DataParallel( self.audio_deri_encoder.cuda(), device_ids=device_ids) self.encoder = nn.DataParallel(self.encoder.cuda(), device_ids=device_ids) self.flows_gen = nn.DataParallel(self.flows_gen.cuda(), device_ids=device_ids) self.bce_loss_fn = self.bce_loss_fn.cuda() self.mse_loss_fn = self.mse_loss_fn.cuda() self.l1_loss_fn = self.l1_loss_fn.cuda() self.cosine_loss_fn = self.cosine_loss_fn.cuda() self.ones = self.ones.cuda(async=True) self.zeros = self.zeros.cuda(async=True) self.one_corr = self.one_corr.cuda(async=True) self.config = config if config.load_model: self.start_epoch = config.start_epoch self.load(config.model_dir, self.start_epoch - 1) else: self.start_epoch = 0
def __init__(self, config): self.generator = Generator() self.discriminator = Discriminator() flownet = flownets(config.flownet_pth) print('flownet_pth: {}'.format(config.flownet_pth)) self.flows_gen = FlowsGen(flownet) print(self.generator) print(self.discriminator) self.bce_loss_fn = nn.BCELoss() self.mse_loss_fn = nn.MSELoss() self.opt_g = torch.optim.Adam(filter(lambda p: p.requires_grad, self.generator.parameters()), lr=config.lr, betas=(config.beta1, config.beta2)) self.opt_d = torch.optim.Adam(self.discriminator.parameters(), lr=config.lr, betas=(config.beta1, config.beta2)) self.opt_flow = torch.optim.Adam(self.flows_gen.parameters(), lr=config.lr_flownet, betas=(0.9, 0.999), weight_decay=4e-4) if config.dataset == 'grid': self.dataset = VaganDataset(config.dataset_dir, train=config.is_train) elif config.dataset == 'lrw': self.dataset = LRWdataset(config.dataset_dir, train=config.is_train) self.data_loader = DataLoader(self.dataset, batch_size=config.batch_size, num_workers=config.num_thread, shuffle=True, drop_last=True) self.ones = Variable(torch.ones(config.batch_size), requires_grad=False) self.zeros = Variable(torch.zeros(config.batch_size), requires_grad=False) device_ids = [int(i) for i in config.device_ids.split(',')] if config.cuda: if len(device_ids) == 1: self.generator = self.generator.cuda(device_ids[0]) self.discriminator = self.discriminator.cuda(device_ids[0]) self.flows_gen = self.flows_gen.cuda(device_ids[0]) else: self.generator = nn.DataParallel(self.generator.cuda(), device_ids=device_ids) self.discriminator = nn.DataParallel(self.discriminator.cuda(), device_ids=device_ids) self.flows_gen = nn.DataParallel(self.flows_gen.cuda(), device_ids=device_ids) self.bce_loss_fn = self.bce_loss_fn.cuda() self.mse_loss_fn = self.mse_loss_fn.cuda() self.ones = self.ones.cuda() self.zeros = self.zeros.cuda() self.config = config self.start_epoch = 0
class Trainer(): def __init__(self, config): self.generator = Generator() self.discriminator = Discriminator() self.flow_encoder = FlowEncoder(flow_type='flownet') self.audio_deri_encoder = AudioDeriEncoder() self.encoder = Encoder() self.encoder.load_state_dict( torch.load( '/mnt/disk0/dat/lchen63/grid/model/model_embedding/encoder_6.pth' )) for param in self.encoder.parameters(): param.requires_grad = False if config.load_model: flownet = flownets() else: flownet = flownets(config.flownet_pth) self.flows_gen = FlowsGen(flownet) print(self.generator) print(self.discriminator) self.bce_loss_fn = nn.BCELoss() self.mse_loss_fn = nn.MSELoss() self.l1_loss_fn = nn.L1Loss() # self.cosine_loss_fn = CosineSimiLoss(config.batch_size) self.cosine_loss_fn = nn.CosineEmbeddingLoss() self.opt_g = torch.optim.Adam( [p for p in self.generator.parameters() if p.requires_grad], lr=config.lr, betas=(config.beta1, config.beta2)) self.opt_d = torch.optim.Adam(self.discriminator.parameters(), lr=config.lr, betas=(config.beta1, config.beta2)) self.opt_corr = torch.optim.Adam(itertools.chain( self.flow_encoder.parameters(), self.audio_deri_encoder.parameters()), lr=config.lr_corr, betas=(0.9, 0.999), weight_decay=0.0005) self.opt_flownet = torch.optim.Adam(self.flows_gen.parameters(), lr=config.lr_flownet, betas=(0.9, 0.999), weight_decay=4e-4) if config.dataset == 'grid': self.dataset = VaganDataset(config.dataset_dir, train=config.is_train) elif config.dataset == 'lrw': self.dataset = LRWdataset(config.dataset_dir, train=config.is_train) self.data_loader = DataLoader(self.dataset, batch_size=config.batch_size, num_workers=config.num_thread, shuffle=True, drop_last=True) self.ones = Variable(torch.ones(config.batch_size)) self.zeros = Variable(torch.zeros(config.batch_size)) self.one_corr = Variable(torch.ones(config.batch_size)) if config.cuda: device_ids = [int(i) for i in config.device_ids.split(',')] self.generator = nn.DataParallel(self.generator.cuda(), device_ids=device_ids) self.discriminator = nn.DataParallel(self.discriminator.cuda(), device_ids=device_ids) self.flow_encoder = nn.DataParallel(self.flow_encoder.cuda(), device_ids=device_ids) self.audio_deri_encoder = nn.DataParallel( self.audio_deri_encoder.cuda(), device_ids=device_ids) self.encoder = nn.DataParallel(self.encoder.cuda(), device_ids=device_ids) self.flows_gen = nn.DataParallel(self.flows_gen.cuda(), device_ids=device_ids) self.bce_loss_fn = self.bce_loss_fn.cuda() self.mse_loss_fn = self.mse_loss_fn.cuda() self.l1_loss_fn = self.l1_loss_fn.cuda() self.cosine_loss_fn = self.cosine_loss_fn.cuda() self.ones = self.ones.cuda(async=True) self.zeros = self.zeros.cuda(async=True) self.one_corr = self.one_corr.cuda(async=True) self.config = config if config.load_model: self.start_epoch = config.start_epoch self.load(config.model_dir, self.start_epoch - 1) else: self.start_epoch = 0 def fit(self): config = self.config configure("{}".format(config.log_dir), flush_secs=5) num_steps_per_epoch = len(self.data_loader) cc = 0 for epoch in range(self.start_epoch, config.max_epochs): for step, (example, real_im, landmarks, right_audio, wrong_audio) in enumerate(self.data_loader): t1 = time.time() if config.cuda: example = Variable(example).cuda(async=True) real_im = Variable(real_im).cuda(async=True) right_audio = Variable(right_audio).cuda(async=True) wrong_audio = Variable(wrong_audio).cuda(async=True) else: example = Variable(example) real_im = Variable(real_im) right_audio = Variable(right_audio) wrong_audio = Variable(wrong_audio) fake_im, _ = self.generator(example, right_audio) # Train the discriminator D_real = self.discriminator(real_im, right_audio) D_wrong = self.discriminator(real_im, wrong_audio) D_fake = self.discriminator(fake_im.detach(), right_audio) loss_real = self.bce_loss_fn(D_real, self.ones) loss_wrong = self.bce_loss_fn(D_wrong, self.zeros) loss_fake = self.bce_loss_fn(D_fake, self.zeros) loss_disc = loss_real + 0.5 * (loss_fake + loss_wrong) loss_disc.backward() self.opt_d.step() self._reset_gradients() # Train the generator fake_im, audio_feature = self.generator(example, right_audio) #####perceptual loss ############## fea_r = self.encoder(real_im)[1] fea_f = self.encoder(fake_im)[1] loss_perceptual = self.mse_loss_fn(fea_f, fea_r) ########################### l1 loss l1_loss_g = self.l1_loss_fn(fake_im, real_im) D_fake = self.discriminator(fake_im, right_audio) real_flows = self.flows_gen(real_im) fake_flows = self.flows_gen(fake_im) f_audio_deri = self.audio_deri_encoder(audio_feature) real_f_flows = self.flow_encoder(real_flows) avg_f_audio_deri = F.avg_pool1d(f_audio_deri, 16, 16).view(-1, 15) real_avg_f_flows = F.avg_pool1d(real_f_flows, 16, 16).view(-1, 15) loss_gen = self.bce_loss_fn(D_fake, self.ones) real_loss_cosine = self.cosine_loss_fn(avg_f_audio_deri, real_avg_f_flows, self.one_corr) loss_g = loss_gen + 0.5 * real_loss_cosine + loss_perceptual + 0.2 * l1_loss_g if config.fake_corr: fake_f_flows = self.flow_encoder(fake_flows) fake_avg_f_flows = F.avg_pool1d(fake_f_flows, 16, 16).view(-1, 15) fake_loss_cosine = self.cosine_loss_fn( avg_f_audio_deri, fake_avg_f_flows, self.one_corr) loss_g += 0.5 * fake_loss_cosine loss_g.backward() self.opt_g.step() self.opt_corr.step() self.opt_flownet.step() self._reset_gradients() t2 = time.time() if (step + 1) % 1 == 0 or (step + 1) == num_steps_per_epoch: steps_remain = num_steps_per_epoch - step + 1 + \ (config.max_epochs - epoch + 1) * num_steps_per_epoch eta = int((t2 - t1) * steps_remain) if not config.fake_corr: print( "[{}/{}][{}/{}] Loss_D: {:.4f} Loss_G: {:.4f}, cosine_loss: {}, percetual_loss: {}, l1_loss: {}, ETA: {} second" .format(epoch + 1, config.max_epochs, step + 1, num_steps_per_epoch, loss_disc.data[0], loss_gen.data[0], real_loss_cosine.data[0], loss_perceptual.data[0], 0.2 * l1_loss_g.data[0], eta)) else: print( "[{}/{}][{}/{}] Loss_D: {:.4f} Loss_G: {:.4f}, cosine_loss: {}, fake_cosine_loss: {}, percetual_loss: {}, l1_loss: {}, ETA: {} second" .format(epoch + 1, config.max_epochs, step + 1, num_steps_per_epoch, loss_disc.data[0], loss_gen.data[0], real_loss_cosine.data[0], fake_loss_cosine.data[0], loss_perceptual.data[0], 0.2 * l1_loss_g.data[0], eta)) log_value('discriminator_loss', loss_disc.data[0], step + num_steps_per_epoch * epoch) log_value('generator_loss', loss_gen.data[0], step + num_steps_per_epoch * epoch) log_value('percetual_loss', loss_perceptual.data[0], step + num_steps_per_epoch * epoch) log_value('l1_loss', 0.2 * l1_loss_g.data[0], step + num_steps_per_epoch * epoch) log_value('real_cosine_loss', real_loss_cosine.data[0], step + num_steps_per_epoch * epoch) if config.fake_corr: log_value('fake_cosine_loss', fake_loss_cosine.data[0], step + num_steps_per_epoch * epoch) if (step) % (num_steps_per_epoch / 10) == 0: fake_store = fake_im.data.permute( 0, 2, 1, 3, 4).contiguous().view(config.batch_size * 16, 3, 64, 64) torchvision.utils.save_image(fake_store, "{}fake_{}.png".format( config.sample_dir, cc), nrow=16, normalize=True) real_store = real_im.data.permute( 0, 2, 1, 3, 4).contiguous().view(config.batch_size * 16, 3, 64, 64) torchvision.utils.save_image(real_store, "{}real_{}.png".format( config.sample_dir, cc), nrow=16, normalize=True) cc += 1 if epoch % 1 == 0: torch.save( self.generator.state_dict(), "{}/generator_{}.pth".format(config.model_dir, epoch)) torch.save( self.discriminator.state_dict(), "{}/discriminator_{}.pth".format(config.model_dir, epoch)) torch.save( self.audio_deri_encoder.state_dict(), "{}/audio_deri_encoder_{}.pth".format( config.model_dir, epoch)) torch.save( self.flow_encoder.state_dict(), "{}/flow_encoder_{}.pth".format(config.model_dir, epoch)) torch.save(self.flows_gen.state_dict(), "{}/flownet_{}.pth".format(config.model_dir, epoch)) def load(self, directory, epoch): gen_path = os.path.join(directory, 'generator_{}.pth'.format(epoch)) disc_path = os.path.join(directory, 'discriminator_{}.pth'.format(epoch)) flow_encoder_path = os.path.join(directory, 'flow_encoder_{}.pth'.format(epoch)) audio_deri_encoder_path = os.path.join( directory, 'audio_deri_encoder_{}.pth'.format(epoch)) flownet_path = os.path.join(directory, 'flownet_{}.pth'.format(epoch)) self.generator.load_state_dict(torch.load(gen_path)) self.discriminator.load_state_dict(torch.load(disc_path)) self.audio_deri_encoder.load_state_dict( torch.load(audio_deri_encoder_path)) self.flow_encoder.load_state_dict(torch.load(flow_encoder_path)) self.flows_gen.load_state_dict(torch.load(flownet_path)) self.start_epoch = int(gen_path.split(".")[0].split("_")[-1]) print("Load pretrained [{}, {}, {}, {}, {}]".format( gen_path, disc_path, audio_deri_encoder_path, flow_encoder_path, flownet_path)) def _reset_gradients(self): self.generator.zero_grad() self.discriminator.zero_grad() self.audio_deri_encoder.zero_grad() self.flow_encoder.zero_grad() self.flows_gen.zero_grad()
class Trainer(): def __init__(self, config): self.generator = Generator() self.discriminator = Discriminator() flownet = flownets(config.flownet_pth) print('flownet_pth: {}'.format(config.flownet_pth)) self.flows_gen = FlowsGen(flownet) print(self.generator) print(self.discriminator) self.bce_loss_fn = nn.BCELoss() self.mse_loss_fn = nn.MSELoss() self.opt_g = torch.optim.Adam(filter(lambda p: p.requires_grad, self.generator.parameters()), lr=config.lr, betas=(config.beta1, config.beta2)) self.opt_d = torch.optim.Adam(self.discriminator.parameters(), lr=config.lr, betas=(config.beta1, config.beta2)) self.opt_flow = torch.optim.Adam(self.flows_gen.parameters(), lr=config.lr_flownet, betas=(0.9, 0.999), weight_decay=4e-4) if config.dataset == 'grid': self.dataset = VaganDataset(config.dataset_dir, train=config.is_train) elif config.dataset == 'lrw': self.dataset = LRWdataset(config.dataset_dir, train=config.is_train) self.data_loader = DataLoader(self.dataset, batch_size=config.batch_size, num_workers=config.num_thread, shuffle=True, drop_last=True) self.ones = Variable(torch.ones(config.batch_size), requires_grad=False) self.zeros = Variable(torch.zeros(config.batch_size), requires_grad=False) device_ids = [int(i) for i in config.device_ids.split(',')] if config.cuda: if len(device_ids) == 1: self.generator = self.generator.cuda(device_ids[0]) self.discriminator = self.discriminator.cuda(device_ids[0]) self.flows_gen = self.flows_gen.cuda(device_ids[0]) else: self.generator = nn.DataParallel(self.generator.cuda(), device_ids=device_ids) self.discriminator = nn.DataParallel(self.discriminator.cuda(), device_ids=device_ids) self.flows_gen = nn.DataParallel(self.flows_gen.cuda(), device_ids=device_ids) self.bce_loss_fn = self.bce_loss_fn.cuda() self.mse_loss_fn = self.mse_loss_fn.cuda() self.ones = self.ones.cuda() self.zeros = self.zeros.cuda() self.config = config self.start_epoch = 0 def fit(self): config = self.config configure("{}/".format(config.log_dir), flush_secs=5) num_steps_per_epoch = len(self.data_loader) cc = 0 for epoch in range(self.start_epoch, config.max_epochs): for step, (example, real_im, landmarks, right_audio, wrong_audio) in enumerate(self.data_loader): t1 = time.time() if config.cuda: example = Variable(example).cuda() real_im = Variable(real_im).cuda() right_audio = Variable(right_audio).cuda() wrong_audio = Variable(wrong_audio).cuda() else: example = Variable(example) real_im = Variable(real_im) right_audio = Variable(right_audio) wrong_audio = Variable(wrong_audio) fake_im = self.generator(example, right_audio) # Train the discriminator real_flows = self.flows_gen(real_im) fake_flows = self.flows_gen(fake_im) D_real = self.discriminator(real_im, real_flows, right_audio) D_wrong = self.discriminator(real_im, real_flows, wrong_audio) D_fake = self.discriminator(fake_im.detach(), fake_flows, right_audio) loss_real = self.bce_loss_fn(D_real, self.ones) loss_wrong = self.bce_loss_fn(D_wrong, self.zeros) loss_fake = self.bce_loss_fn(D_fake, self.zeros) loss_disc = loss_real + 0.5 * (loss_fake + loss_wrong) loss_disc.backward() self.opt_d.step() self.opt_flow.step() self._reset_gradients() # Train the generator fake_im = self.generator(example, right_audio) fake_flows = self.flows_gen(fake_im) D_fake = self.discriminator(fake_im, fake_flows, right_audio) loss_gen = self.bce_loss_fn(D_fake, self.ones) loss_gen.backward() self.opt_g.step() self.opt_flow.step() self._reset_gradients() t2 = time.time() if (step + 1) % 1 == 0 or (step + 1) == num_steps_per_epoch: steps_remain = num_steps_per_epoch - step + 1 + \ (config.max_epochs - epoch + 1) * num_steps_per_epoch eta = int((t2 - t1) * steps_remain) print( "[{}/{}][{}/{}] Loss_D: {:.4f} Loss_G: {:.4f}, ETA: {} second" .format(epoch + 1, config.max_epochs, step + 1, num_steps_per_epoch, loss_disc.data[0], loss_gen.data[0], eta)) log_value('discriminator_loss', loss_disc.data[0], step + num_steps_per_epoch * epoch) log_value('generator_loss', loss_gen.data[0], step + num_steps_per_epoch * epoch) if (step) % (num_steps_per_epoch / 3) == 0: fake_store = fake_im.data.permute( 0, 2, 1, 3, 4).contiguous().view(config.batch_size * 16, 3, 64, 64) torchvision.utils.save_image(fake_store, "{}fake_{}.png".format( config.sample_dir, cc), nrow=16, normalize=True) real_store = real_im.data.permute( 0, 2, 1, 3, 4).contiguous().view(config.batch_size * 16, 3, 64, 64) torchvision.utils.save_image(real_store, "{}real_{}.png".format( config.sample_dir, cc), nrow=16, normalize=True) cc += 1 if epoch % 1 == 0: torch.save( self.generator.state_dict(), "{}/generator_{}.pth".format(config.model_dir, epoch)) torch.save( self.discriminator.state_dict(), "{}/discriminator_{}.pth".format(config.model_dir, epoch)) torch.save( self.flows_gen.state_dict(), "{}/flows_gen_{}.pth".format(config.model_dir, epoch)) def load(self, directory): # paths = glob.glob(os.path.join(directory, "*.pth")) # gen_path = [path for path in paths if "generator" in path][0] # disc_path = [path for path in paths if "discriminator" in path][0] gen_path = '/mnt/disk0/dat/lchen63/grid/model/model_difference/generator_5.pth' disc_path = '/mnt/disk0/dat/lchen63/grid/model/model_difference/discriminator_5.pth' flows_path = '/mnt/disk0/dat/lchen63/grid/model/model_difference/flows_gen_5.pth' print("Loading pretrained [{}, {}, {}]".format(gen_path, disc_path, flows_path)) # gen_state_dict = torch.load(gen_path) # new_gen_state_dict = OrderedDict() # for k, v in gen_state_dict.items(): # name = 'model.' + k # new_gen_state_dict[name] = v # # load params # self.generator.load_state_dict(new_gen_state_dict) # disc_state_dict = torch.load(disc_path) # new_disc_state_dict = OrderedDict() # for k, v in disc_state_dict.items(): # name = 'model.' + k # new_disc_state_dict[name] = v # # load params # self.discriminator.load_state_dict(new_disc_state_dict) self.generator.load_state_dict(torch.load(gen_path)) self.discriminator.load_state_dict(torch.load(disc_path)) self.flows_gen.load_state_dict(torch.load(flows_path)) self.start_epoch = int(gen_path.split(".")[0].split("_")[-1]) print self.start_epoch def _reset_gradients(self): self.generator.zero_grad() self.discriminator.zero_grad() self.flows_gen.zero_grad()