from library import loss_triplegan, evaluation import library.loss_cla as loss_classifier FLAGS = flags.FLAGS KEY_ARGUMENTS = config.load_config(FLAGS.config_file) text_logger, MODELS_FOLDER, SUMMARIES_FOLDER = save_context( __file__, KEY_ARGUMENTS) torch.manual_seed(1234) torch.cuda.manual_seed(1235) np.random.seed(1236) torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = True device = torch.device("cuda" if torch.cuda.is_available() else "cpu") itr = inputs.get_data_iter(batch_size=FLAGS.bs_c, subset=1000) itr_u = inputs.get_data_iter(batch_size=FLAGS.bs_c) netG, optim_G = inputs.get_generator_optimizer() netD, optim_D = inputs.get_discriminator_optimizer() netC, optim_c = inputs.get_classifier_optimizer() netG, netD, netC = netG.to(device), netD.to(device), netC.to(device) netG = nn.DataParallel(netG) netD = nn.DataParallel(netD) netC = nn.DataParallel(netC) netC_T, _ = inputs.get_classifier_optimizer() netC_T = netC_T.to(device) netC_T = nn.DataParallel(netC_T) netC.train() netC_T.train() Torture.update_average(netC_T, netC, 0) for p in netC_T.parameters():
FLAGS.device = device = torch.device( "cuda" if torch.cuda.is_available() else "cpu") n_iter_d = 5 if "sngan" in FLAGS.g_model_name else 1 def sigmoid_rampup(global_step, start_iter, end_iter): if global_step < start_iter: return 0.0 rampup_length = end_iter - start_iter cur_ramp = global_step - start_iter cur_ramp = np.clip(cur_ramp, 0, rampup_length) phase = 1.0 - cur_ramp / rampup_length return np.exp(-5.0 * phase * phase) itr = inputs.get_data_iter(batch_size=FLAGS.bs_c, subset=FLAGS.n_labels) # itr_u = inputs.get_data_iter(batch_size=FLAGS.bs_c) netG, optim_G = inputs.get_generator_optimizer() netD, optim_D = inputs.get_discriminator_optimizer() netC, optim_c = inputs.get_classifier_optimizer() netG, netD, netC = netG.to(device), netD.to(device), netC.to(device) netG = nn.DataParallel(netG) netD = nn.DataParallel(netD) netC = nn.DataParallel(netC) netC_T, _ = inputs.get_classifier_optimizer() netC_T = netC_T.to(device) netC_T = nn.DataParallel(netC_T) netC.train() netC_T.train() Torture.update_average(netC_T, netC, 0) for p in netC_T.parameters():
from library.mean_teacher import optim_weight_swa FLAGS = flags.FLAGS KEY_ARGUMENTS = config.load_config(FLAGS.config_file) text_logger, MODELS_FOLDER, SUMMARIES_FOLDER = save_context( __file__, KEY_ARGUMENTS) torch.manual_seed(1234) torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = True np.random.seed(1235) torch.cuda.manual_seed(1236) device = torch.device("cuda" if torch.cuda.is_available() else "cpu") FLAGS.device = device itr = inputs.get_data_iter(subset=FLAGS.n_labels) itr_u = inputs.get_data_iter() netC, optim_c = inputs.get_classifier_optimizer() netC = netC.to(device) netC = nn.DataParallel(netC) netC_T, _ = inputs.get_classifier_optimizer() netC_T = netC_T.to(device) netC_T = nn.DataParallel(netC_T) netC.train() netC_T.train() Torture.update_average(netC_T, netC, 0) for p in netC_T.parameters(): p.requires_grad_(False) if FLAGS.c_step == "ramp_swa": netC_swa, _ = inputs.get_classifier_optimizer() netC_swa = netC_swa.to(device)
torch.manual_seed(1234) torch.cuda.manual_seed(1235) np.random.seed(1236) torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = True device = torch.device("cuda" if torch.cuda.is_available() else "cpu") FLAGS.device = device nlabels = FLAGS.y_dist.dim batch_size = FLAGS.training.batch_size checkpoint_io = CheckpointIO(checkpoint_dir=MODELS_FOLDER) logger = Logger(log_dir=SUMMARIES_FOLDER) itr = inputs.get_data_iter(batch_size) GNet, GOptim = inputs.get_generator_optimizer() DNet, DOptim = inputs.get_discriminator_optimizer() GNet_test = copy.deepcopy(GNet) update_average(GNet_test, GNet, 0.0) ydist = get_ydist(**vars(FLAGS.y_dist)) zdist = get_zdist(**vars(FLAGS.z_dist)) checkpoint_io.register_modules(GNet=GNet, GOptim=GOptim, DNet=DNet, DOptim=DOptim) checkpoint_io.register_modules(GNet_test=GNet_test) trainer_dict = {"baseline": trainer_baseline, "fd": trainer_fd} trainer_used = trainer_dict[FLAGS.trainer.name]
new_dict = pickle.load(f) new_dict["gpu"] = FLAGS.gpu FLAGS.set_dict(new_dict) FLAGS.old_model = "loaded" text_logger, MODELS_FOLDER, SUMMARIES_FOLDER = save_context( __file__, KEY_ARGUMENTS) torch.manual_seed(1234) torch.cuda.manual_seed(1235) np.random.seed(1236) torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = True device = torch.device("cuda" if torch.cuda.is_available() else "cpu") print(device) itr = inputs.get_data_iter(batch_size=100, subset=1000) itr_u = inputs.get_data_iter(batch_size=100) netG, optim_G = inputs.get_generator_optimizer() netD, optim_D = inputs.get_discriminator_optimizer() netC, optim_c = inputs.get_classifier_optimizer() netC_T, _ = inputs.get_classifier_optimizer() netG, netD, netC = netG.to(device), netD.to(device), netC.to(device) netG = nn.DataParallel(netG) netD = nn.DataParallel(netD) netC = nn.DataParallel(netC) netC_T = nn.DataParallel(netC_T) checkpoint_io = Torture.utils.checkpoint.CheckpointIO( checkpoint_dir=MODELS_FOLDER) checkpoint_io.register_modules(
np.random.seed(1236) torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = True device = torch.device("cuda" if torch.cuda.is_available() else "cpu") FLAGS.device = device # model = FLAGS.old_model_c # dirname = os.path.dirname(model) # basename = os.path.basename(model) # config_path = os.path.join(dirname, "..", "source", "configs_dict.pkl") # summary_path = os.path.join(dirname, "..", "summary") # with open(config_path, "rb") as f: # new_dict = pickle.load(f) # FLAGS.set_dict(new_dict) itr = inputs.get_data_iter(batch_size=FLAGS.bs_c, subset=1000) itr_u = inputs.get_data_iter(batch_size=FLAGS.bs_c) itr_test = inputs.get_data_iter(batch_size=FLAGS.bs_c, train=False) # itr_t = inputs.get_data_iter_twice(subset=1000) # itr_ut = inputs.get_data_iter_twice() netG, optim_G = inputs.get_generator_optimizer() netD, optim_D = inputs.get_discriminator_optimizer() netC, optim_c = inputs.get_classifier_optimizer() netC_T, _ = inputs.get_classifier_optimizer() netG, netD, netC = netG.to(device), netD.to(device), netC.to(device) netG = nn.DataParallel(netG) netD = nn.DataParallel(netD) netC = nn.DataParallel(netC) netC_T = nn.DataParallel(netC_T)
new_dict = pickle.load(f) new_dict["gpu"] = FLAGS.gpu FLAGS.set_dict(new_dict) FLAGS.old_model = "loaded" text_logger, MODELS_FOLDER, SUMMARIES_FOLDER = save_context( __file__, KEY_ARGUMENTS) torch.manual_seed(1234) torch.cuda.manual_seed(1235) np.random.seed(1236) torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = True device = torch.device("cuda" if torch.cuda.is_available() else "cpu") print(device) itr = inputs.get_data_iter(batch_size=100, subset=1000) itr_u = inputs.get_data_iter(batch_size=100, infinity=False) netG, optim_G = inputs.get_generator_optimizer() netD, optim_D = inputs.get_discriminator_optimizer() netC, optim_c = inputs.get_classifier_optimizer() netC_T, _ = inputs.get_classifier_optimizer() netG, netD, netC = netG.to(device), netD.to(device), netC.to(device) netG = nn.DataParallel(netG) netC = nn.DataParallel(netC) netC_T = nn.DataParallel(netC_T) checkpoint_io = Torture.utils.checkpoint.CheckpointIO( checkpoint_dir=MODELS_FOLDER) checkpoint_io.register_modules( netG=netG,