Esempio n. 1
0
                                                     newClass)
    vaX, vaY, vaidx = load_fashion_mnist_classSelect('val', class_use,
                                                     newClass)
    teX, teY, teidx = load_fashion_mnist_classSelect('test', class_use,
                                                     newClass)
else:
    print('dataset must be ' 'mnist' ' or ' 'fmnist' '!')

# --- train ---
batch_idxs = len(trX) // batch_size
batch_idxs_val = len(vaX) // test_size
ce_loss = nn.CrossEntropyLoss()
#
from models.CNN_classifier import CNN

classifier = CNN(y_dim).to(device)
optimizer = torch.optim.SGD(classifier.parameters(), lr=lr, momentum=momentum)
scheduler = StepLR(optimizer, step_size=1, gamma=gamma)
#
loss_total = np.zeros((epochs * batch_idxs))
test_loss_total = np.zeros((epochs))
percent_correct = np.zeros((epochs))
start_time = time.time()
counter = 0
for epoch in range(0, epochs):
    for idx in range(0, batch_idxs):
        batch_labels = torch.from_numpy(trY[idx * batch_size:(idx + 1) *
                                            batch_size]).long().to(device)
        batch_images = trX[idx * batch_size:(idx + 1) * batch_size]
        batch_images_torch = torch.from_numpy(batch_images)
        batch_images_torch = batch_images_torch.permute(0, 3, 1, 2).float()
Esempio n. 2
0
def train_explainer(dataset, classes_used, K, L, lam, print_train_losses=True):
    # --- parameters ---
    # dataset
    data_classes_lst = [int(i) for i in str(classes_used)]
    dataset_name_full = dataset + '_' + str(classes_used)

    # classifier
    classifier_path = './pretrained_models/{}_classifier'.format(
        dataset_name_full)

    # GCE params
    randseed = 0
    gce_path = os.path.join(
        'outputs', dataset_name_full +
        '_gce_K{}_L{}_lambda{}'.format(K, L,
                                       str(lam).replace('.', "")))
    retrain_gce = True  # train explanatory VAE from scratch
    save_gce = True  # save/overwrite pretrained explanatory VAE at gce_path

    # other train params
    train_steps = 3000  #8000
    Nalpha = 25
    Nbeta = 70
    batch_size = 64
    lr = 5e-4

    # --- initialize ---
    device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
    if randseed is not None:
        np.random.seed(randseed)
        torch.manual_seed(randseed)
    ylabels = range(0, len(data_classes_lst))

    # --- load data ---
    from load_mnist import load_mnist_classSelect, load_fashion_mnist_classSelect

    if dataset == 'mnist':
        fn = load_mnist_classSelect
    elif dataset == 'fmnist':
        fn = load_fashion_mnist_classSelect
    elif dataset == 'cifar':
        fn = load_cifar_classSelect
    else:
        print('dataset not correctly specified')

    X, Y, tridx = fn('train', data_classes_lst, ylabels)
    vaX, vaY, vaidx = fn('val', data_classes_lst, ylabels)

    if dataset == "cifar":
        X, vaX = X / 255, vaX / 255

    ntrain, nrow, ncol, c_dim = X.shape
    x_dim = nrow * ncol

    # --- load classifier ---
    from models.CNN_classifier import CNN
    classifier = CNN(len(data_classes_lst), c_dim).to(device)
    checkpoint = torch.load('%s/model.pt' % classifier_path,
                            map_location=device)
    classifier.load_state_dict(checkpoint['model_state_dict_classifier'])

    # --- train/load GCE ---
    from models.CVAEImageNet import Decoder, Encoder
    if retrain_gce:
        encoder = Encoder(K + L, c_dim, x_dim).to(device)
        decoder = Decoder(K + L, c_dim, x_dim).to(device)
        encoder.apply(util.weights_init_normal)
        decoder.apply(util.weights_init_normal)
        gce = GenerativeCausalExplainer(classifier,
                                        decoder,
                                        encoder,
                                        device,
                                        save_output=True,
                                        save_model_params=False,
                                        save_dir=gce_path,
                                        debug_print=print_train_losses)
        traininfo = gce.train(X,
                              K,
                              L,
                              steps=train_steps,
                              Nalpha=Nalpha,
                              Nbeta=Nbeta,
                              lam=lam,
                              batch_size=batch_size,
                              lr=lr)
        if save_gce:
            if not os.path.exists(gce_path):
                os.makedirs(gce_path)
            torch.save(gce, os.path.join(gce_path, 'model.pt'))
            sio.savemat(
                os.path.join(gce_path, 'training-info.mat'), {
                    'data_classes_lst': data_classes_lst,
                    'classifier_path': classifier_path,
                    'K': K,
                    'L': L,
                    'train_step': train_steps,
                    'Nalpha': Nalpha,
                    'Nbeta': Nbeta,
                    'lam': lam,
                    'batch_size': batch_size,
                    'lr': lr,
                    'randseed': randseed,
                    'traininfo': traininfo
                })
    else:  # load pretrained model
        gce = torch.load(os.path.join(gce_path, 'model.pt'),
                         map_location=device)
        traininfo = None

    # --- compute final information flow ---
    I = gce.informationFlow()
    Is = gce.informationFlow_singledim(range(0, K + L))
    print('Information flow of K=%d causal factors on classifier output:' % K)
    print(Is[:K])
    print('Information flow of L=%d noncausal factors on classifier output:' %
          L)
    print(Is[K:])

    # --- generate explanation and create figure ---
    nr_labels = len(data_classes_lst)
    nr_samples_fig = 8
    sample_ind = np.empty(0, dtype=int)

    # retrieve samples from each class
    samples_per_class = math.ceil(nr_samples_fig / nr_labels)
    for i in range(nr_labels):
        samples_per_class = math.ceil(
            (nr_samples_fig - i * samples_per_class) / (nr_labels - i))
        sample_ind = np.int_(
            np.concatenate(
                [sample_ind,
                 np.where(vaY == i)[0][:samples_per_class]]))

    x = torch.from_numpy(vaX[sample_ind])
    zs_sweep = [-3., -2., -1., 0., 1., 2., 3.]
    Xhats, yhats = gce.explain(x, zs_sweep)
    plot_save_dir = os.path.join(gce_path, 'figs/')
    if not os.path.exists(plot_save_dir):
        os.makedirs(plot_save_dir)
    plotting.plotExplanation(1. - Xhats, yhats, save_path=plot_save_dir)

    return traininfo
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
if randseed is not None:
    np.random.seed(randseed)
    torch.manual_seed(randseed)
ylabels = range(0, len(data_classes))

# --- load data ---
from load_mnist import load_mnist_classSelect
X, Y, tridx = load_mnist_classSelect('train', data_classes, ylabels)
vaX, vaY, vaidx = load_mnist_classSelect('val', data_classes, ylabels)
ntrain, nrow, ncol, c_dim = X.shape
x_dim = nrow * ncol

# --- load classifier ---
from models.CNN_classifier import CNN
classifier = CNN(len(data_classes)).to(device)
checkpoint = torch.load('%s/model.pt' % classifier_path, map_location=device)
classifier.load_state_dict(checkpoint['model_state_dict_classifier'])

# --- train/load GCE ---
from models.CVAE import Decoder, Encoder
if retrain_gce:
    encoder = Encoder(K + L, c_dim, x_dim).to(device)
    decoder = Decoder(K + L, c_dim, x_dim).to(device)
    encoder.apply(util.weights_init_normal)
    decoder.apply(util.weights_init_normal)
    gce = GenerativeCausalExplainer(classifier, decoder, encoder, device)
    traininfo = gce.train(X,
                          K,
                          L,
                          steps=train_steps,
vaX, vaY, va_idx = load_mnist_classSelect('val', class_use, newClass)
trX_3ch = np.tile(trX, (1, 1, 1, 3))
vaX_3ch = np.tile(vaX, (1, 1, 1, 3))
sample_inputs = vaX[0:test_size]
sample_inputs_torch = torch.from_numpy(sample_inputs)
sample_inputs_torch = sample_inputs_torch.permute(0, 3, 1,
                                                  2).float().to(device)
ntrain = trX.shape[0]
# data sample to provide local explanation for
x3 = vaX[np.where(1 - vaY)[0][0]]
x8 = vaX[np.where(vaY)[0][0]]

# --- load trained classifier ---
from models.CNN_classifier import CNN

classifier = CNN(y_dim).to(device)
batch_orig = 64
checkpoint = torch.load(classifier_save_dir, map_location=device)
classifier.load_state_dict(checkpoint['model_state_dict_classifier'])

trYhat = classifier(torch.from_numpy(trX).permute(
    0, 3, 1, 2).float())[0].detach().numpy()
vaYhat = classifier(torch.from_numpy(vaX).permute(
    0, 3, 1, 2).float())[0].detach().numpy()
classifier_accuracy_train = np.sum(np.round(trYhat[:, 1]) == trY) / len(trY)
classifier_accuracy_val = np.sum(np.round(vaYhat[:, 1]) == vaY) / len(vaY)

# --- generate integrated gradients explanation ---
"""
Compute integrated gradients explanation
INPUTS