Example #1
0
def run():
    LATENT_SIZE = 3

    train_x = np.load("__mnist.npz")['train_x']

    encoder = Input(28 * 28)
    encoder = Tanh(Affine(encoder, 300))
    encoder = Gauss(mean=Affine(encoder, LATENT_SIZE),
                    logstd=Params(LATENT_SIZE))

    dkl = DKLUninormal(mean=encoder.mean, logstd=encoder.logstd)

    decoder_input = Input(LATENT_SIZE)
    decoder = Tanh(Affine(decoder_input, 300))
    decoder = Gauss(mean=Affine(decoder, 28 * 28),
                    logstd=Const(np.zeros(28 * 28) - 3))

    encOptimizer = Adam(encoder.get_params(), horizon=10, lr=0.01)
    decOptimizer = Adam(decoder.get_params(), horizon=10, lr=0.01)

    for i in range(10000):
        idx = np.random.choice(len(train_x), size=128)
        pics = train_x[idx]

        encoder.load_params(encOptimizer.get_value())
        decoder.load_params(decOptimizer.get_value())

        representation, encBackprop = encoder.sample.evaluate(pics)

        picsLogprob, decBackprop = decoder.logprob.evaluate(representation,
                                                            sample=pics)

        dklValue, dklBackprop = dkl.evaluate(pics)

        decOptimizer.apply_gradient(decBackprop(np.ones(128)))

        encOptimizer.apply_gradient(
            dklBackprop(-np.ones(128)) +
            encBackprop(decoder_input.last_gradient))

        print("Logprob:", bar(np.mean(picsLogprob), 20000), "DKL:",
              bar(np.mean(dklValue), 200))

        if i % 100 == 99:
            plt.clf()
            fig, plots = plt.subplots(2)
            changedPic = decoder.mean(representation[43])
            plots[0].imshow(changedPic.reshape(28, 28),
                            cmap="gray",
                            vmin=0,
                            vmax=1)
            plots[1].imshow(pics[43].reshape(28, 28),
                            cmap="gray",
                            vmin=0,
                            vmax=1)
            fig.savefig("step_%05d.png" % (i + 1), dpi=100)

        if i % 1000 == 999:
            np.save("step_%05d_decoder.npy" % (i + 1), decoder.get_params())
Example #2
0
def run():
    train_x = np.load("__mnist.npz")['train_x']

    encoder = Input(28, 28)
    encoder = Tanh(Affine(encoder, 256))
    encoder = Tanh(Affine(encoder, 256))
    encoder = Gauss(
        mean=Affine(encoder, 3, init=0.1),
        logstd=Clip(Affine(encoder, 3), -6.0, 0.0)
    )

    dkl = DKLUninormal(mean=encoder.mean, logstd=encoder.logstd)

    decoder = encoder.sample
    decoder = Tanh(Affine(decoder, 256))
    decoder = Tanh(Affine(decoder, 256))
    decoder = Gauss(
        mean=Affine(decoder, 28, 28, init=0.1),
        logstd=Clip(Affine(decoder, 28, 28), -6.0, 0.0)
    )

    momentum = 0.0

    for i in range(10000):
        inps = train_x[np.random.choice(len(train_x), size=128)]

        logprob, backprop = decoder.logprob.evaluate(inps, sample=inps)
        grad1 = backprop(np.ones(128))

        dkl_value, backprop = dkl.evaluate(inps)
        grad2 = backprop(-np.ones(128))

        grad1[:len(grad2)] += grad2
        momentum = momentum * 0.9 + grad1 * 0.1
        momentum = np.clip(momentum, -1.0, 1.0)
        decoder.load_params(decoder.get_params() + 0.001 * momentum)

        print(
            "Logprob:", bar(np.mean(logprob), 2000.0, length=20),
            "DKL:", bar(np.mean(dkl_value), 200.0, length=20),
        )

        if i % 100 == 99:
            fig, plots = plt.subplots(3, 2)
            for inp, pair in zip(inps, plots):
                for img, plot in zip([inp, decoder.mean(inp)], pair):
                    plot.imshow(img, cmap="gray")
            fig.set_size_inches(4, 6)
            fig.savefig("step_%05d.png"%(i+1), dpi=100)
            plt.close(fig)
Example #3
0
def run():
    LATENT_SIZE = 2

    encoder = Input(gen_traj().size)
    encoder = Tanh(Affine(encoder, 300))
    encoder = Affine(encoder, LATENT_SIZE), Params(LATENT_SIZE)

    dkl = DKLUninormal(mean=encoder[0], logstd=encoder[1])
    encoder = Gauss(mean=encoder[0], logstd=encoder[1])

    decoder_input = Input(LATENT_SIZE)
    decoder = Tanh(Affine(decoder_input, 300))
    decoder = Affine(decoder, gen_traj().size)
    mean_dec = decoder
    decoder = Gauss(mean=decoder)

    encOptimizer = Adam(encoder.get_params(), horizon=10, lr=0.01)
    decOptimizer = Adam(decoder.get_params(), horizon=10, lr=0.01)

    for i in range(10000):
        inps = [gen_traj() for i in range(128)]

        encoder.load_params(encOptimizer.get_value())
        decoder.load_params(decOptimizer.get_value())

        representation, encBackprop = encoder.sample.evaluate(inps)

        inpsLogprob, decBackprop = decoder.logprob.evaluate(representation,
                                                            sample=inps)

        dklValue, dklBackprop = dkl.evaluate(inps)

        decOptimizer.apply_gradient(decBackprop(np.ones(128)))

        encOptimizer.apply_gradient(
            dklBackprop(-np.ones(128)) +
            encBackprop(decoder_input.last_gradient))
        print(decoder_input.last_gradient)

        print("Logprob:", bar(np.mean(inpsLogprob), 20000), "DKL:",
              bar(np.mean(dklValue), 200))

        if i % 25 == 24:
            save_plot(
                "test_%05d.png" % (i + 1),
                mean_dec(np.random.randn(64, LATENT_SIZE)),
                inps,
                mean_dec(representation),
            )
Example #4
0
def run():
    data = np.load("__mnist.npz")
    data = {k: data[k] for k in data}

    model = Input(28, 28)
    for _ in range(2):
        model = LReLU(Affine(model, 128))
    model = Affine(model, 10, init=0.1)

    opt = Adam(model.get_params(), horizon=10, lr=0.001)

    def sgd_step(inps, lbls):
        outs, backprop = model.evaluate(inps)
        grad = lbls - softmax(outs) - outs * 0.01
        opt.apply_gradient(backprop(grad))
        model.load_params(opt.get_value())

    for epoch in range(10):
        for _ in range(len(data["train_x"]) // 128):
            idx = np.random.randint(len(data["train_x"]), size=128)
            sgd_step(data["train_x"][idx], data["train_y"][idx])

        idx = np.random.randint(len(data["test_x"]), size=4096)
        print(
            bar(100.0 *
                accuracy(model(data["test_x"][idx]), data["test_y"][idx])))
Example #5
0
 def print_line(s, r):
     nonlocal n_lines
     if n_lines < 100:
         n_lines += 1
         if "LOG_FILE" in os.environ:
             with open(os.environ["LOG_FILE"], "a") as f:
                 f.write("%d %.2f\n" % (s, r))
                 f.flush()
         else:
             print("%8d steps:" % s, bar(r, max_rew), flush=True)
Example #6
0
from mannequin import bar
from test_setup import timer

for i in range(-1500, 1500, 234):
    print(bar(i * 0.1))

print(bar(1000.0))
print(bar(10000.0))
print(bar(100000.0))
print(bar(1000000.0))

for i in range(-16, 17):
    print(bar(i / 16.0, 1.0, length=2))

print(bar(0.12))
print(bar(0.13))
print(bar(0.37))
print(bar(0.38))

assert timer(print_info=False) < 0.01