Esempio n. 1
0
def FRQ_2():
    train_features, test_features, train_targets, test_targets = \
    load_data('synthetic', fraction=1.0)
    bias_arr = [0.5, -0.5, -1.5, -2.5, -3.5, -4.5,-5.5]

    indices = [0, 1, 4, 5]
    train_features = np.array(train_features)[indices[:]]
    train_targets = np.array(train_targets)[indices[:]]

    train_features = np.append(train_features, np.ones((train_features.shape[0],1)), axis=1)
    zero_one_loss = ZeroOneLoss()

    print("features: ", train_features)
    print("targets: ", train_targets)

    loss_arr = []
    for i_bias in range(len(bias_arr)):
        w = np.append( np.ones(train_features.shape[1] - 1), bias_arr[i_bias])
        loss = zero_one_loss.forward(train_features, w, train_targets)
        loss_arr.append(loss)

    plt.plot(bias_arr, loss_arr)
    plt.title("loss landscape")
    plt.xlabel("bias")
    plt.ylabel("loss")
    plt.show()
Esempio n. 2
0
def FRQ_1():
    train_features, test_features, train_targets, test_targets = \
        load_data('mnist-binary', fraction=1.0)
    learner = GradientDescent( learning_rate=1e-4, loss='hinge')
    # print ("size: ", train_features.shape)
    # print (train_targets.shape)
    learner.fit(train_features, train_targets,  batch_size=1, max_iter=1000)
Esempio n. 3
0
def test_gradient_descent_blobs():
    """
    Tests the ability of the gradient descent algorithm to classify a linearly
    separable dataset.
    """
    features, _, targets, _ = load_data('blobs')

    hinge = make_predictions(features, targets, 'hinge', None)
Esempio n. 4
0
def FRQ_3():
    train_features, test_features, train_targets, test_targets = \
    load_data('mnist-multiclass', fraction=0.75)
    learner = MultiClassGradientDescent(loss='squared', regularization='l1')
    learner.fit(train_features, train_targets, batch_size=20, max_iter=700)
    predictions = learner.predict(test_features)

    print("predictionsL: ", predictions.shape)
    print("test_targets: ", test_targets.shape)
    c_m = confusion_matrix(test_targets, predictions.astype(int))
    visualized_c_m = np.append(np.arange(-1,5).reshape(1,6), np.append(np.arange(5).reshape(5,1), c_m, axis = 1),  axis=0 )
    print("confusion matrix: ", visualized_c_m)
Esempio n. 5
0
def test_multiclass_gradient_descent_blobs():
    """
    Tests that the multiclass classifier also works on binary tasks
    """
    from your_code import MultiClassGradientDescent

    np.random.seed(0)

    features, _, targets, _ = load_data('blobs')

    learner = MultiClassGradientDescent(loss='squared', regularization=None,
                                        learning_rate=0.01, reg_param=0.05)
    learner.fit(features, targets, batch_size=None, max_iter=1000)
    predictions = learner.predict(features)

    print("predictions: ", predictions)
    print("targets: ", targets)
def test_gradient_descent_mnist_binary():
    """
    Tests the ability of the gradient descent classifier to classify a
    non-trivial problem with a reasonable accuracy.
    """
    from your_code import GradientDescent, accuracy

    train_features, test_features, train_targets, test_targets = \
        load_data('mnist-binary', fraction=0.8)

    np.random.seed(0)
    learner = GradientDescent(loss='squared',
                              regularization=None,
                              learning_rate=0.01,
                              reg_param=0.05)
    learner.fit(train_features, train_targets, batch_size=None, max_iter=1000)
    predictions = learner.predict(test_features)

    assert accuracy(test_targets, predictions) > 0.97
Esempio n. 7
0
def FRQ_4_c():
    train_features, test_features, train_targets, test_targets = \
    load_data('mnist-binary', fraction=1.0)
    batch_size = 50
    max_iter = 2000
    l1_learner = GradientDescent( loss='squared', learning_rate=1e-5, regularization='l1', reg_param=1)
    l1_iter = l1_learner.fit(train_features, train_targets, batch_size=batch_size, max_iter=max_iter)

    data = np.copy(l1_learner.model)
    data = np.delete(data, -1)
    for i in range(len(data)):
        if abs(data[i])<0.001:
            data[i] = 0
        else:
            data[i] = 1
    data = data.reshape(28,28)

    plt.imshow(data,cmap='Greys')
    plt.title("heat map")
    plt.show()
Esempio n. 8
0
def FRQ_4():
    train_features, test_features, train_targets, test_targets = \
    load_data('mnist-binary', fraction=1.0)
    print ("# of 0: ", np.sum([1 for target in train_targets if target==-1]))
    print ("# of 1: ", np.sum([1 for target in train_targets if target==1]))

    lambda_arr = [1e-3, 1e-2, 1e-1, 1, 10, 100]
    l1_iter_arr = []
    l2_iter_arr = []
    l1_non_0_num_arr = []
    l2_non_0_num_arr = []

    batch_size = 50
    max_iter = 2000
    for _lambda in lambda_arr:
        l1_learner = GradientDescent( loss='squared', learning_rate=1e-5, regularization='l1', reg_param=_lambda)
        l2_learner = GradientDescent( loss='squared', learning_rate=1e-5, regularization='l2', reg_param=_lambda)

        l1_iter = l1_learner.fit(train_features, train_targets, batch_size=batch_size, max_iter=max_iter)
        l2_iter = l2_learner.fit(train_features, train_targets, batch_size=batch_size, max_iter=max_iter)

        l1_iter_arr.append(l1_iter)
        l2_iter_arr.append(l2_iter)

        l1_non_0_num = np.sum( [1 for i in l1_learner.model if abs(i)>=0.001 ] ) - 1
        l2_non_0_num = np.sum( [1 for i in l2_learner.model if abs(i)>=0.001 ] ) - 1

        l1_non_0_num_arr.append(l1_non_0_num)
        l2_non_0_num_arr.append(l2_non_0_num)

    print("l1 iterations: ", l1_iter_arr)
    print("l2 iterations: ", l1_iter_arr)

    plt.plot(np.log10(lambda_arr), l1_non_0_num_arr, label="L1")
    plt.plot(np.log10(lambda_arr), l2_non_0_num_arr, label="L2")
    plt.title("lambda values vs number of non zero weights with L1, L2 regularizer")
    plt.legend(loc="lower right")
    plt.xlabel("lambda ")
    plt.ylabel("Number of non zero weights")
    plt.show()
def test_gradient_descent_blobs():
    """
    Tests the ability of the gradient descent algorithm to classify a linearly
    separable dataset.
    """
    features, _, targets, _ = load_data('blobs')

    hinge = make_predictions(features, targets, 'hinge', None)
    assert np.all(hinge == targets)

    l1_hinge = make_predictions(features, targets, 'hinge', 'l1')
    assert np.all(l1_hinge == targets)

    l2_hinge = make_predictions(features, targets, 'hinge', 'l2')
    assert np.all(l2_hinge == targets)

    squared = make_predictions(features, targets, 'squared', None)
    assert np.all(squared == targets)

    l1_squared = make_predictions(features, targets, 'squared', 'l1')
    assert np.all(l1_squared == targets)

    l2_squared = make_predictions(features, targets, 'squared', 'l2')
    assert np.all(l2_squared == targets)
Esempio n. 10
0
        for i in range(len(self.classes)):
            gradient_descent = self.model[i]
            confidence = gradient_descent.confidence(features)
            confidence_matrix[:, i] = confidence

        predictions = np.zeros((features.shape[0]))
        for r in range(len(confidence_matrix)):
            max_index = np.argmax(confidence_matrix[r], axis=0)
            predictions[r] = self.classes[max_index]

        return predictions


print('Question 3a')
max_iter = 1000
batch_size = 1
fraction = 0.75
learning_rate = 1e-4
reg_param = 0.05
loss = 'squared'
regularization = 'l1'

train_features, test_features, train_targets, test_targets = \
    load_data('mnist-multiclass', fraction=fraction)
learner = MultiClassGradientDescentQ3(loss=loss, regularization=regularization)
learner.fit(train_features, train_targets)
predictions = learner.predict(test_features)

confusion_matrix = metrics.confusion_matrix(test_targets, predictions)
print("confusion_matrix = ", confusion_matrix)
        """
        confidence = features.dot(self.model)
        return confidence


print('Question 1a')
max_iter = 1000
batch_size = None
fraction = 1
learning_rate = 1e-4
reg_param = 0.05
loss = 'hinge'
regularization = None

train_features, test_features, train_targets, test_targets = \
    load_data('mnist-binary', fraction=fraction)
learner = GradientDescentQ1(loss=loss,
                            regularization=regularization,
                            learning_rate=learning_rate,
                            reg_param=reg_param,
                            question='1a')
learner.fit(train_features, train_targets)
predictions = learner.predict(test_features)

print('Question 1b')
max_iter = 1000000
batch_size = 1
fraction = 1
learning_rate = 1e-4
reg_param = 0.05
loss = 'hinge'
Esempio n. 12
0
from your_code import GradientDescent, load_data

print('Starting example experiment')

train_features, test_features, train_targets, test_targets = \
    load_data('blobs', fraction=0.6)
learner = GradientDescent('squared')
learner.fit(train_features, train_targets)
predictions = learner.predict(test_features)

print('Finished example experiment')
Esempio n. 13
0
import numpy as np
import matplotlib.pyplot as plt
from your_code import load_data, ZeroOneLoss

print('Question 2a')
train_features, test_features, train_targets, test_targets = load_data(
    'synthetic', fraction=1)
ones_mat = np.ones((train_features.shape[0], 1))
train_features = np.append(train_features, ones_mat, axis=1)
loss = ZeroOneLoss()

bias = np.array([0.5, -0.5, -1.5, -2.5, -3.5, -4.5, -5.5])
w = np.array([1, 0])
loss_landscape = np.zeros(len(bias))
i = 0
for b in bias:
    w[1] = b
    loss_landscape[i] = loss.forward(train_features, w, train_targets)
    i += 1

plt.figure()
plt.plot(bias, loss_landscape, color='orange', label='Loss Landscape')
plt.title('Loss Landscape')
plt.xlabel('Bias')
plt.ylabel('Loss')
plt.legend(loc="best")
plt.savefig("Q2a.png")

print('Question 2a')

train_features, test_features, train_targets, test_targets = load_data(