def test_variable_weight_one_input(self): #function x**2*a+x*b+c xv = np.array([1.3]) x = Input(['x'], 'x') av = np.array([2.1]) bv = np.array([3.2]) cv = np.array([5.1]) a = VWeight(1, weights=av) b = VWeight(1, weights=bv) c = VWeight(1, weights=cv) net = ComputationalGraphLayer(a * x**2 + b * x + c) out = net.forward(xv) self.assertEqual(out.shape, (1, )) assert_almost_equal(out, xv**2 * av + xv * bv + cv) dJdy = net.backward(np.array([1.0])) self.assertEqual(dJdy.shape, (1, )) assert_almost_equal(dJdy, 2 * xv * av + bv) net = ComputationalGraphLayer(x * x * a + x * b + c) out = net.forward(xv) self.assertEqual(out.shape, (1, )) assert_almost_equal(out, xv**2 * av + xv * bv + cv) dJdy = net.backward(np.array([1.0])) self.assertEqual(dJdy.shape, (1, )) assert_almost_equal(dJdy, 2 * xv * av + bv)
def test_variable_mw_and_w_multi_input(self): #test one layer W*x+b+W*y+b list_var = ['x', 'y'] x = Input(list_var, 'x') y = Input(list_var, 'y') xv = np.array([1.5, 1.1, 7.5]) yv = np.array([3.5, 3.1, 2.5, 4.3]) xyv = [xv, yv] Wv1 = np.array([[2.1, 3.1, 2.2], [2.2, 3.2, 4.2]]) Wv2 = np.array([[6.1, 5.1, 2.2, 4.3], [1.2, 1.2, 5.2, 5.1]]) W1 = MWeight(3, 2, weights=Wv1) W2 = MWeight(4, 2, weights=Wv2) bv = np.array([1.3, 5.1]) b = VWeight(2, weights=bv) net = ComputationalGraphLayer(W1.dot(x) + b + W2.dot(y) - b) out = net.forward(xyv) self.assertEqual(out.shape, (2, )) assert_almost_equal(out, Wv1.dot(xv) + bv + Wv2.dot(yv) - bv) dJdy = net.backward(np.array([1.0, 1.0])) self.assertEqual(len(dJdy), 2) gradvett = [np.sum(Wv1, 0), np.sum(Wv2, 0)] for ind, element in enumerate(dJdy): self.assertEqual(element.shape, xyv[ind].shape) assert_almost_equal(dJdy[ind], gradvett[ind])
def __init__(self, input_size, output_size, memory_size, window_size, Wxh='gaussian', Whh='gaussian', Why='gaussian', bh='zeros', by='zeros'): self.memory_size = memory_size self.window_size = window_size self.window_step = 0 x = Input(['x', 'h'], 'x') h = Input(['x', 'h'], 'h') s = Input(['s'], 's') self.statenet = [] self.outputnet = [] self.state = [] self.dJdh = [] self.Wxh = SharedWeights(Wxh, input_size, memory_size) self.Whh = SharedWeights(Whh, memory_size, memory_size) self.Why = SharedWeights(Why, memory_size, output_size) self.bh = SharedWeights(bh, 1, memory_size) self.by = SharedWeights(by, 1, output_size) for ind in range(window_size): cWxh = MWeight(input_size, memory_size, weights=self.Wxh) cWhh = MWeight(memory_size, memory_size, weights=self.Whh) cbh = VWeight(memory_size, weights=self.bh) cWhy = MWeight(memory_size, output_size, weights=self.Why) cby = VWeight(output_size, weights=self.by) self.statenet.append( ComputationalGraphLayer(Tanh(cWxh * x + cWhh * h + cbh))) self.outputnet.append( ComputationalGraphLayer( # Softmax(cWhy*s+cby) cWhy * s + cby)) self.state.append(np.zeros(memory_size)) self.dJdh.append(np.zeros(memory_size))
def test_complex(self): xv = np.array([0.5, 0.1]) hv = np.array([0.2, 0.4, 0.5]) cv = np.array([1.3, 2.4, 0.2]) vars2 = ['xh', 'c'] xsize = 2 hcsize = 3 xh = Input(vars2, 'xh') c = Input(vars2, 'c') Wf = MWeight(xsize + hcsize, hcsize) bf = VWeight(hcsize) net = Sequential(VariableDictLayer(vars2), ComputationalGraphLayer(Sigmoid(Wf.dot(xh) + bf) * c)) xhc = {'xh': np.hstack([xv, hv]), 'c': cv} out = net.forward(xhc) self.assertEqual(out.shape, (3, )) assert_almost_equal( out, sigmoid(Wf.net.W.get().dot(np.hstack([xv, hv])) + bf.net.W.get()) * cv) vars3 = ['x', 'h', 'c'] x = Input(vars3, 'x') #xsize h = Input(vars3, 'h') #hcsize c = Input(vars3, 'c') #hcsize Wf = MWeight(xsize + hcsize, hcsize) bf = VWeight(hcsize) net = Sequential( VariableDictLayer(vars3), ComputationalGraphLayer(Sigmoid(Wf.dot(Concat([x, h])) + bf) * c)) print net xhc = {'x': xv, 'h': hv, 'c': cv} out = net.forward(xhc) self.assertEqual(out.shape, (3, )) assert_almost_equal( out, sigmoid(Wf.net.W.get().dot(np.hstack([xv, hv])) + bf.net.W.get()) * cv)
def __init__(self, input_size, output_size, memory_size, Wxh='gaussian', Whh='gaussian', Why='gaussian', bh='zeros', by='zeros'): self.memory_size = memory_size x = Input(['x', 'h'], 'x') h = Input(['x', 'h'], 'h') s = Input(['s'], 's') self.Wxh = MWeight(input_size, memory_size, weights=Wxh) self.Whh = MWeight(memory_size, memory_size, weights=Whh) self.bh = VWeight(memory_size, weights=bh) self.Why = MWeight(memory_size, output_size, weights=Why) self.by = VWeight(output_size, weights=by) self.statenet = ComputationalGraphLayer( Tanh(self.Wxh.dot(x) + self.Whh.dot(h) + self.bh)) self.outputnet = ComputationalGraphLayer(self.Why.dot(s) + self.by) self.state = np.zeros(memory_size) self.dJdstate = np.zeros(memory_size)
def test_variable_mw_and_w_one_input(self): #test one layer W*x+b xv = np.array([1.5, 1.1, 7.5]) x = Input(['x'], 'x') Wv = np.array([[2.1, 3.1, 2.2], [2.2, 3.2, 4.2]]) W = MWeight(3, 2, weights=Wv) bv = np.array([1.3, 5.1]) b = VWeight(2, weights=bv) net = ComputationalGraphLayer(W.dot(x) + b) out = net.forward(xv) self.assertEqual(out.shape, (2, )) assert_almost_equal(out, Wv.dot(xv) + bv) dJdy = net.backward(np.array([1.0, 1.0])) self.assertEqual(dJdy.shape, (3, )) assert_almost_equal(dJdy, np.sum(Wv, 0))
def test_neuron_one_input(self): xv = np.array([0.5, 0.1, 0.5]) x = Input(['x'], 'x') Wv = np.array([[0.1, 0.1, 0.2], [0.5, 0.2, 0.2]]) W = MWeight(3, 2, weights=Wv) bv = np.array([0.3, 0.1]) b = VWeight(2, weights=bv) net = ComputationalGraphLayer(Sigmoid(W.dot(x) + b)) out = net.forward(xv) self.assertEqual(out.shape, (2, )) check_out = 1.0 / (1.0 + np.exp(-Wv.dot(xv) - bv)) assert_almost_equal(out, check_out) dJdy = net.backward(np.array([1.0, 1.0])) self.assertEqual(dJdy.shape, (3, )) assert_almost_equal(dJdy, np.sum(net.numeric_gradient(xv), 0)) assert_almost_equal(dJdy, (check_out * (1 - check_out)).dot(Wv)) net2 = Sequential( LinearLayer(3, 2, weights=np.hstack([Wv, bv.reshape(2, 1)])), SigmoidLayer) out2 = net2.forward(xv) assert_almost_equal(out, out2) dJdy2 = net.backward(np.array([1.0, 1.0])) assert_almost_equal(dJdy, dJdy2)
# Sequential( # ParallelGroup(GenericLayer,GenericLayer), # MulGroup( # GenericLayer, # VWeightLayer(1) # ) # ), # VWeightLayer(1) # ) # ) epochs = 100 #equal to varx = Input('x', 'x') a = VWeight(1, L2=0.001, weights=np.array([10.0])) b = VWeight(1, L2=0.001, weights=np.array([10.0])) c = VWeight(1, L2=0.001, weights=np.array([10.0])) # a = VWeight(1,weights=np.array([10.0])) # b = VWeight(1,weights=np.array([10.0])) # c = VWeight(1,weights=np.array([10.0])) # x = Input(lv,'x') n = ComputationalGraphLayer(a * varx**2 + b * varx + c) # train = [] for i, x in enumerate(np.linspace(-0.2, 0.2, 50)): train.append((np.array([x]), np.array([5.2 * x + 7.1]))) test = [] for i, x in enumerate(np.linspace(-10, 10, 50)):