def test_clip_grad(self): #This is testing for the issue #633 x, y = floats('xy') a = theano.tensor.clip(x, y, x) g = theano.gradient.grad(a, x) fn = gof.DualLinker().accept(FunctionGraph([x, y], [g])).make_function() # Test the other way around as well a2 = theano.tensor.clip(x, x, y) g2 = theano.gradient.grad(a2, x) fn2 = gof.DualLinker().accept(FunctionGraph([x, y], [g2])).make_function() # Test for the equal case too . a3 = theano.tensor.clip(x, x, x) g3 = theano.gradient.grad(a3, x) fn3 = gof.DualLinker().accept(FunctionGraph([x], [g3])).make_function() rng = np.random.RandomState(utt.fetch_seed()) ntests = 50 for i in xrange(ntests): xval = rng.rand(1) #To ensure that the min < x . yval_mn = rng.rand(1) - 1.0 #To ensure that the max > x. yval_mx = rng.rand(1) + 1.0 aval = fn(xval, yval_mn) aval2 = fn2(xval, yval_mx) aval3 = fn3(xval) self.assertTrue(aval == 1.) self.assertTrue(aval2 == 1.) self.assertTrue(aval3 == 1.)
def test_straightforward(self): x, y, z = floats("xyz") e = mul(add(x, y), div_proxy(x, y)) C = Composite([x, y], [e]) c = C.make_node(x, y) # print c.c_code(['x', 'y'], ['z'], dict(id = 0)) g = FunctionGraph([x, y], [c.out]) fn = gof.DualLinker().accept(g).make_function() assert fn(1.0, 2.0) == 1.5
def test_with_constants(self): x, y, z = floats("xyz") e = mul(add(70.0, y), div_proxy(x, y)) C = Composite([x, y], [e]) c = C.make_node(x, y) assert "70.0" in c.op.c_code(c, "dummy", ["x", "y"], ["z"], dict(id=0)) # print c.c_code(['x', 'y'], ['z'], dict(id = 0)) g = FunctionGraph([x, y], [c.out]) fn = gof.DualLinker().accept(g).make_function() assert fn(1.0, 2.0) == 36.0
def test_composite_neg_bool(self): # Check that taking the negation of a Boolean intermediate value # works correctly with Python code. It used to be an issue because # `-numpy.bool_(True)` is False and `-numpy.bool_(False)` is True. x = floats('x') y = - (x > 0) z = Composite([x], [y]).make_node(x).outputs[0] f = theano.function([x], z, mode=theano.Mode(linker='py')) for inp, out in zip([-1, 0, 1], [0, 0, -1]): self.assertTrue(f(inp) == out)
def test_flatten(self): # Test that we flatten multiple Composite. x, y, z = floats("xyz") C = Composite([x, y], [x + y]) CC = Composite([x, y], [C(x * y, y)]) assert not isinstance(CC.outputs[0].owner.op, Composite) # Test with multiple outputs CC = Composite([x, y, z], [C(x * y, y), C(x * z, y)]) # We don't flatten that case. assert isinstance(CC.outputs[0].owner.op, Composite)
def test_many_outputs(self): x, y, z = floats("xyz") e0 = x + y + z e1 = x + y * z e2 = x / y C = Composite([x, y, z], [e0, e1, e2]) c = C.make_node(x, y, z) # print c.c_code(['x', 'y', 'z'], ['out0', 'out1', 'out2'], dict(id = 0)) g = FunctionGraph([x, y, z], c.outputs) fn = gof.DualLinker().accept(g).make_function() assert fn(1.0, 2.0, 3.0) == [6.0, 7.0, 0.5]
def test_composite_printing(self): x, y, z = floats('xyz') e0 = x + y + z e1 = x + y * z e2 = x / y e3 = x // 5 e4 = -x e5 = x - y e6 = x**y + (-z) e7 = x % 3 C = Composite([x, y, z], [e0, e1, e2, e3, e4, e5, e6, e7]) c = C.make_node(x, y, z) g = FunctionGraph([x, y, z], c.outputs) gof.DualLinker().accept(g).make_function() assert str(g) == ('[*1 -> Composite{((i0 + i1) + i2),' ' (i0 + (i1 * i2)), (i0 / i1), ' '(i0 // Constant{5}), ' '(-i0), (i0 - i1), ((i0 ** i1) + (-i2)),' ' (i0 % Constant{3})}(x, y, z), ' '*1::1, *1::2, *1::3, *1::4, *1::5, *1::6, *1::7]')
def test_composite_printing(self): x, y, z = floats('xyz') e0 = x + y + z e1 = x + y * z e2 = x / y e3 = x // 5 e4 = -x e5 = x - y e6 = x ** y + (-z) e7 = x % 3 C = Composite([x, y, z], [e0, e1, e2, e3, e4, e5, e6, e7]) c = C.make_node(x, y, z) g = FunctionGraph([x, y, z], c.outputs) fn = gof.DualLinker().accept(g).make_function() assert str(g) == ('[*1 -> Composite{((i0 + i1) + i2),' ' (i0 + (i1 * i2)), (i0 / i1), ' '(i0 // Constant{5}), ' '(-i0), (i0 - i1), ((i0 ** i1) + (-i2)),' ' (i0 % Constant{3})}(x, y, z), ' '*1::1, *1::2, *1::3, *1::4, *1::5, *1::6, *1::7]')
def test_clip_grad(self): # This is testing for the issue #633 x, y = floats('xy') a = theano.tensor.clip(x, y, x) g = theano.gradient.grad(a, x) fn = gof.DualLinker().accept(FunctionGraph([x, y], [g])).make_function() # Test the other way around as well a2 = theano.tensor.clip(x, x, y) g2 = theano.gradient.grad(a2, x) fn2 = gof.DualLinker().accept(FunctionGraph([x, y], [g2])).make_function() # Test for the equal case too . a3 = theano.tensor.clip(x, x, x) g3 = theano.gradient.grad(a3, x) fn3 = gof.DualLinker().accept(FunctionGraph([x], [g3])).make_function() rng = np.random.RandomState(utt.fetch_seed()) ntests = 50 for i in xrange(ntests): xval = rng.rand(1) # To ensure that the min < x . yval_mn = rng.rand(1) - 1.0 # To ensure that the max > x. yval_mx = rng.rand(1) + 1.0 aval = fn(xval, yval_mn) aval2 = fn2(xval, yval_mx) aval3 = fn3(xval) self.assertTrue(aval == 1.) self.assertTrue(aval2 == 1.) self.assertTrue(aval3 == 1.)
from theano.scalar.basic_sympy import SymPyCCode from theano.scalar.basic import floats import theano try: import sympy xs = sympy.Symbol('x') ys = sympy.Symbol('y') except ImportError: from nose.plugins.skip import SkipTest raise SkipTest('optional package sympy disabled') xt, yt = floats('xy') def test_SymPyCCode(): op = SymPyCCode([xs, ys], xs + ys) e = op(xt, yt) g = theano.gof.FunctionGraph([xt, yt], [e]) fn = theano.gof.CLinker().accept(g).make_function() assert fn(1.0, 2.0) == 3.0 def test_grad(): op = SymPyCCode([xs], xs**2) zt = op(xt) ztprime = theano.grad(zt, xt) assert ztprime.owner.op.expr == 2 * xs def test_multivar_grad():
def inputs(): return floats("xyz")
def test_mul_add_div_proxy(): x, y, z = floats("xyz") e = mul(add(x, y), div_proxy(x, y)) g = FunctionGraph([x, y], [e]) fn = gof.DualLinker().accept(g).make_function() assert fn(1.0, 2.0) == 1.5
def test_ge(self): x, y, z = floats("xyz") fn = DualLinker().accept(FunctionGraph([x, y], [x >= y])).make_function() for a, b in ((3.0, 9), (3, 0.9), (3, 3)): assert fn(a, b) == (a >= b)
def inputs(): return floats('xyz')
from __future__ import absolute_import, print_function, division import theano from theano.scalar.basic_sympy import SymPyCCode from theano.scalar.basic import floats from nose.plugins.skip import SkipTest try: import sympy xs = sympy.Symbol('x') ys = sympy.Symbol('y') except ImportError: raise SkipTest('optional package sympy disabled') xt, yt = floats('xy') def test_SymPyCCode(): if not theano.config.cxx: raise SkipTest("Need cxx for this test") op = SymPyCCode([xs, ys], xs + ys) e = op(xt, yt) g = theano.gof.FunctionGraph([xt, yt], [e]) fn = theano.gof.CLinker().accept(g).make_function() assert fn(1.0, 2.0) == 3.0 def test_grad(): op = SymPyCCode([xs], xs**2)
def test_neq(self): x, y, z = floats("xyz") fn = gof.DualLinker().accept(FunctionGraph([x, y], [neq(x, y)])).make_function() for a, b in ((3.0, 9), (3, 0.9), (3, 3)): assert fn(a, b) == (a != b)
import pytest import theano sympy = pytest.importorskip("sympy") from theano.scalar.basic import floats from theano.scalar.basic_sympy import SymPyCCode xs = sympy.Symbol("x") ys = sympy.Symbol("y") xt, yt = floats("xy") @pytest.mark.skipif(not theano.config.cxx, reason="Need cxx for this test") def test_SymPyCCode(): op = SymPyCCode([xs, ys], xs + ys) e = op(xt, yt) g = theano.gof.FunctionGraph([xt, yt], [e]) fn = theano.link.c.basic.CLinker().accept(g).make_function() assert fn(1.0, 2.0) == 3.0 def test_grad(): op = SymPyCCode([xs], xs**2) zt = op(xt) ztprime = theano.grad(zt, xt) assert ztprime.owner.op.expr == 2 * xs