提交 6a01c31e authored 作者: Iban Harlouchet's avatar Iban Harlouchet

__props__ for theano/tests/test_gradient.py

上级 c89e1bc2
...@@ -36,6 +36,7 @@ class testgrad_sources_inputs(unittest.TestCase): ...@@ -36,6 +36,7 @@ class testgrad_sources_inputs(unittest.TestCase):
def test_retNone1(self): def test_retNone1(self):
"""Test that it is not ok to return None from op.grad()""" """Test that it is not ok to return None from op.grad()"""
class retNone(gof.op.Op): class retNone(gof.op.Op):
__props__ = ()
def make_node(self): def make_node(self):
inputs = [theano.tensor.vector()] inputs = [theano.tensor.vector()]
outputs = [theano.tensor.vector()] outputs = [theano.tensor.vector()]
...@@ -52,6 +53,7 @@ class testgrad_sources_inputs(unittest.TestCase): ...@@ -52,6 +53,7 @@ class testgrad_sources_inputs(unittest.TestCase):
"""Test that it is not ok to return the wrong number of gradient terms """Test that it is not ok to return the wrong number of gradient terms
""" """
class retOne(gof.op.Op): class retOne(gof.op.Op):
__props__ = ()
def make_node(self, *inputs): def make_node(self, *inputs):
outputs = [theano.tensor.vector()] outputs = [theano.tensor.vector()]
return gof.Apply(self, inputs, outputs) return gof.Apply(self, inputs, outputs)
...@@ -72,6 +74,7 @@ class testgrad_sources_inputs(unittest.TestCase): ...@@ -72,6 +74,7 @@ class testgrad_sources_inputs(unittest.TestCase):
gval = theano.tensor.matrix() gval = theano.tensor.matrix()
class O(gof.op.Op): class O(gof.op.Op):
__props__ = ()
def make_node(self): def make_node(self):
inputs = [theano.tensor.matrix()] inputs = [theano.tensor.matrix()]
outputs = [theano.tensor.matrix()] outputs = [theano.tensor.matrix()]
...@@ -88,6 +91,7 @@ class testgrad_sources_inputs(unittest.TestCase): ...@@ -88,6 +91,7 @@ class testgrad_sources_inputs(unittest.TestCase):
gval = theano.tensor.matrix() gval = theano.tensor.matrix()
class O(gof.op.Op): class O(gof.op.Op):
__props__ = ()
def make_node(self): def make_node(self):
inputs = [theano.tensor.matrix()] inputs = [theano.tensor.matrix()]
outputs = [theano.tensor.scalar(), theano.tensor.scalar()] outputs = [theano.tensor.scalar(), theano.tensor.scalar()]
...@@ -107,6 +111,7 @@ class testgrad_sources_inputs(unittest.TestCase): ...@@ -107,6 +111,7 @@ class testgrad_sources_inputs(unittest.TestCase):
gval1 = theano.tensor.scalar() gval1 = theano.tensor.scalar()
class O(gof.op.Op): class O(gof.op.Op):
__props__ = ()
def make_node(self): def make_node(self):
inputs = [theano.tensor.scalar(), theano.tensor.scalar()] inputs = [theano.tensor.scalar(), theano.tensor.scalar()]
outputs = [theano.tensor.matrix()] outputs = [theano.tensor.matrix()]
...@@ -127,6 +132,7 @@ class testgrad_sources_inputs(unittest.TestCase): ...@@ -127,6 +132,7 @@ class testgrad_sources_inputs(unittest.TestCase):
gval1 = theano.tensor.matrix() gval1 = theano.tensor.matrix()
class O(gof.op.Op): class O(gof.op.Op):
__props__ = ()
def make_node(self): def make_node(self):
inputs = [theano.tensor.matrix(), theano.tensor.matrix()] inputs = [theano.tensor.matrix(), theano.tensor.matrix()]
outputs = [theano.tensor.matrix(), theano.tensor.matrix()] outputs = [theano.tensor.matrix(), theano.tensor.matrix()]
...@@ -161,6 +167,7 @@ class test_grad(unittest.TestCase): ...@@ -161,6 +167,7 @@ class test_grad(unittest.TestCase):
# tests that unimplemented grads are caught in the grad method # tests that unimplemented grads are caught in the grad method
class DummyOp(gof.Op): class DummyOp(gof.Op):
__props__ = ()
def make_node(self, x): def make_node(self, x):
return gof.Apply(self, [x], [x.type()]) return gof.Apply(self, [x], [x.type()])
...@@ -350,6 +357,7 @@ class test_grad(unittest.TestCase): ...@@ -350,6 +357,7 @@ class test_grad(unittest.TestCase):
# Op1 has two outputs, f and g # Op1 has two outputs, f and g
# x is connected to f but not to g # x is connected to f but not to g
class Op1(theano.gof.Op): class Op1(theano.gof.Op):
__props__ = ()
def make_node(self, x): def make_node(self, x):
return theano.Apply(self, inputs=[x], return theano.Apply(self, inputs=[x],
outputs=[x.type(), theano.tensor.scalar()]) outputs=[x.type(), theano.tensor.scalar()])
...@@ -363,6 +371,7 @@ class test_grad(unittest.TestCase): ...@@ -363,6 +371,7 @@ class test_grad(unittest.TestCase):
# Op2 has two inputs, f and g # Op2 has two inputs, f and g
# Its gradient with respect to g is not defined # Its gradient with respect to g is not defined
class Op2(theano.gof.Op): class Op2(theano.gof.Op):
__props__ = ()
def make_node(self, f, g): def make_node(self, f, g):
return theano.Apply(self, inputs=[f, g], return theano.Apply(self, inputs=[f, g],
outputs=[theano.tensor.scalar()]) outputs=[theano.tensor.scalar()])
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论