提交 73146b4a authored 作者: Brandon T. Willard's avatar Brandon T. Willard 提交者: Brandon T. Willard

Remove unnecessary print statements and comments

上级 fdaa40ba
...@@ -579,11 +579,6 @@ def get_scalar_constant_value( ...@@ -579,11 +579,6 @@ def get_scalar_constant_value(
raise NotScalarConstantError(v) raise NotScalarConstantError(v)
#########################
# Casting Operations
#########################
class TensorFromScalar(Op): class TensorFromScalar(Op):
__props__ = () __props__ = ()
......
...@@ -1810,7 +1810,6 @@ def test_join_inplace(): ...@@ -1810,7 +1810,6 @@ def test_join_inplace():
f = aesara.function([In(x, borrow=True), s], Out(c, borrow=True)) f = aesara.function([In(x, borrow=True), s], Out(c, borrow=True))
data = np.array([3, 4, 5], dtype=config.floatX) data = np.array([3, 4, 5], dtype=config.floatX)
print(f(data, 0))
if config.mode not in ["DebugMode", "DEBUG_MODE"]: if config.mode not in ["DebugMode", "DEBUG_MODE"]:
assert f(data, 0) is data assert f(data, 0) is data
...@@ -3441,7 +3440,7 @@ class TestAllocDiag: ...@@ -3441,7 +3440,7 @@ class TestAllocDiag:
# Test infer_shape # Test infer_shape
f_shape = aesara.function([x], adiag_op(x).shape, mode="FAST_RUN") f_shape = aesara.function([x], adiag_op(x).shape, mode="FAST_RUN")
aesara.printing.debugprint(f_shape.maker.fgraph.outputs[0]) # aesara.printing.debugprint(f_shape.maker.fgraph.outputs[0])
output_shape = f_shape(test_val) output_shape = f_shape(test_val)
assert not any( assert not any(
isinstance(node.op, self.alloc_diag) isinstance(node.op, self.alloc_diag)
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论