提交 b1f7fb76 authored 作者: Olivier Delalleau's avatar Olivier Delalleau

Fixed typos and line too long in warning message

上级 e45087b7
...@@ -1052,8 +1052,13 @@ def local_argmax_pushdown(node): ...@@ -1052,8 +1052,13 @@ def local_argmax_pushdown(node):
(softmax, softplus, tensor.exp, tensor.log, tensor.tanh, sigmoid, (softmax, softplus, tensor.exp, tensor.log, tensor.tanh, sigmoid,
softmax_with_bias): softmax_with_bias):
if theano.config.warn.argmax_pushdown_bug: if theano.config.warn.argmax_pushdown_bug:
logging.getLogger('theano.tensor.nnet.nnet').warn("WARNING: their was a bug in Theano fixed the 27 may 2010 in this case. I.E. when we take the max of a softplus, softmax, exp, log, tanh, sigmoid, softmax_with_bias op, we where doing the max of the parent of the input. To remove this warning set the Theano flags 'warn.argmax_pushdown_bug' to False") logging.getLogger('theano.tensor.nnet.nnet').warn("WARNING: there "
"was a bug in Theano fixed on May 27th, 2010 in this case."
" I.E. when we take the max of a softplus, softmax, exp, "
"log, tanh, sigmoid, softmax_with_bias op, we were doing "
"the max of the parent of the input. To remove this "
"warning set the Theano flags 'warn.argmax_pushdown_bug' "
"to False")
if node.op == tensor._max_and_argmax and node.inputs[0].owner and len(node.outputs[0].clients)==0: if node.op == tensor._max_and_argmax and node.inputs[0].owner and len(node.outputs[0].clients)==0:
x_max, x_argmax = node.outputs x_max, x_argmax = node.outputs
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论