提交 48fe99af authored 作者: Frederic Bastien's avatar Frederic Bastien

Small changes following code review

上级 70c22839
...@@ -3030,9 +3030,6 @@ def local_log_softmax_dnn(node): ...@@ -3030,9 +3030,6 @@ def local_log_softmax_dnn(node):
node.inputs[0].owner and node.inputs[0].owner and
isinstance(node.inputs[0].owner.op, GpuDnnSoftmax) and isinstance(node.inputs[0].owner.op, GpuDnnSoftmax) and
len(node.inputs[0].clients) == 1): len(node.inputs[0].clients) == 1):
if version(raises=False) < 3000:
# No log-softmax before cudnn v3
return
softmax_node = node.inputs[0].owner softmax_node = node.inputs[0].owner
new_softmax = GpuDnnSoftmax('log', softmax_node.op.mode) new_softmax = GpuDnnSoftmax('log', softmax_node.op.mode)
return [new_softmax(softmax_node.inputs[0])] return [new_softmax(softmax_node.inputs[0])]
......
...@@ -22,6 +22,7 @@ if theano.config.mode == 'FAST_COMPILE': ...@@ -22,6 +22,7 @@ if theano.config.mode == 'FAST_COMPILE':
else: else:
mode_with_gpu = theano.compile.mode.get_default_mode().including('gpuarray').excluding('gpu') mode_with_gpu = theano.compile.mode.get_default_mode().including('gpuarray').excluding('gpu')
mode_without_gpu = theano.compile.mode.get_default_mode().excluding('gpuarray') mode_without_gpu = theano.compile.mode.get_default_mode().excluding('gpuarray')
mode_without_gpu.check_py_code = False
# If using float16, cast reference input to float32 # If using float16, cast reference input to float32
......
...@@ -27,9 +27,8 @@ from .rnn_support import Model, GRU, LSTM, WrapperLayer ...@@ -27,9 +27,8 @@ from .rnn_support import Model, GRU, LSTM, WrapperLayer
from theano.configdefaults import SUPPORTED_DNN_CONV_ALGO_FWD from theano.configdefaults import SUPPORTED_DNN_CONV_ALGO_FWD
mode_with_gpu = mode_with_gpu.including() mode_with_gpu = mode_with_gpu.including()
mode_without_gpu = mode_without_gpu.including() # Globally disabled for mode_without_gpu
mode_with_gpu.check_py_code = False mode_with_gpu.check_py_code = False
mode_without_gpu.check_py_code = False
# If using float16, set CUDNN precision to float32 # If using float16, set CUDNN precision to float32
......
...@@ -3708,10 +3708,10 @@ def local_gpu_batch_norm_inference_inplace(node): ...@@ -3708,10 +3708,10 @@ def local_gpu_batch_norm_inference_inplace(node):
def values_eq_approx_high_tol(a, b): def values_eq_approx_high_tol(a, b):
""" """
This fct is needed to don't have DebugMode raise useless This fct is needed to don't have DebugMode raise useless
error due to ronding error. errors due to rounding error.
This happen as We reduce on the two last dimensions, so this This happen as we reduce on the two last dimensions, so this
can raise the absolute error if the number of element we can raise the absolute error if the number of elements we
reduce on is significant. reduce on is significant.
""" """
......
...@@ -203,7 +203,8 @@ class TensorType(Type): ...@@ -203,7 +203,8 @@ class TensorType(Type):
""" """
Convert a symbolic Variable into a TensorType, if compatible. Convert a symbolic Variable into a TensorType, if compatible.
For the moment, only a TensorType or CudaNdarrayType will be For the moment, only a TensorType, GpuArrayType and
CudaNdarrayType will be
converted, provided they have the same number of dimensions and converted, provided they have the same number of dimensions and
dtype and have "compatible" broadcastable pattern. dtype and have "compatible" broadcastable pattern.
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论