提交 6a3770d5 authored 作者: Pascal Lamblin's avatar Pascal Lamblin

Remove printing statements, update C code version

上级 67671bfa
...@@ -94,9 +94,6 @@ def c_set_tensor4d(var, desc, err, fail): ...@@ -94,9 +94,6 @@ def c_set_tensor4d(var, desc, err, fail):
str2 = CudaNdarray_HOST_STRIDES(%(var)s)[2]?CudaNdarray_HOST_STRIDES(%(var)s)[2]:CudaNdarray_HOST_DIMS(%(var)s)[3]; str2 = CudaNdarray_HOST_STRIDES(%(var)s)[2]?CudaNdarray_HOST_STRIDES(%(var)s)[2]:CudaNdarray_HOST_DIMS(%(var)s)[3];
str1 = CudaNdarray_HOST_STRIDES(%(var)s)[1]?CudaNdarray_HOST_STRIDES(%(var)s)[1]:CudaNdarray_HOST_DIMS(%(var)s)[2]*CudaNdarray_HOST_DIMS(%(var)s)[3]; str1 = CudaNdarray_HOST_STRIDES(%(var)s)[1]?CudaNdarray_HOST_STRIDES(%(var)s)[1]:CudaNdarray_HOST_DIMS(%(var)s)[2]*CudaNdarray_HOST_DIMS(%(var)s)[3];
str0 = CudaNdarray_HOST_STRIDES(%(var)s)[0]?CudaNdarray_HOST_STRIDES(%(var)s)[0]:CudaNdarray_HOST_DIMS(%(var)s)[2]*CudaNdarray_HOST_DIMS(%(var)s)[3]*CudaNdarray_HOST_DIMS(%(var)s)[1]; str0 = CudaNdarray_HOST_STRIDES(%(var)s)[0]?CudaNdarray_HOST_STRIDES(%(var)s)[0]:CudaNdarray_HOST_DIMS(%(var)s)[2]*CudaNdarray_HOST_DIMS(%(var)s)[3]*CudaNdarray_HOST_DIMS(%(var)s)[1];
printf("str0=%%d str1=%%d str2=%%d str3=%%d\\n",
str0, str1, str2, str3
);
%(err)s = cudnnSetTensor4dDescriptorEx( %(err)s = cudnnSetTensor4dDescriptorEx(
%(desc)s, CUDNN_DATA_FLOAT, %(desc)s, CUDNN_DATA_FLOAT,
CudaNdarray_HOST_DIMS(%(var)s)[0], CudaNdarray_HOST_DIMS(%(var)s)[0],
...@@ -917,8 +914,7 @@ if (err%(name)s != CUDNN_STATUS_SUCCESS) { ...@@ -917,8 +914,7 @@ if (err%(name)s != CUDNN_STATUS_SUCCESS) {
return [[1], [0]] return [[1], [0]]
def c_code_cache_version(self): def c_code_cache_version(self):
return return (6, version())
return (5, version())
class GpuDnnPoolGrad(DnnBase): class GpuDnnPoolGrad(DnnBase):
...@@ -1077,26 +1073,29 @@ _handle, ...@@ -1077,26 +1073,29 @@ _handle,
#endif #endif
if (err%(name)s != CUDNN_STATUS_SUCCESS) { if (err%(name)s != CUDNN_STATUS_SUCCESS) {
PyErr_Format(PyExc_RuntimeError, PyErr_Format(PyExc_RuntimeError,
"GpuDnnPoolGrad: error doing operation: %%s", "GpuDnnPoolGrad: error doing operation: %%s. "
cudnnGetErrorString(err%(name)s)); "input.shape=(%%d, %%d, %%d, %%d) "
printf("input.shape=(%%d, %%d, %%d, %%d) input_grad.shape=(%%d, %%d, %%d, %%d) output.shape=(%%d, %%d, %%d, %%d) output_grad.shape=(%%d, %%d, %%d, %%d)\\n", "input_grad.shape=(%%d, %%d, %%d, %%d) "
CudaNdarray_HOST_DIMS(%(input)s)[0], "output.shape=(%%d, %%d, %%d, %%d) "
CudaNdarray_HOST_DIMS(%(input)s)[1], "output_grad.shape=(%%d, %%d, %%d, %%d)",
CudaNdarray_HOST_DIMS(%(input)s)[2], cudnnGetErrorString(err%(name)s),
CudaNdarray_HOST_DIMS(%(input)s)[3], CudaNdarray_HOST_DIMS(%(input)s)[0],
CudaNdarray_HOST_DIMS(%(input_grad)s)[0], CudaNdarray_HOST_DIMS(%(input)s)[1],
CudaNdarray_HOST_DIMS(%(input_grad)s)[1], CudaNdarray_HOST_DIMS(%(input)s)[2],
CudaNdarray_HOST_DIMS(%(input_grad)s)[2], CudaNdarray_HOST_DIMS(%(input)s)[3],
CudaNdarray_HOST_DIMS(%(input_grad)s)[3], CudaNdarray_HOST_DIMS(%(input_grad)s)[0],
CudaNdarray_HOST_DIMS(%(output)s)[0], CudaNdarray_HOST_DIMS(%(input_grad)s)[1],
CudaNdarray_HOST_DIMS(%(output)s)[1], CudaNdarray_HOST_DIMS(%(input_grad)s)[2],
CudaNdarray_HOST_DIMS(%(output)s)[2], CudaNdarray_HOST_DIMS(%(input_grad)s)[3],
CudaNdarray_HOST_DIMS(%(output)s)[3], CudaNdarray_HOST_DIMS(%(output)s)[0],
CudaNdarray_HOST_DIMS(%(output_grad)s)[0], CudaNdarray_HOST_DIMS(%(output)s)[1],
CudaNdarray_HOST_DIMS(%(output_grad)s)[1], CudaNdarray_HOST_DIMS(%(output)s)[2],
CudaNdarray_HOST_DIMS(%(output_grad)s)[2], CudaNdarray_HOST_DIMS(%(output)s)[3],
CudaNdarray_HOST_DIMS(%(output_grad)s)[3] CudaNdarray_HOST_DIMS(%(output_grad)s)[0],
); CudaNdarray_HOST_DIMS(%(output_grad)s)[1],
CudaNdarray_HOST_DIMS(%(output_grad)s)[2],
CudaNdarray_HOST_DIMS(%(output_grad)s)[3]
);
%(fail)s %(fail)s
} }
""" % dict(output_grad=out_grad, desc=desc, """ % dict(output_grad=out_grad, desc=desc,
...@@ -1109,8 +1108,7 @@ CudaNdarray_HOST_DIMS(%(output_grad)s)[3] ...@@ -1109,8 +1108,7 @@ CudaNdarray_HOST_DIMS(%(output_grad)s)[3]
output_grad_desc="output_grad"+name) output_grad_desc="output_grad"+name)
def c_code_cache_version(self): def c_code_cache_version(self):
return return (5, version())
return (4, version())
def infer_shape(self, node, shape): def infer_shape(self, node, shape):
return [shape[0]] return [shape[0]]
......
...@@ -103,8 +103,6 @@ def test_pooling(): ...@@ -103,8 +103,6 @@ def test_pooling():
(1, 3, 99, 99), (1, 3, 99, 99),
(32, 1, 147, 197), (32, 1, 147, 197),
]: ]:
print func, pad, ws, stride, shp
data = numpy.random.normal(0, 1, shp).astype("float32") data = numpy.random.normal(0, 1, shp).astype("float32")
a = f1(data).__array__() a = f1(data).__array__()
...@@ -119,7 +117,6 @@ def test_pooling(): ...@@ -119,7 +117,6 @@ def test_pooling():
ws = 2 ws = 2
stride = 2 stride = 2
print func, pad, ws, stride, shp
# This test the CPU grad + opt + GPU implemtentation # This test the CPU grad + opt + GPU implemtentation
def fn(x): def fn(x):
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论