提交 ee22927e authored 作者: notoraptor's avatar notoraptor

Add newline after each debug print.

上级 bf5097b4
...@@ -147,7 +147,7 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns, ...@@ -147,7 +147,7 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
PyErr_SetString(PyExc_RuntimeError, "No best-timed conv fwd algorithm found"); PyErr_SetString(PyExc_RuntimeError, "No best-timed conv fwd algorithm found");
return 1; return 1;
} else { } else {
fprintf(stderr, " (%d best-timed conv fwd algorithms) ", count); fprintf(stderr, "(%d best-timed conv fwd algorithms)\n", count);
} }
#endif #endif
...@@ -174,9 +174,9 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns, ...@@ -174,9 +174,9 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
return 1; return 1;
// NB: This is printed only when algorithm is chosen at runtime. // NB: This is printed only when algorithm is chosen at runtime.
if (reuse_algo) if (reuse_algo)
fprintf(stderr, "(reused %s) ", algorithm_name); fprintf(stderr, "(reused %s)\n", algorithm_name);
else else
fprintf(stderr, "(using %s) ", algorithm_name); fprintf(stderr, "(using %s)\n", algorithm_name);
#endif #endif
if (params->choose_once) { if (params->choose_once) {
...@@ -198,7 +198,7 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns, ...@@ -198,7 +198,7 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
#ifdef DEBUG #ifdef DEBUG
if (0 != theano_enum_to_string_cudnnConvolutionFwdAlgo_t(algo, algorithm_name)) if (0 != theano_enum_to_string_cudnnConvolutionFwdAlgo_t(algo, algorithm_name))
return 1; return 1;
fprintf(stderr, "(%s unsupported for 3D: fallback to CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM) ", algorithm_name); fprintf(stderr, "(%s unsupported for 3D: fallback to CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM)\n", algorithm_name);
#endif #endif
algo = CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM; algo = CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM;
} }
...@@ -211,7 +211,7 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns, ...@@ -211,7 +211,7 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
{ {
#ifdef DEBUG #ifdef DEBUG
fprintf(stderr, "(CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM " fprintf(stderr, "(CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM "
"will fail with batch size > 2^16, fallback to CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM) "); "will fail with batch size > 2^16, fallback to CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM)\n");
#endif #endif
algo = CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM; algo = CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM;
} }
...@@ -283,7 +283,7 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns, ...@@ -283,7 +283,7 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
if (0 != theano_enum_to_string_cudnnConvolutionFwdAlgo_t(algo, algorithm_name)) if (0 != theano_enum_to_string_cudnnConvolutionFwdAlgo_t(algo, algorithm_name))
return 1; return 1;
fprintf(stderr, "(%s error getting worksize: " fprintf(stderr, "(%s error getting worksize: "
"fallback to CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM) ", algorithm_name); "fallback to CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM)\n", algorithm_name);
#endif #endif
algo = CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM; algo = CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM;
......
...@@ -187,7 +187,7 @@ APPLY_SPECIFIC(conv_gi)(PyGpuArrayObject *kerns, PyGpuArrayObject *output, ...@@ -187,7 +187,7 @@ APPLY_SPECIFIC(conv_gi)(PyGpuArrayObject *kerns, PyGpuArrayObject *output,
PyErr_SetString(PyExc_RuntimeError, "No best-timed conv gradinput algorithm found"); PyErr_SetString(PyExc_RuntimeError, "No best-timed conv gradinput algorithm found");
return 1; return 1;
} else { } else {
fprintf(stderr, " (%d best-timed conv gradinput algorithms) ", count); fprintf(stderr, "(%d best-timed conv gradinput algorithms)\n", count);
} }
#endif #endif
...@@ -214,9 +214,9 @@ APPLY_SPECIFIC(conv_gi)(PyGpuArrayObject *kerns, PyGpuArrayObject *output, ...@@ -214,9 +214,9 @@ APPLY_SPECIFIC(conv_gi)(PyGpuArrayObject *kerns, PyGpuArrayObject *output,
return 1; return 1;
// NB: This is printed only when algorithm is chosen at runtime. // NB: This is printed only when algorithm is chosen at runtime.
if (reuse_algo) if (reuse_algo)
fprintf(stderr, "(reused %s) ", algorithm_name); fprintf(stderr, "(reused %s)\n", algorithm_name);
else else
fprintf(stderr, "(using %s) ", algorithm_name); fprintf(stderr, "(using %s)\n", algorithm_name);
#endif #endif
if (params->choose_once) { if (params->choose_once) {
......
...@@ -189,7 +189,7 @@ APPLY_SPECIFIC(conv_gw)(PyGpuArrayObject *input, PyGpuArrayObject *output, ...@@ -189,7 +189,7 @@ APPLY_SPECIFIC(conv_gw)(PyGpuArrayObject *input, PyGpuArrayObject *output,
PyErr_SetString(PyExc_RuntimeError, "No best-timed conv gradweight algorithm found"); PyErr_SetString(PyExc_RuntimeError, "No best-timed conv gradweight algorithm found");
return 1; return 1;
} else { } else {
fprintf(stderr, " (%d best-timed conv gradweight algorithms) ", count); fprintf(stderr, "(%d best-timed conv gradweight algorithms)\n", count);
} }
#endif #endif
...@@ -216,9 +216,9 @@ APPLY_SPECIFIC(conv_gw)(PyGpuArrayObject *input, PyGpuArrayObject *output, ...@@ -216,9 +216,9 @@ APPLY_SPECIFIC(conv_gw)(PyGpuArrayObject *input, PyGpuArrayObject *output,
return 1; return 1;
// NB: This is printed only when algorithm is chosen at runtime. // NB: This is printed only when algorithm is chosen at runtime.
if (reuse_algo) if (reuse_algo)
fprintf(stderr, "(reused %s) ", algorithm_name); fprintf(stderr, "(reused %s)\n", algorithm_name);
else else
fprintf(stderr, "(using %s) ", algorithm_name); fprintf(stderr, "(using %s)\n", algorithm_name);
#endif #endif
if (params->choose_once) { if (params->choose_once) {
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论