提交 50248b0f authored 作者: Olivier Delalleau's avatar Olivier Delalleau

PEP8

上级 f5d8bde1
...@@ -244,7 +244,8 @@ class GpuOp(theano.gof.Op): ...@@ -244,7 +244,8 @@ class GpuOp(theano.gof.Op):
return super(GpuOp, self).make_thunk(node, storage_map, return super(GpuOp, self).make_thunk(node, storage_map,
compute_map, no_recycling) compute_map, no_recycling)
theano.compile.debugmode.default_make_thunk.append(get_unbound_function(GpuOp.make_thunk)) theano.compile.debugmode.default_make_thunk.append(
get_unbound_function(GpuOp.make_thunk))
# We must do those import to be able to create the full doc when # We must do those import to be able to create the full doc when
# nvcc is not available # nvcc is not available
...@@ -271,7 +272,8 @@ if cuda_available: ...@@ -271,7 +272,8 @@ if cuda_available:
shared_constructor = float32_shared_constructor shared_constructor = float32_shared_constructor
import basic_ops import basic_ops
from basic_ops import (GpuFromHost, HostFromGpu, GpuElemwise, from basic_ops import (
GpuFromHost, HostFromGpu, GpuElemwise,
GpuDimShuffle, GpuCAReduce, GpuReshape, GpuContiguous, GpuDimShuffle, GpuCAReduce, GpuReshape, GpuContiguous,
GpuSubtensor, GpuIncSubtensor, GpuSubtensor, GpuIncSubtensor,
GpuAdvancedSubtensor1, GpuAdvancedIncSubtensor1, GpuAdvancedSubtensor1, GpuAdvancedIncSubtensor1,
...@@ -388,7 +390,7 @@ def use(device, ...@@ -388,7 +390,7 @@ def use(device,
cuda_enabled = True cuda_enabled = True
if config.print_active_device: if config.print_active_device:
print >> sys.stderr, "Using gpu device %d: %s" %( print >> sys.stderr, "Using gpu device %d: %s" % (
active_device_number(), active_device_name()) active_device_number(), active_device_name())
if device_properties(use.device_number)['regsPerBlock'] < 16384: if device_properties(use.device_number)['regsPerBlock'] < 16384:
# We will try to use too much register per bloc at many places # We will try to use too much register per bloc at many places
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论