Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
4ad36ddc
提交
4ad36ddc
authored
12月 17, 2015
作者:
Frédéric Bastien
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #3788 from nouiz/carriepl-v4
Rebased cudnn v4
上级
22db3930
32e113c1
显示空白字符变更
内嵌
并排
正在显示
13 个修改的文件
包含
410 行增加
和
299 行删除
+410
-299
configdefaults.py
theano/configdefaults.py
+68
-7
cudnn_helper.h
theano/sandbox/cuda/cudnn_helper.h
+9
-0
dnn.py
theano/sandbox/cuda/dnn.py
+107
-53
dnn_conv_base.c
theano/sandbox/cuda/dnn_conv_base.c
+0
-5
dnn_fwd.c
theano/sandbox/cuda/dnn_fwd.c
+25
-36
dnn_gi.c
theano/sandbox/cuda/dnn_gi.c
+28
-21
dnn_gw.c
theano/sandbox/cuda/dnn_gw.c
+8
-17
conv_desc.c
theano/sandbox/gpuarray/conv_desc.c
+2
-2
cudnn_helper.h
theano/sandbox/gpuarray/cudnn_helper.h
+5
-92
dnn.py
theano/sandbox/gpuarray/dnn.py
+90
-33
dnn_fwd.c
theano/sandbox/gpuarray/dnn_fwd.c
+27
-22
dnn_gi.c
theano/sandbox/gpuarray/dnn_gi.c
+28
-6
dnn_gw.c
theano/sandbox/gpuarray/dnn_gw.c
+13
-5
没有找到文件。
theano/configdefaults.py
浏览文件 @
4ad36ddc
...
@@ -219,30 +219,91 @@ AddConfigVar('gpuarray.sync',
...
@@ -219,30 +219,91 @@ AddConfigVar('gpuarray.sync',
BoolParam
(
False
),
BoolParam
(
False
),
in_c_key
=
True
)
in_c_key
=
True
)
def
safe_no_dnn_workmem
(
workmem
):
"""
Make sure the user is not attempting to use dnn.conv.workmem`.
"""
if
workmem
:
raise
RuntimeError
(
'The option `dnn.conv.workmem` has been removed and should '
'not be used anymore. Please use the option '
'`dnn.conv.algo_fwd` instead.'
)
return
True
AddConfigVar
(
'dnn.conv.workmem'
,
AddConfigVar
(
'dnn.conv.workmem'
,
"This flag is deprecated; use dnn.conv.algo_fwd."
,
"This flag is deprecated; use dnn.conv.algo_fwd."
,
EnumStr
(
''
),
ConfigParam
(
''
,
allow_override
=
False
,
filter
=
safe_no_dnn_workmem
),
in_c_key
=
False
)
in_c_key
=
False
)
def
safe_no_dnn_workmem_bwd
(
workmem
):
"""
Make sure the user is not attempting to use dnn.conv.workmem_bwd`.
"""
if
workmem
:
raise
RuntimeError
(
'The option `dnn.conv.workmem_bwd` has been removed and '
'should not be used anymore. Please use the options '
'`dnn.conv.algo_bwd_filter` and `dnn.conv.algo_bwd_data` instead.'
)
return
True
AddConfigVar
(
'dnn.conv.workmem_bwd'
,
AddConfigVar
(
'dnn.conv.workmem_bwd'
,
"This flag is deprecated; use dnn.conv.algo_bwd."
,
"This flag is deprecated; use dnn.conv.algo_bwd."
,
EnumStr
(
''
),
ConfigParam
(
''
,
allow_override
=
False
,
filter
=
safe_no_dnn_workmem_bwd
),
in_c_key
=
False
)
def
safe_no_dnn_algo_bwd
(
algo
):
"""
Make sure the user is not attempting to use dnn.conv.algo_bwd`.
"""
if
algo
:
raise
RuntimeError
(
'The option `dnn.conv.algo_bwd` has been removed and '
'should not be used anymore. Please use the options '
'`dnn.conv.algo_bwd_filter` and `dnn.conv.algo_bwd_data` instead.'
)
return
True
AddConfigVar
(
'dnn.conv.algo_bwd'
,
"This flag is deprecated; use dnn.conv.algo_bwd_data and "
"dnn.conv.algo_bwd_filter."
,
ConfigParam
(
''
,
allow_override
=
False
,
filter
=
safe_no_dnn_algo_bwd
),
in_c_key
=
False
)
in_c_key
=
False
)
AddConfigVar
(
'dnn.conv.algo_fwd'
,
AddConfigVar
(
'dnn.conv.algo_fwd'
,
"Default implementation to use for CuDNN forward convolution."
,
"Default implementation to use for CuDNN forward convolution."
,
EnumStr
(
'small'
,
'none'
,
'large'
,
'fft'
,
'guess_once'
,
EnumStr
(
'small'
,
'none'
,
'large'
,
'fft'
,
'fft_tiling'
,
'guess_on_shape_change'
,
'time_once'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
),
in_c_key
=
False
)
AddConfigVar
(
'dnn.conv.algo_bwd_data'
,
"Default implementation to use for CuDNN backward convolution to "
"get the gradients of the convolution with regard to the inputs."
,
EnumStr
(
'none'
,
'deterministic'
,
'fft'
,
'fft_tiling'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
),
'time_on_shape_change'
),
in_c_key
=
False
)
in_c_key
=
False
)
AddConfigVar
(
'dnn.conv.algo_bwd'
,
AddConfigVar
(
'dnn.conv.algo_bwd_filter'
,
"Default implementation to use for CuDNN backward convolution."
,
"Default implementation to use for CuDNN backward convolution to "
EnumStr
(
'none'
,
'deterministic'
,
'fft'
,
'guess_once'
,
"get the gradients of the convolution with regard to the "
"filters."
,
EnumStr
(
'none'
,
'deterministic'
,
'fft'
,
'small'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
),
'time_on_shape_change'
),
in_c_key
=
False
)
in_c_key
=
False
)
AddConfigVar
(
'dnn.conv.precision'
,
"Default data precision to use for the computation in CuDNN "
"convolutions (defaults to the same dtype as the inputs of the "
"convolutions)."
,
EnumStr
(
'as_input'
,
'float16'
,
'float32'
,
'float64'
),
in_c_key
=
False
)
def
default_dnn_path
(
suffix
):
def
default_dnn_path
(
suffix
):
def
f
(
suffix
=
suffix
):
def
f
(
suffix
=
suffix
):
...
...
theano/sandbox/cuda/cudnn_helper.h
浏览文件 @
4ad36ddc
...
@@ -3,6 +3,15 @@
...
@@ -3,6 +3,15 @@
#include <cudnn.h>
#include <cudnn.h>
// If needed, define element of the V4 interface in terms of elements of
// previous versions
#if defined(CUDNN_VERSION) && CUDNN_VERSION < 4000
#define CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING 5
#define CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING 3
#endif
#ifndef CUDNN_VERSION
#ifndef CUDNN_VERSION
#include <assert.h>
#include <assert.h>
...
...
theano/sandbox/cuda/dnn.py
浏览文件 @
4ad36ddc
...
@@ -92,27 +92,13 @@ if ((err = cudnnCreate(&_handle)) != CUDNN_STATUS_SUCCESS) {
...
@@ -92,27 +92,13 @@ if ((err = cudnnCreate(&_handle)) != CUDNN_STATUS_SUCCESS) {
" from one version, but we link with"
" from one version, but we link with"
" a different version
%
s"
%
str
(
v
))
" a different version
%
s"
%
str
(
v
))
raise
RuntimeError
(
dnn_available
.
msg
)
raise
RuntimeError
(
dnn_available
.
msg
)
if
v
==
-
1
:
if
v
==
-
1
or
v
[
0
]
<
3007
:
dnn_available
.
avail
=
False
dnn_available
.
msg
=
(
"CuDNN v1 detected. This version is no longer "
"supported by Theano. Update your CuDNN installation "
"to a more recent version"
)
raise
RuntimeError
(
dnn_available
.
msg
)
if
v
==
(
20
,
20
):
dnn_available
.
avail
=
False
dnn_available
.
msg
=
(
"You have installed a release candidate of CuDNN v2."
" This isn't supported anymore."
" Update to CuDNN v2 final version."
)
raise
RuntimeError
(
dnn_available
.
msg
)
if
3000
<=
v
[
0
]
<
3007
:
# 3007 is the final release of cudnn v3
# 3007 is the final release of cudnn v3
dnn_available
.
avail
=
False
dnn_available
.
avail
=
False
dnn_available
.
msg
=
(
dnn_available
.
msg
=
(
"You have
installed a release candidate of CuDNN v3.
"
"You have
an old release of CuDNN (or a release
"
"
This isn't supported anymore.
"
"
candidate) that isn't supported. Please update to
"
"
Update to CuDNN
v3 final version."
)
"
at least
v3 final version."
)
raise
RuntimeError
(
dnn_available
.
msg
)
raise
RuntimeError
(
dnn_available
.
msg
)
return
dnn_available
.
avail
return
dnn_available
.
avail
...
@@ -248,7 +234,7 @@ class GpuDnnConvDesc(GpuOp):
...
@@ -248,7 +234,7 @@ class GpuDnnConvDesc(GpuOp):
"""
"""
__props__
=
(
'border_mode'
,
'subsample'
,
'conv_mode'
)
__props__
=
(
'border_mode'
,
'subsample'
,
'conv_mode'
,
'precision'
)
def
c_headers
(
self
):
def
c_headers
(
self
):
return
[
'cudnn.h'
,
'cudnn_helper.h'
]
return
[
'cudnn.h'
,
'cudnn_helper.h'
]
...
@@ -265,7 +251,8 @@ class GpuDnnConvDesc(GpuOp):
...
@@ -265,7 +251,8 @@ class GpuDnnConvDesc(GpuOp):
def
do_constant_folding
(
self
,
node
):
def
do_constant_folding
(
self
,
node
):
return
False
return
False
def
__init__
(
self
,
border_mode
,
subsample
=
(
1
,
1
),
conv_mode
=
'conv'
):
def
__init__
(
self
,
border_mode
,
subsample
=
(
1
,
1
),
conv_mode
=
'conv'
,
precision
=
"float32"
):
if
isinstance
(
border_mode
,
int
):
if
isinstance
(
border_mode
,
int
):
border_mode
=
(
border_mode
,)
*
len
(
subsample
)
border_mode
=
(
border_mode
,)
*
len
(
subsample
)
if
isinstance
(
border_mode
,
tuple
):
if
isinstance
(
border_mode
,
tuple
):
...
@@ -283,6 +270,9 @@ class GpuDnnConvDesc(GpuOp):
...
@@ -283,6 +270,9 @@ class GpuDnnConvDesc(GpuOp):
assert
conv_mode
in
(
'conv'
,
'cross'
)
assert
conv_mode
in
(
'conv'
,
'cross'
)
self
.
conv_mode
=
conv_mode
self
.
conv_mode
=
conv_mode
assert
precision
in
[
'float16'
,
'float32'
,
'float64'
]
self
.
precision
=
precision
def
make_node
(
self
,
img_shape
,
kern_shape
):
def
make_node
(
self
,
img_shape
,
kern_shape
):
if
img_shape
.
type
.
ndim
!=
1
or
img_shape
.
type
.
dtype
!=
'int64'
:
if
img_shape
.
type
.
ndim
!=
1
or
img_shape
.
type
.
dtype
!=
'int64'
:
raise
TypeError
(
'img must be 1D shape tensor'
)
raise
TypeError
(
'img must be 1D shape tensor'
)
...
@@ -321,6 +311,14 @@ class GpuDnnConvDesc(GpuOp):
...
@@ -321,6 +311,14 @@ class GpuDnnConvDesc(GpuOp):
subsample_str
=
", "
.
join
([
str
(
s
)
for
s
in
self
.
subsample
])
subsample_str
=
", "
.
join
([
str
(
s
)
for
s
in
self
.
subsample
])
upscale_str
=
", "
.
join
([
"1"
]
*
nb_dim
)
upscale_str
=
", "
.
join
([
"1"
]
*
nb_dim
)
if
self
.
precision
==
'float16'
:
precision
=
'CUDNN_DATA_HALF'
elif
self
.
precision
==
'float32'
:
precision
=
'CUDNN_DATA_FLOAT'
else
:
assert
self
.
precision
==
'float64'
precision
=
'CUDNN_DATA_DOUBLE'
return
"""
return
"""
{
{
cudnnStatus_t err;
cudnnStatus_t err;
...
@@ -346,11 +344,11 @@ class GpuDnnConvDesc(GpuOp):
...
@@ -346,11 +344,11 @@ class GpuDnnConvDesc(GpuOp):
}
}
}
}
err = cudnnSetConvolutionNdDescriptor(
err = cudnnSetConvolutionNdDescriptor
_v3
(
%(desc)
s,
%(desc)
s,
%(nb_dim)
d,
%(nb_dim)
d,
pad, subsample, upscale,
pad, subsample, upscale,
%(conv_flag)
s
%(conv_flag)
s
,
%(precision)
s
);
);
#else
#else
PyErr_Format(PyExc_RuntimeError, "could not set op descriptor: CUDNN_VERSION must be >= 30");
PyErr_Format(PyExc_RuntimeError, "could not set op descriptor: CUDNN_VERSION must be >= 30");
...
@@ -364,10 +362,10 @@ class GpuDnnConvDesc(GpuOp):
...
@@ -364,10 +362,10 @@ class GpuDnnConvDesc(GpuOp):
"""
%
dict
(
name
=
name
,
img_shape
=
img_shape
,
kern_shape
=
kern_shape
,
desc
=
desc
,
"""
%
dict
(
name
=
name
,
img_shape
=
img_shape
,
kern_shape
=
kern_shape
,
desc
=
desc
,
bmode
=
bmode
,
conv_flag
=
conv_flag
,
fail
=
sub
[
'fail'
],
bmode
=
bmode
,
conv_flag
=
conv_flag
,
fail
=
sub
[
'fail'
],
pad_str
=
pad_str
,
subsample_str
=
subsample_str
,
pad_str
=
pad_str
,
subsample_str
=
subsample_str
,
upscale_str
=
upscale_str
,
nb_dim
=
nb_dim
)
upscale_str
=
upscale_str
,
nb_dim
=
nb_dim
,
precision
=
precision
)
def
c_code_cache_version
(
self
):
def
c_code_cache_version
(
self
):
return
(
2
,
version
())
return
(
3
,
version
())
# scalar constants
# scalar constants
_zero
=
constant
(
numpy
.
asarray
(
0.0
,
dtype
=
'float32'
))
_zero
=
constant
(
numpy
.
asarray
(
0.0
,
dtype
=
'float32'
))
...
@@ -401,9 +399,8 @@ class GpuDnnConv(DnnBase, COp):
...
@@ -401,9 +399,8 @@ class GpuDnnConv(DnnBase, COp):
workmem
workmem
*deprecated*, use parameter algo instead.
*deprecated*, use parameter algo instead.
algo
algo
['none', 'small', 'large', 'fft', 'guess_once',
['none', 'small', 'large', 'fft', 'fft_tiling', 'guess_once',
'guess_on_shape_change', 'time_once',
'guess_on_shape_change', 'time_once', 'time_on_shape_change']
'time_on_shape_change']
Default is the value of :attr:`config.dnn.conv.algo_fwd`.
Default is the value of :attr:`config.dnn.conv.algo_fwd`.
...
@@ -445,9 +442,15 @@ class GpuDnnConv(DnnBase, COp):
...
@@ -445,9 +442,15 @@ class GpuDnnConv(DnnBase, COp):
raise
RuntimeError
(
"CuDNN convolution timing requires CuDNN "
raise
RuntimeError
(
"CuDNN convolution timing requires CuDNN "
"v3"
)
"v3"
)
assert
self
.
algo
in
[
'none'
,
'small'
,
'large'
,
'fft'
,
'guess_once'
,
# The fft_tiling implementation is only available from CuDNN V4 onward
'guess_on_shape_change'
,
'time_once'
,
if
version
()
<
(
4000
,
4000
):
'time_on_shape_change'
]
if
self
.
algo
==
'fft_tiling'
:
raise
RuntimeError
(
"CuDNN tiled-FFT convolution requires "
"CuDNN v4 or more recent"
)
assert
self
.
algo
in
[
'none'
,
'small'
,
'large'
,
'fft'
,
'fft_tiling'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
]
def
__setstate__
(
self
,
d
):
def
__setstate__
(
self
,
d
):
self
.
__dict__
.
update
(
d
)
self
.
__dict__
.
update
(
d
)
...
@@ -477,8 +480,15 @@ class GpuDnnConv(DnnBase, COp):
...
@@ -477,8 +480,15 @@ class GpuDnnConv(DnnBase, COp):
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM'
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM'
elif
self
.
algo
==
'large'
:
elif
self
.
algo
==
'large'
:
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_GEMM'
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_GEMM'
elif
self
.
algo
==
'direct'
:
# need v2
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_DIRECT'
elif
self
.
algo
==
'fft'
:
elif
self
.
algo
==
'fft'
:
# need v3
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_FFT'
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_FFT'
elif
self
.
algo
==
'fft_tiling'
:
# need v4
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING'
elif
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
]:
elif
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
]:
# The convolution implementation should be choosen according
# The convolution implementation should be choosen according
# to a heuristic
# to a heuristic
...
@@ -652,8 +662,8 @@ class GpuDnnConvGradW(DnnBase, COp):
...
@@ -652,8 +662,8 @@ class GpuDnnConvGradW(DnnBase, COp):
The convolution descriptor.
The convolution descriptor.
workmem
workmem
*deprecated*, use parameter algo instead.
*deprecated*, use parameter algo instead.
algo : {'none', 'deterministic', 'fft', 'guess_once', 'guess_on_shape_change', 'time_once', 'time_on_shape_change'}
algo : {'none', 'deterministic', 'fft', '
small', '
guess_once', 'guess_on_shape_change', 'time_once', 'time_on_shape_change'}
Default is the value of :attr:`config.dnn.conv.algo_bwd`.
Default is the value of :attr:`config.dnn.conv.algo_bwd
_filter
`.
"""
"""
...
@@ -671,15 +681,16 @@ class GpuDnnConvGradW(DnnBase, COp):
...
@@ -671,15 +681,16 @@ class GpuDnnConvGradW(DnnBase, COp):
self
.
algo
=
workmem
self
.
algo
=
workmem
else
:
else
:
if
algo
is
None
:
if
algo
is
None
:
algo
=
config
.
dnn
.
conv
.
algo_bwd
algo
=
config
.
dnn
.
conv
.
algo_bwd
_filter
self
.
algo
=
algo
self
.
algo
=
algo
self
.
inplace
=
inplace
self
.
inplace
=
inplace
if
self
.
inplace
:
if
self
.
inplace
:
self
.
destroy_map
=
{
0
:
[
2
]}
self
.
destroy_map
=
{
0
:
[
2
]}
assert
self
.
algo
in
[
'none'
,
'deterministic'
,
'fft'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
assert
self
.
algo
in
[
'none'
,
'deterministic'
,
'fft'
,
'small'
,
'time_on_shape_change'
]
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
]
def
__setstate__
(
self
,
d
):
def
__setstate__
(
self
,
d
):
self
.
__dict__
.
update
(
d
)
self
.
__dict__
.
update
(
d
)
...
@@ -687,7 +698,7 @@ class GpuDnnConvGradW(DnnBase, COp):
...
@@ -687,7 +698,7 @@ class GpuDnnConvGradW(DnnBase, COp):
if
hasattr
(
self
,
'workmem'
):
if
hasattr
(
self
,
'workmem'
):
self
.
algo
=
self
.
workmem
self
.
algo
=
self
.
workmem
else
:
else
:
self
.
algo
=
config
.
dnn
.
conv
.
algo_bwd
self
.
algo
=
config
.
dnn
.
conv
.
algo_bwd
_filter
if
not
hasattr
(
self
,
'inplace'
):
if
not
hasattr
(
self
,
'inplace'
):
self
.
inplace
=
False
self
.
inplace
=
False
...
@@ -724,11 +735,15 @@ class GpuDnnConvGradW(DnnBase, COp):
...
@@ -724,11 +735,15 @@ class GpuDnnConvGradW(DnnBase, COp):
alg
=
"0"
alg
=
"0"
else
:
else
:
if
self
.
algo
==
'none'
:
if
self
.
algo
==
'none'
:
# non-deterministic
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0'
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0'
elif
self
.
algo
==
'deterministic'
:
elif
self
.
algo
==
'deterministic'
:
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1'
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1'
elif
self
.
algo
==
'fft'
:
elif
self
.
algo
==
'fft'
:
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT'
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT'
elif
self
.
algo
==
'small'
:
# need v3, non-deterministic, small workspace
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3'
elif
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
]:
elif
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
]:
# The convolution implementation should be chosen according
# The convolution implementation should be chosen according
# to a heuristic
# to a heuristic
...
@@ -788,7 +803,7 @@ class GpuDnnConv3dGradW(GpuDnnConvGradW):
...
@@ -788,7 +803,7 @@ class GpuDnnConv3dGradW(GpuDnnConvGradW):
:param workmem:
:param workmem:
*deprecated*, use parameter algo instead.
*deprecated*, use parameter algo instead.
:param algo: ['none', 'guess_once', 'guess_on_shape_change', 'time_once', 'time_on_shape_change']
:param algo: ['none', 'guess_once', 'guess_on_shape_change', 'time_once', 'time_on_shape_change']
Default is the value of :attr:`config.dnn.conv.algo_bwd`.
Default is the value of :attr:`config.dnn.conv.algo_bwd
_filter
`.
"""
"""
__props__
=
(
'algo'
,
'inplace'
,)
__props__
=
(
'algo'
,
'inplace'
,)
...
@@ -856,11 +871,11 @@ class GpuDnnConvGradI(DnnBase, COp):
...
@@ -856,11 +871,11 @@ class GpuDnnConvGradI(DnnBase, COp):
workmem
workmem
*deprecated*, use parameter algo instead.
*deprecated*, use parameter algo instead.
algo
algo
['none', 'deterministic', 'fft', 'guess_once',
['none', 'deterministic', 'fft', '
fft_tiling', '
guess_once',
'guess_on_shape_change', 'time_once',
'guess_on_shape_change', 'time_once',
'time_on_shape_change']
'time_on_shape_change']
Default is the value of :attr:`config.dnn.conv.algo_bwd`.
Default is the value of :attr:`config.dnn.conv.algo_bwd
_data
`.
"""
"""
...
@@ -879,15 +894,23 @@ class GpuDnnConvGradI(DnnBase, COp):
...
@@ -879,15 +894,23 @@ class GpuDnnConvGradI(DnnBase, COp):
self
.
algo
=
workmem
self
.
algo
=
workmem
else
:
else
:
if
algo
is
None
:
if
algo
is
None
:
algo
=
config
.
dnn
.
conv
.
algo_bwd
algo
=
config
.
dnn
.
conv
.
algo_bwd
_data
self
.
algo
=
algo
self
.
algo
=
algo
self
.
inplace
=
inplace
self
.
inplace
=
inplace
if
self
.
inplace
:
if
self
.
inplace
:
self
.
destroy_map
=
{
0
:
[
2
]}
self
.
destroy_map
=
{
0
:
[
2
]}
assert
self
.
algo
in
[
'none'
,
'deterministic'
,
'fft'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
# The small-workspace implementation is only available from CuDNN V4
'time_on_shape_change'
]
# onward.
if
version
()
<
(
4000
,
4000
):
if
self
.
algo
==
'fft_tiling'
:
raise
RuntimeError
(
"CuDNN's tiled-FFT convolution requires "
"CuDNN v4 or more recent"
)
assert
self
.
algo
in
[
'none'
,
'deterministic'
,
'fft'
,
'fft_tiling'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
]
def
__setstate__
(
self
,
d
):
def
__setstate__
(
self
,
d
):
self
.
__dict__
.
update
(
d
)
self
.
__dict__
.
update
(
d
)
...
@@ -895,7 +918,7 @@ class GpuDnnConvGradI(DnnBase, COp):
...
@@ -895,7 +918,7 @@ class GpuDnnConvGradI(DnnBase, COp):
if
hasattr
(
self
,
'workmem'
):
if
hasattr
(
self
,
'workmem'
):
self
.
algo
=
self
.
workmem
self
.
algo
=
self
.
workmem
else
:
else
:
self
.
algo
=
config
.
dnn
.
conv
.
algo_bwd
self
.
algo
=
config
.
dnn
.
conv
.
algo_bwd
_data
if
not
hasattr
(
self
,
'inplace'
):
if
not
hasattr
(
self
,
'inplace'
):
self
.
inplace
=
False
self
.
inplace
=
False
...
@@ -936,7 +959,11 @@ class GpuDnnConvGradI(DnnBase, COp):
...
@@ -936,7 +959,11 @@ class GpuDnnConvGradI(DnnBase, COp):
elif
self
.
algo
==
'deterministic'
:
elif
self
.
algo
==
'deterministic'
:
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_1'
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_1'
elif
self
.
algo
==
'fft'
:
elif
self
.
algo
==
'fft'
:
# need v3, big workspace
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT'
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT'
elif
self
.
algo
==
'fft_tiling'
:
# need v4, big workspace, but less then fft
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING'
elif
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
]:
elif
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
]:
# The convolution implementation should be chosen according
# The convolution implementation should be chosen according
# to a heuristic
# to a heuristic
...
@@ -998,7 +1025,7 @@ class GpuDnnConv3dGradI(GpuDnnConvGradI):
...
@@ -998,7 +1025,7 @@ class GpuDnnConv3dGradI(GpuDnnConvGradI):
:param algo: ['none', 'guess_once', 'guess_on_shape_change',
:param algo: ['none', 'guess_once', 'guess_on_shape_change',
'time_once', 'time_on_shape_change']
'time_once', 'time_on_shape_change']
Default is the value of :attr:`config.dnn.conv.algo_bwd`.
Default is the value of :attr:`config.dnn.conv.algo_bwd
_data
`.
"""
"""
__props__
=
(
'algo'
,
'inplace'
,)
__props__
=
(
'algo'
,
'inplace'
,)
...
@@ -1055,7 +1082,8 @@ class GpuDnnConv3dGradI(GpuDnnConvGradI):
...
@@ -1055,7 +1082,8 @@ class GpuDnnConv3dGradI(GpuDnnConvGradI):
def
dnn_conv
(
img
,
kerns
,
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
def
dnn_conv
(
img
,
kerns
,
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
conv_mode
=
'conv'
,
direction_hint
=
None
,
workmem
=
None
,
algo
=
None
):
conv_mode
=
'conv'
,
direction_hint
=
None
,
workmem
=
None
,
algo
=
None
,
precision
=
None
):
"""
"""
GPU convolution using cuDNN from NVIDIA.
GPU convolution using cuDNN from NVIDIA.
...
@@ -1094,9 +1122,20 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -1094,9 +1122,20 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
Convolution implementation to use. Some of its values may require certain
Convolution implementation to use. Some of its values may require certain
versions of CuDNN to be installed. Default is the value of
versions of CuDNN to be installed. Default is the value of
:attr:`config.dnn.conv.algo_fwd`.
:attr:`config.dnn.conv.algo_fwd`.
precision : {'as_input', 'float16', 'float32', 'float64'}
Description of the dtype in which the computation of the convolution
should be done. Possible values are 'as_input', 'float16', 'float32'
and 'float64'. Default is the value of
:attr:`config.dnn.conv.precision`.
"""
"""
# Establish dtype in which to perform the computation of the convolution
if
precision
is
None
:
precision
=
theano
.
config
.
dnn
.
conv
.
precision
if
precision
==
'as_input'
:
precision
=
theano
.
scalar
.
upcast
(
img
.
dtype
,
kerns
.
dtype
)
# Check if deprecated param 'workmem' is used
# Check if deprecated param 'workmem' is used
if
workmem
is
not
None
:
if
workmem
is
not
None
:
warnings
.
warn
((
"dnn_conv: parameter 'workmem' is deprecated. Use "
warnings
.
warn
((
"dnn_conv: parameter 'workmem' is deprecated. Use "
...
@@ -1123,7 +1162,8 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -1123,7 +1162,8 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
out
=
gpu_alloc_empty
(
shape_i
(
kerns
,
1
,
fgraph
),
out
=
gpu_alloc_empty
(
shape_i
(
kerns
,
1
,
fgraph
),
shape_i
(
img
,
1
,
fgraph
),
shape2
,
shape3
)
shape_i
(
img
,
1
,
fgraph
),
shape2
,
shape3
)
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
conv_mode
=
'cross'
)(
img
.
shape
,
out
.
shape
)
conv_mode
=
'cross'
,
precision
=
precision
)(
img
.
shape
,
out
.
shape
)
conv
=
GpuDnnConvGradW
()(
img
,
kerns
,
out
,
desc
)
conv
=
GpuDnnConvGradW
()(
img
,
kerns
,
out
,
desc
)
return
as_cuda_ndarray_variable
(
conv
.
dimshuffle
(
1
,
0
,
2
,
3
))
return
as_cuda_ndarray_variable
(
conv
.
dimshuffle
(
1
,
0
,
2
,
3
))
...
@@ -1139,7 +1179,8 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -1139,7 +1179,8 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
out
=
gpu_alloc_empty
(
shape_i
(
img
,
0
,
fgraph
),
out
=
gpu_alloc_empty
(
shape_i
(
img
,
0
,
fgraph
),
shape_i
(
kerns
,
1
,
fgraph
),
shape2
,
shape3
)
shape_i
(
kerns
,
1
,
fgraph
),
shape2
,
shape3
)
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
conv_mode
=
conv_mode
)(
out
.
shape
,
kerns
.
shape
)
conv_mode
=
conv_mode
,
precision
=
precision
)(
out
.
shape
,
kerns
.
shape
)
return
GpuDnnConvGradI
()(
kerns
,
img
,
out
,
desc
)
return
GpuDnnConvGradI
()(
kerns
,
img
,
out
,
desc
)
# Standard case: We use GpuDnnConv with suitable padding.
# Standard case: We use GpuDnnConv with suitable padding.
...
@@ -1148,7 +1189,8 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -1148,7 +1189,8 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
img
=
gpu_contiguous
(
img
)
img
=
gpu_contiguous
(
img
)
kerns
=
gpu_contiguous
(
kerns
)
kerns
=
gpu_contiguous
(
kerns
)
desc
=
GpuDnnConvDesc
(
border_mode
=
border_mode
,
subsample
=
subsample
,
desc
=
GpuDnnConvDesc
(
border_mode
=
border_mode
,
subsample
=
subsample
,
conv_mode
=
conv_mode
)(
img
.
shape
,
kerns
.
shape
)
conv_mode
=
conv_mode
,
precision
=
precision
)(
img
.
shape
,
kerns
.
shape
)
desc_op
=
desc
.
owner
.
op
desc_op
=
desc
.
owner
.
op
out_shp
=
GpuDnnConv
.
get_out_shape
(
img
.
shape
,
kerns
.
shape
,
out_shp
=
GpuDnnConv
.
get_out_shape
(
img
.
shape
,
kerns
.
shape
,
desc_op
.
border_mode
,
desc_op
.
border_mode
,
...
@@ -1159,7 +1201,7 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -1159,7 +1201,7 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
def
dnn_conv3d
(
img
,
kerns
,
border_mode
=
'valid'
,
subsample
=
(
1
,
1
,
1
),
def
dnn_conv3d
(
img
,
kerns
,
border_mode
=
'valid'
,
subsample
=
(
1
,
1
,
1
),
conv_mode
=
'conv'
,
direction_hint
=
None
,
workmem
=
None
,
conv_mode
=
'conv'
,
direction_hint
=
None
,
workmem
=
None
,
algo
=
'none'
):
algo
=
'none'
,
precision
=
None
):
"""
"""
GPU convolution using cuDNN from NVIDIA.
GPU convolution using cuDNN from NVIDIA.
...
@@ -1186,6 +1228,10 @@ def dnn_conv3d(img, kerns, border_mode='valid', subsample=(1, 1, 1),
...
@@ -1186,6 +1228,10 @@ def dnn_conv3d(img, kerns, border_mode='valid', subsample=(1, 1, 1),
:param algo: convolution implementation to use. Only 'none' is implemented
:param algo: convolution implementation to use. Only 'none' is implemented
for the conv3d. Default is the value of
for the conv3d. Default is the value of
:attr:`config.dnn.conv.algo_fwd`.
:attr:`config.dnn.conv.algo_fwd`.
:param precision : dtype in which the computation of the convolution
should be done. Possible values are 'as_input', 'float16', 'float32'
and 'float64'. Default is the value of
:attr:`config.dnn.conv.precision`.
:warning: The cuDNN library only works with GPU that have a compute
:warning: The cuDNN library only works with GPU that have a compute
capability of 3.0 or higer. This means that older GPU will not
capability of 3.0 or higer. This means that older GPU will not
...
@@ -1194,6 +1240,12 @@ def dnn_conv3d(img, kerns, border_mode='valid', subsample=(1, 1, 1),
...
@@ -1194,6 +1240,12 @@ def dnn_conv3d(img, kerns, border_mode='valid', subsample=(1, 1, 1),
"""
"""
# Establish dtype in which to perform the computation of the convolution
if
precision
is
None
:
precision
=
theano
.
config
.
dnn
.
conv
.
precision
if
precision
==
'as_input'
:
precision
=
theano
.
scalar
.
upcast
(
img
.
dtype
,
kerns
.
dtype
)
# Check if deprecated param 'workmem' is used
# Check if deprecated param 'workmem' is used
if
workmem
is
not
None
:
if
workmem
is
not
None
:
warnings
.
warn
((
"dnn_conv3d: parameter 'workmem' is deprecated. Use "
warnings
.
warn
((
"dnn_conv3d: parameter 'workmem' is deprecated. Use "
...
@@ -1221,7 +1273,8 @@ def dnn_conv3d(img, kerns, border_mode='valid', subsample=(1, 1, 1),
...
@@ -1221,7 +1273,8 @@ def dnn_conv3d(img, kerns, border_mode='valid', subsample=(1, 1, 1),
out
=
gpu_alloc_empty
(
shape_i
(
kerns
,
1
,
fgraph
),
out
=
gpu_alloc_empty
(
shape_i
(
kerns
,
1
,
fgraph
),
shape_i
(
img
,
1
,
fgraph
),
shape2
,
shape3
,
shape4
)
shape_i
(
img
,
1
,
fgraph
),
shape2
,
shape3
,
shape4
)
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
,
1
),
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
,
1
),
conv_mode
=
'cross'
)(
img
.
shape
,
out
.
shape
)
conv_mode
=
'cross'
,
precision
=
precision
)(
img
.
shape
,
out
.
shape
)
conv
=
GpuDnnConv3dGradW
()(
img
,
kerns
,
out
,
desc
)
conv
=
GpuDnnConv3dGradW
()(
img
,
kerns
,
out
,
desc
)
return
as_cuda_ndarray_variable
(
conv
.
dimshuffle
(
1
,
0
,
2
,
3
,
4
))
return
as_cuda_ndarray_variable
(
conv
.
dimshuffle
(
1
,
0
,
2
,
3
,
4
))
...
@@ -1231,7 +1284,8 @@ def dnn_conv3d(img, kerns, border_mode='valid', subsample=(1, 1, 1),
...
@@ -1231,7 +1284,8 @@ def dnn_conv3d(img, kerns, border_mode='valid', subsample=(1, 1, 1),
img
=
gpu_contiguous
(
img
)
img
=
gpu_contiguous
(
img
)
kerns
=
gpu_contiguous
(
kerns
)
kerns
=
gpu_contiguous
(
kerns
)
desc
=
GpuDnnConvDesc
(
border_mode
=
border_mode
,
subsample
=
subsample
,
desc
=
GpuDnnConvDesc
(
border_mode
=
border_mode
,
subsample
=
subsample
,
conv_mode
=
conv_mode
)(
img
.
shape
,
kerns
.
shape
)
conv_mode
=
conv_mode
,
precision
=
precision
)(
img
.
shape
,
kerns
.
shape
)
desc_op
=
desc
.
owner
.
op
desc_op
=
desc
.
owner
.
op
out_shp
=
GpuDnnConv3d
.
get_out_shape
(
img
.
shape
,
kerns
.
shape
,
out_shp
=
GpuDnnConv3d
.
get_out_shape
(
img
.
shape
,
kerns
.
shape
,
desc_op
.
border_mode
,
desc_op
.
border_mode
,
...
...
theano/sandbox/cuda/dnn_conv_base.c
浏览文件 @
4ad36ddc
...
@@ -15,11 +15,8 @@ int APPLY_SPECIFIC(previous_kerns_shape)[5];
...
@@ -15,11 +15,8 @@ int APPLY_SPECIFIC(previous_kerns_shape)[5];
int
APPLY_SPECIFIC
(
previous_output_shape
)[
5
];
int
APPLY_SPECIFIC
(
previous_output_shape
)[
5
];
bool
APPLY_SPECIFIC
(
previous_algo_set
);
bool
APPLY_SPECIFIC
(
previous_algo_set
);
cudnnConvolutionFwdAlgo_t
APPLY_SPECIFIC
(
previous_algo
);
cudnnConvolutionFwdAlgo_t
APPLY_SPECIFIC
(
previous_algo
);
#if defined(CUDNN_VERSION) && CUDNN_VERSION >= 3000
cudnnConvolutionBwdFilterAlgo_t
APPLY_SPECIFIC
(
previous_bwd_f_algo
);
cudnnConvolutionBwdFilterAlgo_t
APPLY_SPECIFIC
(
previous_bwd_f_algo
);
cudnnConvolutionBwdDataAlgo_t
APPLY_SPECIFIC
(
previous_bwd_d_algo
);
cudnnConvolutionBwdDataAlgo_t
APPLY_SPECIFIC
(
previous_bwd_d_algo
);
#endif
#section init_code_struct
#section init_code_struct
...
@@ -55,10 +52,8 @@ APPLY_SPECIFIC(previous_algo_set) = false;
...
@@ -55,10 +52,8 @@ APPLY_SPECIFIC(previous_algo_set) = false;
// Select default implementations for the case where the convolution
// Select default implementations for the case where the convolution
// implementations should be selected based on the size of the data.
// implementations should be selected based on the size of the data.
APPLY_SPECIFIC
(
previous_algo
)
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM
;
APPLY_SPECIFIC
(
previous_algo
)
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM
;
#if defined(CUDNN_VERSION) && CUDNN_VERSION >= 3000
APPLY_SPECIFIC
(
previous_bwd_f_algo
)
=
CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0
;
APPLY_SPECIFIC
(
previous_bwd_f_algo
)
=
CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0
;
APPLY_SPECIFIC
(
previous_bwd_d_algo
)
=
CUDNN_CONVOLUTION_BWD_DATA_ALGO_0
;
APPLY_SPECIFIC
(
previous_bwd_d_algo
)
=
CUDNN_CONVOLUTION_BWD_DATA_ALGO_0
;
#endif
#section cleanup_code_struct
#section cleanup_code_struct
...
...
theano/sandbox/cuda/dnn_fwd.c
浏览文件 @
4ad36ddc
...
@@ -81,7 +81,6 @@ APPLY_SPECIFIC(conv_fwd)(CudaNdarray *input, CudaNdarray *kerns,
...
@@ -81,7 +81,6 @@ APPLY_SPECIFIC(conv_fwd)(CudaNdarray *input, CudaNdarray *kerns,
// CuDNN time every implementation and choose the best one.
// CuDNN time every implementation and choose the best one.
if
(
CHOOSE_ALGO_TIME
)
if
(
CHOOSE_ALGO_TIME
)
{
{
#if defined(CUDNN_VERSION) && CUDNN_VERSION >= 3000
// Time the different implementations to choose the best one
// Time the different implementations to choose the best one
int
requestedCount
=
1
;
int
requestedCount
=
1
;
int
count
;
int
count
;
...
@@ -102,7 +101,6 @@ APPLY_SPECIFIC(conv_fwd)(CudaNdarray *input, CudaNdarray *kerns,
...
@@ -102,7 +101,6 @@ APPLY_SPECIFIC(conv_fwd)(CudaNdarray *input, CudaNdarray *kerns,
}
}
chosen_algo
=
choosen_algo_perf
.
algo
;
chosen_algo
=
choosen_algo_perf
.
algo
;
#endif
}
}
else
else
{
{
...
@@ -161,24 +159,28 @@ APPLY_SPECIFIC(conv_fwd)(CudaNdarray *input, CudaNdarray *kerns,
...
@@ -161,24 +159,28 @@ APPLY_SPECIFIC(conv_fwd)(CudaNdarray *input, CudaNdarray *kerns,
chosen_algo
=
CONV_ALGO
;
chosen_algo
=
CONV_ALGO
;
}
}
#if defined(CUDNN_VERSION) && CUDNN_VERSION >= 3000
// The FFT implementation (only in V3 and onward) does not support strides,
// The FFT implementation (only in V3 and onward) does not support strides,
// 1x1 filters or inputs with a spatial dimension larger than 1024.
// 1x1 filters or inputs with a spatial dimension larger than 1024.
// If the chosen implementation is FFT, validate that it can be used
// The tiled-FFT implementation (only in V4 onward) does not support
// on the current data and default on a safe implementation if it
// strides.
// If the chosen implementation is FFT or tiled-FFT, validate that it can
// be used on the current data and default on a safe implementation if it
// can't.
// can't.
// Following code is 2d-specific, but it is fine as ftt is defined only for
// Following code is 2d-specific, but it is fine as FFT and tiled-FFT are
// 2d-filters
// defined only for 2d-filters
if
(
chosen_algo
==
CUDNN_CONVOLUTION_FWD_ALGO_FFT
&&
nb_dim
==
4
)
if
((
chosen_algo
==
CUDNN_CONVOLUTION_FWD_ALGO_FFT
||
chosen_algo
==
CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING
)
&&
nb_dim
==
4
)
{
{
// Extract the properties of the convolution descriptor
// Extract the properties of the convolution descriptor
int
pad_h
,
pad_w
,
stride_v
,
stride_h
,
upscale_x
,
upscale_y
;
int
nd
;
int
pad
[
2
];
int
stride
[
2
];
int
upscale
[
2
];
cudnnConvolutionMode_t
mode
;
cudnnConvolutionMode_t
mode
;
err
=
cudnnGetConvolution2dDescriptor
(
desc
,
&
pad_h
,
&
pad_w
,
cudnnDataType_t
data_type
;
&
stride_v
,
&
stride_h
,
err
=
cudnnGetConvolutionNdDescriptor_v3
(
desc
,
2
,
&
nd
,
pad
,
stride
,
&
upscale_x
,
&
upscale_y
,
upscale
,
&
mode
,
&
data_type
);
&
mode
);
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
PyErr_Format
(
PyExc_RuntimeError
,
PyErr_Format
(
PyExc_RuntimeError
,
...
@@ -197,36 +199,23 @@ APPLY_SPECIFIC(conv_fwd)(CudaNdarray *input, CudaNdarray *kerns,
...
@@ -197,36 +199,23 @@ APPLY_SPECIFIC(conv_fwd)(CudaNdarray *input, CudaNdarray *kerns,
// Ensure that the selected implementation supports the requested
// Ensure that the selected implementation supports the requested
// convolution. Fall back to a safe implementation otherwise.
// convolution. Fall back to a safe implementation otherwise.
if
(
stride_v
!=
1
||
stride_h
!=
1
||
input_h
>
1024
||
if
(
chosen_algo
==
CUDNN_CONVOLUTION_FWD_ALGO_FFT
)
{
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
||
input_h
>
1024
||
input_w
>
1024
||
(
filter_h
==
1
&&
filter_w
==
1
))
input_w
>
1024
||
(
filter_h
==
1
&&
filter_w
==
1
))
{
{
chosen_algo
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM
;
chosen_algo
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM
;
}
}
}
}
#endif
else
{
#if defined(CUDNN_VERSION) && CUDNN_VERSION < 3000
// chosen_algo == CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING
// In versions before V3, CuDNN did not support kernels larger than the
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
)
// inputs in any spatial dimension, even if padding was used such that the
{
// padded inputs were larger than the kernels. If the kernels are larger
chosen_algo
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM
;
// then the inputs, raise an error message.
}
bool
shape_mismatch
=
false
;
for
(
int
i
=
2
;
i
<
nb_dim
;
i
++
){
shape_mismatch
=
shape_mismatch
||
(
CudaNdarray_HOST_DIMS
(
kerns
)[
i
]
>
CudaNdarray_HOST_DIMS
(
input
)[
i
]);
}
}
if
(
shape_mismatch
){
PyErr_Format
(
PyExc_RuntimeError
,
"GpuDnnConv: the current version of CuDNN does not support "
"kernels larger than the inputs in any spatial dimension, "
"even if the inputs are padded such that the padded inputs "
"are larger than the kernels. Update your installation of "
"CuDNN to V3 or more recent to solve the issue."
);
return
1
;
}
}
#endif
err
=
cudnnGetConvolutionForwardWorkspaceSize
(
_handle
,
err
=
cudnnGetConvolutionForwardWorkspaceSize
(
_handle
,
APPLY_SPECIFIC
(
input
),
APPLY_SPECIFIC
(
input
),
...
...
theano/sandbox/cuda/dnn_gi.c
浏览文件 @
4ad36ddc
...
@@ -33,7 +33,6 @@ APPLY_SPECIFIC(conv_gi)(CudaNdarray *kerns, CudaNdarray *output,
...
@@ -33,7 +33,6 @@ APPLY_SPECIFIC(conv_gi)(CudaNdarray *kerns, CudaNdarray *output,
if
(
c_set_tensorNd
(
*
input
,
APPLY_SPECIFIC
(
input
))
==
-
1
)
if
(
c_set_tensorNd
(
*
input
,
APPLY_SPECIFIC
(
input
))
==
-
1
)
return
1
;
return
1
;
#if defined(CUDNN_VERSION) && CUDNN_VERSION >= 3000
{
{
size_t
worksize
;
size_t
worksize
;
void
*
workspace
;
void
*
workspace
;
...
@@ -159,21 +158,28 @@ APPLY_SPECIFIC(conv_gi)(CudaNdarray *kerns, CudaNdarray *output,
...
@@ -159,21 +158,28 @@ APPLY_SPECIFIC(conv_gi)(CudaNdarray *kerns, CudaNdarray *output,
chosen_algo
=
CONV_ALGO
;
chosen_algo
=
CONV_ALGO
;
}
}
// The FFT implementation (only in
v
3 and onward) does not support strides,
// The FFT implementation (only in
V
3 and onward) does not support strides,
// 1x1 filters or inputs with a spatial dimension larger than 1024.
// 1x1 filters or inputs with a spatial dimension larger than 1024.
// If the chosen implementation is FFT, validate that it can be used
// The tiled-FFT implementation (only in V4 onward) does not support
// on the current data and default on a safe implementation if it
// strides.
// If the chosen implementation is FFT or tiled-FFT, validate that it can
// be used on the current data and default on a safe implementation if it
// can't.
// can't.
if
(
chosen_algo
==
CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT
&&
nb_dim
==
4
)
// Following code is 2d-specific, but it is fine as FFT and tiled-FFT are
// defined only for 2d-filters
if
((
chosen_algo
==
CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING
||
chosen_algo
==
CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT
)
&&
nb_dim
==
4
)
{
{
// Extract the properties of the convolution descriptor
// Extract the properties of the convolution descriptor
int
pad_h
,
pad_w
,
stride_v
,
stride_h
,
upscale_x
,
upscale_y
;
int
nd
;
int
pad
[
2
];
int
stride
[
2
];
int
upscale
[
2
];
cudnnConvolutionMode_t
mode
;
cudnnConvolutionMode_t
mode
;
err
=
cudnnGetConvolution2dDescriptor
(
desc
,
&
pad_h
,
&
pad_w
,
cudnnDataType_t
data_type
;
&
stride_v
,
&
stride_h
,
err
=
cudnnGetConvolutionNdDescriptor_v3
(
desc
,
2
,
&
nd
,
pad
,
stride
,
&
upscale_x
,
&
upscale_y
,
upscale
,
&
mode
,
&
data_type
);
&
mode
);
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
PyErr_Format
(
PyExc_RuntimeError
,
PyErr_Format
(
PyExc_RuntimeError
,
...
@@ -192,12 +198,23 @@ APPLY_SPECIFIC(conv_gi)(CudaNdarray *kerns, CudaNdarray *output,
...
@@ -192,12 +198,23 @@ APPLY_SPECIFIC(conv_gi)(CudaNdarray *kerns, CudaNdarray *output,
// Ensure that the selected implementation supports the requested
// Ensure that the selected implementation supports the requested
// convolution. Fall back to a safe implementation otherwise.
// convolution. Fall back to a safe implementation otherwise.
if
(
stride_v
!=
1
||
stride_h
!=
1
||
input_h
>
1024
||
if
(
chosen_algo
==
CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT
)
{
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
||
input_h
>
1024
||
input_w
>
1024
||
(
filter_h
==
1
&&
filter_w
==
1
))
input_w
>
1024
||
(
filter_h
==
1
&&
filter_w
==
1
))
{
{
chosen_algo
=
CUDNN_CONVOLUTION_BWD_DATA_ALGO_0
;
chosen_algo
=
CUDNN_CONVOLUTION_BWD_DATA_ALGO_0
;
}
}
}
}
else
{
// chosen_algo == CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
)
{
chosen_algo
=
CUDNN_CONVOLUTION_BWD_DATA_ALGO_0
;
}
}
}
// Infer required workspace size from the chosen implementation
// Infer required workspace size from the chosen implementation
err
=
cudnnGetConvolutionBackwardDataWorkspaceSize
(
_handle
,
err
=
cudnnGetConvolutionBackwardDataWorkspaceSize
(
_handle
,
...
@@ -231,16 +248,6 @@ APPLY_SPECIFIC(conv_gi)(CudaNdarray *kerns, CudaNdarray *output,
...
@@ -231,16 +248,6 @@ APPLY_SPECIFIC(conv_gi)(CudaNdarray *kerns, CudaNdarray *output,
(
void
*
)
&
beta
,
(
void
*
)
&
beta
,
APPLY_SPECIFIC
(
input
),
CudaNdarray_DEV_DATA
(
*
input
));
APPLY_SPECIFIC
(
input
),
CudaNdarray_DEV_DATA
(
*
input
));
}
}
#else
err
=
cudnnConvolutionBackwardData
(
_handle
,
(
void
*
)
&
alpha
,
APPLY_SPECIFIC
(
kerns
),
CudaNdarray_DEV_DATA
(
kerns
),
APPLY_SPECIFIC
(
output
),
CudaNdarray_DEV_DATA
(
output
),
desc
,
(
void
*
)
&
beta
,
APPLY_SPECIFIC
(
input
),
CudaNdarray_DEV_DATA
(
*
input
));
#endif
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
PyErr_Format
(
PyExc_RuntimeError
,
"GpuDnnConvGradI: error doing operation: %s"
,
PyErr_Format
(
PyExc_RuntimeError
,
"GpuDnnConvGradI: error doing operation: %s"
,
...
...
theano/sandbox/cuda/dnn_gw.c
浏览文件 @
4ad36ddc
...
@@ -33,7 +33,6 @@ APPLY_SPECIFIC(conv_gw)(CudaNdarray *input, CudaNdarray *output,
...
@@ -33,7 +33,6 @@ APPLY_SPECIFIC(conv_gw)(CudaNdarray *input, CudaNdarray *output,
if
(
c_set_filterNd
(
*
kerns
,
APPLY_SPECIFIC
(
kerns
))
==
-
1
)
if
(
c_set_filterNd
(
*
kerns
,
APPLY_SPECIFIC
(
kerns
))
==
-
1
)
return
1
;
return
1
;
#if defined(CUDNN_VERSION) && CUDNN_VERSION >= 3000
{
{
size_t
worksize
;
size_t
worksize
;
void
*
workspace
;
void
*
workspace
;
...
@@ -168,12 +167,14 @@ APPLY_SPECIFIC(conv_gw)(CudaNdarray *input, CudaNdarray *output,
...
@@ -168,12 +167,14 @@ APPLY_SPECIFIC(conv_gw)(CudaNdarray *input, CudaNdarray *output,
{
{
// Extract the properties of the convolution descriptor
// Extract the properties of the convolution descriptor
int
pad_h
,
pad_w
,
stride_v
,
stride_h
,
upscale_x
,
upscale_y
;
int
nd
;
int
pad
[
2
];
int
stride
[
2
];
int
upscale
[
2
];
cudnnConvolutionMode_t
mode
;
cudnnConvolutionMode_t
mode
;
err
=
cudnnGetConvolution2dDescriptor
(
desc
,
&
pad_h
,
&
pad_w
,
cudnnDataType_t
data_type
;
&
stride_v
,
&
stride_h
,
err
=
cudnnGetConvolutionNdDescriptor_v3
(
desc
,
2
,
&
nd
,
pad
,
stride
,
&
upscale_x
,
&
upscale_y
,
upscale
,
&
mode
,
&
data_type
);
&
mode
);
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
PyErr_Format
(
PyExc_RuntimeError
,
PyErr_Format
(
PyExc_RuntimeError
,
...
@@ -192,7 +193,7 @@ APPLY_SPECIFIC(conv_gw)(CudaNdarray *input, CudaNdarray *output,
...
@@ -192,7 +193,7 @@ APPLY_SPECIFIC(conv_gw)(CudaNdarray *input, CudaNdarray *output,
// Ensure that the selected implementation supports the requested
// Ensure that the selected implementation supports the requested
// convolution. Fall back to a safe implementation otherwise.
// convolution. Fall back to a safe implementation otherwise.
if
(
stride
_v
!=
1
||
stride_h
!=
1
||
input_h
>
1024
||
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
||
input_h
>
1024
||
input_w
>
1024
||
(
filter_h
==
1
&&
filter_w
==
1
))
input_w
>
1024
||
(
filter_h
==
1
&&
filter_w
==
1
))
{
{
chosen_algo
=
CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0
;
chosen_algo
=
CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0
;
...
@@ -232,16 +233,6 @@ APPLY_SPECIFIC(conv_gw)(CudaNdarray *input, CudaNdarray *output,
...
@@ -232,16 +233,6 @@ APPLY_SPECIFIC(conv_gw)(CudaNdarray *input, CudaNdarray *output,
APPLY_SPECIFIC
(
kerns
),
CudaNdarray_DEV_DATA
(
*
kerns
));
APPLY_SPECIFIC
(
kerns
),
CudaNdarray_DEV_DATA
(
*
kerns
));
}
}
#else
err
=
cudnnConvolutionBackwardFilter
(
_handle
,
(
void
*
)
&
alpha
,
APPLY_SPECIFIC
(
input
),
CudaNdarray_DEV_DATA
(
input
),
APPLY_SPECIFIC
(
output
),
CudaNdarray_DEV_DATA
(
output
),
desc
,
(
void
*
)
&
beta
,
APPLY_SPECIFIC
(
kerns
),
CudaNdarray_DEV_DATA
(
*
kerns
));
#endif
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
PyErr_Format
(
PyExc_RuntimeError
,
"GpuDnnConvGradW: error doing operation: %s"
,
PyErr_Format
(
PyExc_RuntimeError
,
"GpuDnnConvGradW: error doing operation: %s"
,
...
...
theano/sandbox/gpuarray/conv_desc.c
浏览文件 @
4ad36ddc
...
@@ -29,7 +29,7 @@ int APPLY_SPECIFIC(conv_desc)(PyArrayObject *filt_shp,
...
@@ -29,7 +29,7 @@ int APPLY_SPECIFIC(conv_desc)(PyArrayObject *filt_shp,
return
-
1
;
return
-
1
;
}
}
err
=
cudnnSetConvolutionNdDescriptor
(
*
desc
,
NB_DIMS
,
pad
,
strides
,
upscale
,
err
=
cudnnSetConvolutionNdDescriptor
_v3
(
*
desc
,
NB_DIMS
,
pad
,
strides
,
CONV_MODE
);
upscale
,
CONV_MODE
,
PRECISION
);
return
0
;
return
0
;
}
}
theano/sandbox/gpuarray/cudnn_helper.h
浏览文件 @
4ad36ddc
...
@@ -13,99 +13,12 @@ static inline int cudnnGetVersion() {
...
@@ -13,99 +13,12 @@ static inline int cudnnGetVersion() {
#include <assert.h>
#include <assert.h>
#if CUDNN_VERSION < 3000
// If needed, define element of the V4 interface in terms of elements of
//
Here we define the R3 API in terms of functions in the R2 interface
//
previous versions
// This is only for what we use
#if defined(CUDNN_VERSION) && CUDNN_VERSION < 4000
typedef
int
cudnnConvolutionBwdDataAlgo_t
;
#define CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING 5
#define CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING 3
#define CUDNN_CONVOLUTION_BWD_DATA_ALGO_0 0
#define CUDNN_CONVOLUTION_BWD_DATA_ALGO_1 1
#define CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT 2
static
cudnnStatus_t
cudnnGetConvolutionBackwardDataWorkspaceSize
(
cudnnHandle_t
handle
,
const
cudnnFilterDescriptor_t
filterDesc
,
const
cudnnTensorDescriptor_t
diffDesc
,
const
cudnnConvolutionDescriptor_t
convDesc
,
const
cudnnTensorDescriptor_t
gradDesc
,
cudnnConvolutionBwdDataAlgo_t
algo
,
size_t
*
sizeInBytes
)
{
*
sizeInBytes
=
0
;
return
CUDNN_STATUS_SUCCESS
;
}
static
cudnnStatus_t
cudnnConvolutionBackwardData_v3
(
cudnnHandle_t
handle
,
const
void
*
alpha
,
const
cudnnFilterDescriptor_t
filterDesc
,
const
void
*
filterData
,
const
cudnnTensorDescriptor_t
diffDesc
,
const
void
*
diffData
,
const
cudnnConvolutionDescriptor_t
convDesc
,
cudnnConvolutionBwdDataAlgo_t
algo
,
void
*
workspace
,
size_t
workspaceSizeInBytes
,
const
void
*
beta
,
const
cudnnTensorDescriptor_t
gradDesc
,
void
*
gradData
)
{
return
cudnnConvolutionBackwardData
(
handle
,
alpha
,
filterDesc
,
filterData
,
diffDesc
,
diffData
,
convDesc
,
beta
,
gradDesc
,
gradData
);
}
typedef
int
cudnnConvolutionBwdFilterAlgo_t
;
#define CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0 0
#define CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1 1
#define CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT 2
static
cudnnStatus_t
cudnnGetConvolutionBackwardFilterWorkspaceSize
(
cudnnHandle_t
handle
,
const
cudnnTensorDescriptor_t
filterDesc
,
const
cudnnTensorDescriptor_t
diffDesc
,
const
cudnnConvolutionDescriptor_t
convDesc
,
const
cudnnFilterDescriptor_t
gradDesc
,
cudnnConvolutionBwdDataAlgo_t
algo
,
size_t
*
sizeInBytes
)
{
*
sizeInBytes
=
0
;
return
CUDNN_STATUS_SUCCESS
;
}
static
cudnnStatus_t
cudnnConvolutionBackwardFilter_v3
(
cudnnHandle_t
handle
,
const
void
*
alpha
,
const
cudnnTensorDescriptor_t
srcDesc
,
const
void
*
srcData
,
const
cudnnTensorDescriptor_t
diffDesc
,
const
void
*
diffData
,
const
cudnnConvolutionDescriptor_t
convDesc
,
cudnnConvolutionBwdFilterAlgo_t
algo
,
void
*
workspace
,
size_t
workspaceSizeInBytes
,
const
void
*
beta
,
const
cudnnFilterDescriptor_t
gradDesc
,
void
*
gradData
)
{
return
cudnnConvolutionBackwardFilter
(
handle
,
alpha
,
srcDesc
,
srcData
,
diffDesc
,
diffData
,
convDesc
,
beta
,
gradDesc
,
gradData
);
}
#endif
#endif
...
...
theano/sandbox/gpuarray/dnn.py
浏览文件 @
4ad36ddc
...
@@ -75,15 +75,11 @@ if ((err = cudnnCreate(&_handle)) != CUDNN_STATUS_SUCCESS) {
...
@@ -75,15 +75,11 @@ if ((err = cudnnCreate(&_handle)) != CUDNN_STATUS_SUCCESS) {
def
_dnn_check_version
():
def
_dnn_check_version
():
v
=
version
()
v
=
version
()
if
v
<
2000
:
if
v
<
3007
:
return
False
,
(
return
False
,
(
"You have an old release of CuDNN (or a release candidate) "
"You have an old release of CuDNN (or a release candidate) "
"that isn't supported. Please update to at least v
2
final "
"that isn't supported. Please update to at least v
3
final "
"version."
)
"version."
)
if
3000
<=
v
<
3007
:
return
False
,
(
"You have installed a release candidate of CuDNN v3. This "
"isn't supported. Please update to v3 final version."
)
return
True
,
None
return
True
,
None
...
@@ -241,7 +237,7 @@ class GpuDnnConvDesc(COp):
...
@@ -241,7 +237,7 @@ class GpuDnnConvDesc(COp):
"""
"""
__props__
=
(
'border_mode'
,
'subsample'
,
'conv_mode'
)
__props__
=
(
'border_mode'
,
'subsample'
,
'conv_mode'
,
'precision'
)
def
c_headers
(
self
):
def
c_headers
(
self
):
return
[
'cudnn.h'
,
'cudnn_helper.h'
]
return
[
'cudnn.h'
,
'cudnn_helper.h'
]
...
@@ -258,7 +254,8 @@ class GpuDnnConvDesc(COp):
...
@@ -258,7 +254,8 @@ class GpuDnnConvDesc(COp):
def
do_constant_folding
(
self
,
node
):
def
do_constant_folding
(
self
,
node
):
return
False
return
False
def
__init__
(
self
,
border_mode
,
subsample
=
(
1
,
1
),
conv_mode
=
'conv'
):
def
__init__
(
self
,
border_mode
,
subsample
=
(
1
,
1
),
conv_mode
=
'conv'
,
precision
=
"float32"
):
COp
.
__init__
(
self
,
[
"conv_desc.c"
],
"APPLY_SPECIFIC(conv_desc)"
)
COp
.
__init__
(
self
,
[
"conv_desc.c"
],
"APPLY_SPECIFIC(conv_desc)"
)
if
isinstance
(
border_mode
,
int
):
if
isinstance
(
border_mode
,
int
):
...
@@ -278,6 +275,9 @@ class GpuDnnConvDesc(COp):
...
@@ -278,6 +275,9 @@ class GpuDnnConvDesc(COp):
assert
conv_mode
in
(
'conv'
,
'cross'
)
assert
conv_mode
in
(
'conv'
,
'cross'
)
self
.
conv_mode
=
conv_mode
self
.
conv_mode
=
conv_mode
assert
precision
in
[
'float16'
,
'float32'
,
'float64'
]
self
.
precision
=
precision
def
make_node
(
self
,
kern_shape
):
def
make_node
(
self
,
kern_shape
):
if
kern_shape
.
type
.
ndim
!=
1
or
kern_shape
.
type
.
dtype
!=
'int64'
:
if
kern_shape
.
type
.
ndim
!=
1
or
kern_shape
.
type
.
dtype
!=
'int64'
:
raise
TypeError
(
'kern must be 1D shape tensor'
)
raise
TypeError
(
'kern must be 1D shape tensor'
)
...
@@ -315,11 +315,20 @@ class GpuDnnConvDesc(COp):
...
@@ -315,11 +315,20 @@ class GpuDnnConvDesc(COp):
else
:
else
:
sub2
=
'0'
sub2
=
'0'
if
self
.
precision
==
'float16'
:
precision
=
'CUDNN_DATA_HALF'
elif
self
.
precision
==
'float32'
:
precision
=
'CUDNN_DATA_FLOAT'
else
:
assert
self
.
precision
==
'float64'
precision
=
'CUDNN_DATA_DOUBLE'
return
[(
'NB_DIMS'
,
str
(
len
(
self
.
subsample
))),
return
[(
'NB_DIMS'
,
str
(
len
(
self
.
subsample
))),
(
'BORDER_MODE'
,
bmode
),
(
'BORDER_MODE'
,
bmode
),
(
'PAD_0'
,
pad0
),
(
'PAD_1'
,
pad1
),
(
'PAD_2'
,
pad2
),
(
'PAD_0'
,
pad0
),
(
'PAD_1'
,
pad1
),
(
'PAD_2'
,
pad2
),
(
'CONV_MODE'
,
conv_flag
),
(
'CONV_MODE'
,
conv_flag
),
(
'SUB_0'
,
sub0
),
(
'SUB_1'
,
sub1
),
(
'SUB_2'
,
sub2
)]
(
'SUB_0'
,
sub0
),
(
'SUB_1'
,
sub1
),
(
'SUB_2'
,
sub2
),
(
'PRECISION'
,
precision
)]
def
c_code_cache_version
(
self
):
def
c_code_cache_version
(
self
):
return
(
super
(
GpuDnnConvDesc
,
self
)
.
c_code_cache_version
(),
version
())
return
(
super
(
GpuDnnConvDesc
,
self
)
.
c_code_cache_version
(),
version
())
...
@@ -353,7 +362,8 @@ class GpuDnnConv(DnnBase):
...
@@ -353,7 +362,8 @@ class GpuDnnConv(DnnBase):
kernel
kernel
descr
descr
The convolution descriptor.
The convolution descriptor.
algo : {'small', 'none', 'large', 'fft', 'guess_once', 'guess_on_shape_change', 'time_once', 'time_on_shape_change'}
algo : {'small', 'none', 'large', 'fft', 'fft_tiling', 'guess_once',
'guess_on_shape_change', 'time_once', 'time_on_shape_change'}
Default is the value of :attr:`config.dnn.conv.algo_fwd`.
Default is the value of :attr:`config.dnn.conv.algo_fwd`.
"""
"""
...
@@ -382,9 +392,15 @@ class GpuDnnConv(DnnBase):
...
@@ -382,9 +392,15 @@ class GpuDnnConv(DnnBase):
elif
self
.
algo
in
[
'time_once'
,
'time_on_shape_change'
]:
elif
self
.
algo
in
[
'time_once'
,
'time_on_shape_change'
]:
raise
RuntimeError
(
"CuDNN convolution timing requires CuDNN v3"
)
raise
RuntimeError
(
"CuDNN convolution timing requires CuDNN v3"
)
assert
self
.
algo
in
[
'none'
,
'small'
,
'large'
,
'fft'
,
'guess_once'
,
# The fft_tiling implementation is only available from CuDNN V4 onward
'guess_on_shape_change'
,
'time_once'
,
if
version
()
<
4000
:
'time_on_shape_change'
]
if
self
.
algo
==
'fft_tiling'
:
raise
RuntimeError
(
"CuDNN tiled-FFT convolution requires "
"CuDNN v4 or more recent"
)
assert
self
.
algo
in
[
'none'
,
'small'
,
'large'
,
'fft'
,
'fft_tiling'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
]
def
__setstate__
(
self
,
d
):
def
__setstate__
(
self
,
d
):
self
.
__dict__
.
update
(
d
)
self
.
__dict__
.
update
(
d
)
...
@@ -408,8 +424,13 @@ class GpuDnnConv(DnnBase):
...
@@ -408,8 +424,13 @@ class GpuDnnConv(DnnBase):
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM'
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM'
elif
self
.
algo
==
'large'
:
elif
self
.
algo
==
'large'
:
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_GEMM'
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_GEMM'
elif
self
.
algo
==
'direct'
:
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_DIRECT'
elif
self
.
algo
==
'fft'
:
elif
self
.
algo
==
'fft'
:
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_FFT'
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_FFT'
elif
self
.
algo
==
'fft_tiling'
:
# need v4
alg
=
'CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING'
defs
.
append
((
'CONV_ALGO'
,
alg
))
defs
.
append
((
'CONV_ALGO'
,
alg
))
if
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
,
if
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
,
...
@@ -439,9 +460,10 @@ class GpuDnnConv(DnnBase):
...
@@ -439,9 +460,10 @@ class GpuDnnConv(DnnBase):
raise
TypeError
(
"The number of dimensions of "
raise
TypeError
(
"The number of dimensions of "
"img, kern and output must match"
)
"img, kern and output must match"
)
if
img
.
type
.
ndim
==
5
and
self
.
algo
==
'fft'
:
if
(
img
.
type
.
ndim
==
5
and
raise
ValueError
(
"convolution algo fft can't be used for "
self
.
algo
in
[
'small'
,
'large'
,
'fft'
,
'fft_tiling'
]):
"3d convolutions"
)
raise
ValueError
(
"convolution algo
%
s can't be used for "
"3d convolutions"
,
(
self
.
algo
,))
if
(
not
isinstance
(
desc
.
type
,
CDataType
)
or
if
(
not
isinstance
(
desc
.
type
,
CDataType
)
or
desc
.
type
.
ctype
!=
'cudnnConvolutionDescriptor_t'
):
desc
.
type
.
ctype
!=
'cudnnConvolutionDescriptor_t'
):
...
@@ -479,6 +501,14 @@ class GpuDnnConv(DnnBase):
...
@@ -479,6 +501,14 @@ class GpuDnnConv(DnnBase):
or scalar.
or scalar.
"""
"""
# if ishape and/or kshape are not tuples or list, but rather symbolic
# vectors, turn them into lists of symbolic scalars.
if
not
isinstance
(
ishape
,
(
list
,
tuple
)):
ishape
=
[
ishape
[
i
]
for
i
in
range
(
len
(
subsample
)
+
2
)]
if
not
isinstance
(
kshape
,
(
list
,
tuple
)):
kshape
=
[
kshape
[
i
]
for
i
in
range
(
len
(
subsample
)
+
2
)]
return
get_conv_output_shape
(
return
get_conv_output_shape
(
ishape
,
ishape
,
kshape
,
kshape
,
...
@@ -511,18 +541,19 @@ class GpuDnnConvGradW(DnnBase):
...
@@ -511,18 +541,19 @@ class GpuDnnConvGradW(DnnBase):
if
self
.
inplace
:
if
self
.
inplace
:
self
.
destroy_map
=
{
0
:
[
2
]}
self
.
destroy_map
=
{
0
:
[
2
]}
if
algo
is
None
:
if
algo
is
None
:
algo
=
config
.
dnn
.
conv
.
algo_bwd
algo
=
config
.
dnn
.
conv
.
algo_bwd
_filter
self
.
algo
=
algo
self
.
algo
=
algo
assert
self
.
algo
in
[
'none'
,
'deterministic'
,
'fft'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
assert
self
.
algo
in
[
'none'
,
'deterministic'
,
'fft'
,
'small'
,
'time_on_shape_change'
]
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
]
def
__setstate__
(
self
,
d
):
def
__setstate__
(
self
,
d
):
self
.
__dict__
.
update
(
d
)
self
.
__dict__
.
update
(
d
)
if
not
hasattr
(
self
,
'inplace'
):
if
not
hasattr
(
self
,
'inplace'
):
self
.
inplace
=
False
self
.
inplace
=
False
if
not
hasattr
(
self
,
'algo'
):
if
not
hasattr
(
self
,
'algo'
):
self
.
algo
=
config
.
dnn
.
conv
.
algo_bwd
self
.
algo
=
config
.
dnn
.
conv
.
algo_bwd
_filter
def
grad
(
self
,
inp
,
grads
):
def
grad
(
self
,
inp
,
grads
):
img
,
top
,
output
,
desc
,
alpha
,
beta
=
inp
img
,
top
,
output
,
desc
,
alpha
,
beta
=
inp
...
@@ -557,7 +588,9 @@ class GpuDnnConvGradW(DnnBase):
...
@@ -557,7 +588,9 @@ class GpuDnnConvGradW(DnnBase):
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1'
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1'
if
self
.
algo
==
'fft'
:
if
self
.
algo
==
'fft'
:
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT'
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT'
if
self
.
algo
==
'small'
:
# non-deterministic, small workspace
alg
=
'CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3'
if
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
,
if
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
]:
'time_once'
,
'time_on_shape_change'
]:
defs
.
append
((
'CHOOSE_ALGO'
,
''
))
defs
.
append
((
'CHOOSE_ALGO'
,
''
))
...
@@ -587,7 +620,8 @@ class GpuDnnConvGradW(DnnBase):
...
@@ -587,7 +620,8 @@ class GpuDnnConvGradW(DnnBase):
raise
TypeError
(
"The number of dimensions of "
raise
TypeError
(
"The number of dimensions of "
"img, topgrad and output must match"
)
"img, topgrad and output must match"
)
if
img
.
type
.
ndim
==
5
and
self
.
algo
in
[
'fft'
,
'deterministic'
]:
if
(
img
.
type
.
ndim
==
5
and
self
.
algo
in
[
'fft'
,
'deterministic'
,
'small'
]):
raise
ValueError
(
"convolution algo
%
s can't be used for "
raise
ValueError
(
"convolution algo
%
s can't be used for "
"3d convolutions"
,
(
self
.
algo
,))
"3d convolutions"
,
(
self
.
algo
,))
...
@@ -627,16 +661,23 @@ class GpuDnnConvGradI(DnnBase):
...
@@ -627,16 +661,23 @@ class GpuDnnConvGradI(DnnBase):
if
self
.
inplace
:
if
self
.
inplace
:
self
.
destroy_map
=
{
0
:
[
2
]}
self
.
destroy_map
=
{
0
:
[
2
]}
if
algo
is
None
:
if
algo
is
None
:
algo
=
config
.
dnn
.
conv
.
algo_bwd
algo
=
config
.
dnn
.
conv
.
algo_bwd
_data
self
.
algo
=
algo
self
.
algo
=
algo
assert
self
.
algo
in
[
'none'
,
'deterministic'
,
'fft'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
# The small-workspace implementation is only available from CuDNN V4
'time_on_shape_change'
]
# onward.
if
version
()
<
4000
and
self
.
algo
==
'fft_tiling'
:
raise
RuntimeError
(
"CuDNN's tiled-FFT convolution requires CuDNN "
"v4 or more recent"
)
assert
self
.
algo
in
[
'none'
,
'deterministic'
,
'fft'
,
'fft_tiling'
,
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
]
def
__setstate__
(
self
,
d
):
def
__setstate__
(
self
,
d
):
self
.
__dict__
.
update
(
d
)
self
.
__dict__
.
update
(
d
)
if
not
hasattr
(
self
,
'algo'
):
if
not
hasattr
(
self
,
'algo'
):
self
.
algo
=
config
.
dnn
.
conv
.
algo_bwd
self
.
algo
=
config
.
dnn
.
conv
.
algo_bwd
_data
if
not
hasattr
(
self
,
'inplace'
):
if
not
hasattr
(
self
,
'inplace'
):
self
.
inplace
=
False
self
.
inplace
=
False
...
@@ -673,6 +714,9 @@ class GpuDnnConvGradI(DnnBase):
...
@@ -673,6 +714,9 @@ class GpuDnnConvGradI(DnnBase):
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_1'
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_1'
if
self
.
algo
==
'fft'
:
if
self
.
algo
==
'fft'
:
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT'
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT'
if
self
.
algo
==
'fft_tiling'
:
# big workspace but less than fft
alg
=
'CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING'
if
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
,
if
self
.
algo
in
[
'guess_once'
,
'guess_on_shape_change'
,
'time_once'
,
'time_on_shape_change'
]:
'time_once'
,
'time_on_shape_change'
]:
...
@@ -703,7 +747,8 @@ class GpuDnnConvGradI(DnnBase):
...
@@ -703,7 +747,8 @@ class GpuDnnConvGradI(DnnBase):
raise
TypeError
(
"The number of dimensions of "
raise
TypeError
(
"The number of dimensions of "
"kern, topgrad and output must match"
)
"kern, topgrad and output must match"
)
if
kern
.
type
.
ndim
==
5
and
self
.
algo
in
[
'fft'
,
'deterministic'
]:
if
(
kern
.
type
.
ndim
==
5
and
self
.
algo
in
[
'fft'
,
'deterministic'
,
'fft_tiling'
]):
raise
ValueError
(
"convolution algo
%
s can't be used for "
raise
ValueError
(
"convolution algo
%
s can't be used for "
"3d convolutions"
,
(
self
.
algo
,))
"3d convolutions"
,
(
self
.
algo
,))
...
@@ -723,7 +768,7 @@ class GpuDnnConvGradI(DnnBase):
...
@@ -723,7 +768,7 @@ class GpuDnnConvGradI(DnnBase):
def
dnn_conv
(
img
,
kerns
,
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
def
dnn_conv
(
img
,
kerns
,
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
conv_mode
=
'conv'
,
direction_hint
=
None
,
workmem
=
None
,
conv_mode
=
'conv'
,
direction_hint
=
None
,
workmem
=
None
,
algo
=
None
):
algo
=
None
,
precision
=
None
):
"""
"""
GPU convolution using cuDNN from NVIDIA.
GPU convolution using cuDNN from NVIDIA.
...
@@ -757,12 +802,24 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -757,12 +802,24 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
Convolution implementation to use. Some of its values may
Convolution implementation to use. Some of its values may
require certain versions of CuDNN to be installed. Default is
require certain versions of CuDNN to be installed. Default is
the value of :attr:`config.dnn.conv.algo_fwd`.
the value of :attr:`config.dnn.conv.algo_fwd`.
precision : {'as_input', 'float16', 'float32', 'float64'}
Description of the dtype in which the computation of the convolution
should be done. Possible values are 'as_input', 'float16', 'float32'
and 'float64'. Default is the value of
:attr:`config.dnn.conv.precision`.
.. warning:: The cuDNN library only works with GPUs that have a compute
.. warning:: The cuDNN library only works with GPUs that have a compute
capability of 3.0 or higer. This means that older GPUs will not
capability of 3.0 or higer. This means that older GPUs will not
work with this Op.
work with this Op.
"""
"""
# Establish dtype in which to perform the computation of the convolution
if
precision
is
None
:
precision
=
theano
.
config
.
dnn
.
conv
.
precision
if
precision
==
'as_input'
:
precision
=
theano
.
scalar
.
upcast
(
img
.
dtype
,
kerns
.
dtype
)
if
workmem
is
not
None
:
if
workmem
is
not
None
:
if
algo
is
not
None
:
if
algo
is
not
None
:
raise
ValueError
(
"You can't use both algo and workmem"
)
raise
ValueError
(
"You can't use both algo and workmem"
)
...
@@ -786,7 +843,7 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -786,7 +843,7 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
shape_i
(
kerns
,
1
,
fgraph
),
shape_i
(
kerns
,
1
,
fgraph
),
shape_i
(
img
,
1
,
fgraph
),
shape2
,
shape3
)
shape_i
(
img
,
1
,
fgraph
),
shape2
,
shape3
)
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
conv_mode
=
'cross'
)(
out
.
shape
)
conv_mode
=
'cross'
,
precision
=
precision
)(
out
.
shape
)
conv
=
GpuDnnConvGradW
()(
img
,
kerns
,
out
,
desc
)
conv
=
GpuDnnConvGradW
()(
img
,
kerns
,
out
,
desc
)
return
as_gpuarray_variable
(
conv
.
dimshuffle
(
1
,
0
,
2
,
3
),
ctx_name
)
return
as_gpuarray_variable
(
conv
.
dimshuffle
(
1
,
0
,
2
,
3
),
ctx_name
)
...
@@ -804,7 +861,7 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -804,7 +861,7 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
shape_i
(
kerns
,
1
,
fgraph
),
shape_i
(
kerns
,
1
,
fgraph
),
shape2
,
shape3
)
shape2
,
shape3
)
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
desc
=
GpuDnnConvDesc
(
border_mode
=
'valid'
,
subsample
=
(
1
,
1
),
conv_mode
=
conv_mode
)(
kerns
.
shape
)
conv_mode
=
conv_mode
,
precision
=
precision
)(
kerns
.
shape
)
return
GpuDnnConvGradI
()(
kerns
,
img
,
out
,
desc
)
return
GpuDnnConvGradI
()(
kerns
,
img
,
out
,
desc
)
# Standard case: We use GpuDnnConv with suitable padding.
# Standard case: We use GpuDnnConv with suitable padding.
...
@@ -813,7 +870,7 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
...
@@ -813,7 +870,7 @@ def dnn_conv(img, kerns, border_mode='valid', subsample=(1, 1),
img
=
gpu_contiguous
(
img
)
img
=
gpu_contiguous
(
img
)
kerns
=
gpu_contiguous
(
kerns
)
kerns
=
gpu_contiguous
(
kerns
)
desc
=
GpuDnnConvDesc
(
border_mode
=
border_mode
,
subsample
=
subsample
,
desc
=
GpuDnnConvDesc
(
border_mode
=
border_mode
,
subsample
=
subsample
,
conv_mode
=
conv_mode
)(
kerns
.
shape
)
conv_mode
=
conv_mode
,
precision
=
precision
)(
kerns
.
shape
)
desc_op
=
desc
.
owner
.
op
desc_op
=
desc
.
owner
.
op
out_shp
=
GpuDnnConv
.
get_out_shape
(
img
.
shape
,
kerns
.
shape
,
out_shp
=
GpuDnnConv
.
get_out_shape
(
img
.
shape
,
kerns
.
shape
,
desc_op
.
border_mode
,
desc_op
.
border_mode
,
...
...
theano/sandbox/gpuarray/dnn_fwd.c
浏览文件 @
4ad36ddc
...
@@ -136,15 +136,26 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
...
@@ -136,15 +136,26 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
algo
==
CUDNN_CONVOLUTION_FWD_ALGO_GEMM
))
algo
==
CUDNN_CONVOLUTION_FWD_ALGO_GEMM
))
algo
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM
;
algo
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM
;
#if CUDNN_VERSION > 3000
// The FFT implementation does not support strides, 1x1 filters or inputs
if
(
algo
==
CUDNN_CONVOLUTION_FWD_ALGO_FFT
)
{
// with a spatial dimension larger than 1024. The tiled-FFT implementation
// does not support strides.
// If the chosen implementation is FFT or tiled-FFT, validate that it can
// be used on the current data and default to a safe implementation if it
// can't.
// The following code is 2d-specific but it is fine as FFT and tiled-FFT are
// defined only for 2d filters
if
((
algo
==
CUDNN_CONVOLUTION_FWD_ALGO_FFT
||
algo
==
CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING
)
&&
PyGpuArray_NDIM
(
input
)
==
4
)
{
// Extract the properties of the convolution descriptor
int
nd
;
int
nd
;
int
pad
[
2
];
int
pad
[
2
];
int
stride
[
2
];
int
stride
[
2
];
int
upscale
[
2
];
int
upscale
[
2
];
cudnnConvolutionMode_t
mode
;
cudnnConvolutionMode_t
mode
;
err
=
cudnnGetConvolutionNdDescriptor
(
desc
,
2
,
&
nd
,
pad
,
stride
,
cudnnDataType_t
data_type
;
upscale
,
&
mode
);
err
=
cudnnGetConvolutionNdDescriptor_v3
(
desc
,
2
,
&
nd
,
pad
,
stride
,
upscale
,
&
mode
,
&
data_type
);
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
PyErr_Format
(
PyExc_RuntimeError
,
PyErr_Format
(
PyExc_RuntimeError
,
"error getting convolution properties: %s"
,
"error getting convolution properties: %s"
,
...
@@ -153,30 +164,24 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
...
@@ -153,30 +164,24 @@ APPLY_SPECIFIC(conv_fwd)(PyGpuArrayObject *input, PyGpuArrayObject *kerns,
return
1
;
return
1
;
}
}
if
(
algo
==
CUDNN_CONVOLUTION_FWD_ALGO_FFT
)
{
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
||
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
||
PyGpuArray_DIM
(
input
,
2
)
>
1024
||
PyGpuArray_DIM
(
input
,
3
)
>
1024
||
PyGpuArray_DIM
(
input
,
2
)
>
1024
||
PyGpuArray_DIM
(
input
,
3
)
>
1024
||
(
PyGpuArray_DIM
(
kerns
,
2
)
==
1
&&
PyGpuArray_DIM
(
kerns
,
3
)
==
1
))
{
(
PyGpuArray_DIM
(
kerns
,
2
)
==
1
&&
PyGpuArray_DIM
(
kerns
,
3
)
==
1
))
algo
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM
;
{
algo
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM
;
}
}
}
}
#endif
else
{
#if CUDNN_VERSION < 3000
// algo == CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING
/* cuDNN before v3 does not support kernels larger than input even
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
)
* if appropriate padding is selected. */
{
for
(
unsigned
int
i
=
2
;
i
<
PyGpuArray_NDIM
(
input
);
i
++
)
{
algo
=
CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM
;
if
(
PyGpuArray_DIM
(
kerns
,
i
)
>
PyGpuArray_DIM
(
input
,
i
))
{
}
PyErr_SetString
(
PyExc_RuntimeError
,
"the current version "
"of CuDNN does not support kernels larger than the "
"inputs in any spatial dimension, even if the inputs "
"are padded such that the padded inputs are larger "
"than the kernels. Update your installation of CuDNN "
"to V3 or more recent to solve the issue."
);
cuda_exit
(
c
->
ctx
);
return
1
;
}
}
}
}
#endif
{
{
size_t
worksize
;
size_t
worksize
;
...
...
theano/sandbox/gpuarray/dnn_gi.c
浏览文件 @
4ad36ddc
...
@@ -128,15 +128,26 @@ APPLY_SPECIFIC(conv_gi)(PyGpuArrayObject *kerns, PyGpuArrayObject *output,
...
@@ -128,15 +128,26 @@ APPLY_SPECIFIC(conv_gi)(PyGpuArrayObject *kerns, PyGpuArrayObject *output,
#endif
#endif
#if CUDNN_VERSION > 3000
// The FFT implementation does not support strides, 1x1 filters or inputs
if
(
algo
==
CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT
)
{
// with a spatial dimension larger than 1024. The tiled-FFT implementation
// does not support strides.
// If the chosen implementation is FFT or tiled-FFT, validate that it can
// be used on the current data and default to a safe implementation if it
// can't.
// The following code is 2d-specific but it is fine as FFT and tiled-FFT are
// defined only for 2d filters
if
((
algo
==
CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING
||
algo
==
CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT
)
&&
PyGpuArray_NDIM
(
kerns
)
==
4
)
{
// Extract the properties of the convolution descriptor
int
nd
;
int
nd
;
int
pad
[
2
];
int
pad
[
2
];
int
stride
[
2
];
int
stride
[
2
];
int
upscale
[
2
];
int
upscale
[
2
];
cudnnConvolutionMode_t
mode
;
cudnnConvolutionMode_t
mode
;
err
=
cudnnGetConvolutionNdDescriptor
(
desc
,
2
,
&
nd
,
pad
,
stride
,
cudnnDataType_t
data_type
;
upscale
,
&
mode
);
err
=
cudnnGetConvolutionNdDescriptor_v3
(
desc
,
2
,
&
nd
,
pad
,
stride
,
upscale
,
&
mode
,
&
data_type
);
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
PyErr_Format
(
PyExc_RuntimeError
,
PyErr_Format
(
PyExc_RuntimeError
,
"error getting convolution properties: %s"
,
"error getting convolution properties: %s"
,
...
@@ -145,13 +156,24 @@ APPLY_SPECIFIC(conv_gi)(PyGpuArrayObject *kerns, PyGpuArrayObject *output,
...
@@ -145,13 +156,24 @@ APPLY_SPECIFIC(conv_gi)(PyGpuArrayObject *kerns, PyGpuArrayObject *output,
return
1
;
return
1
;
}
}
if
(
algo
==
CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT
)
{
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
||
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
||
PyGpuArray_DIM
(
*
input
,
2
)
>
1024
||
PyGpuArray_DIM
(
*
input
,
3
)
>
1024
||
PyGpuArray_DIM
(
*
input
,
2
)
>
1024
||
PyGpuArray_DIM
(
*
input
,
3
)
>
1024
||
(
PyGpuArray_DIM
(
kerns
,
2
)
==
1
&&
PyGpuArray_DIM
(
kerns
,
3
)
==
1
))
{
(
PyGpuArray_DIM
(
kerns
,
2
)
==
1
&&
PyGpuArray_DIM
(
kerns
,
3
)
==
1
))
{
algo
=
CUDNN_CONVOLUTION_BWD_DATA_ALGO_0
;
algo
=
CUDNN_CONVOLUTION_BWD_DATA_ALGO_0
;
}
}
}
}
#endif
else
{
// algo == CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING
if
(
stride
[
0
]
!=
1
||
stride
[
1
]
!=
1
)
{
algo
=
CUDNN_CONVOLUTION_BWD_DATA_ALGO_0
;
}
}
}
size_t
worksize
;
size_t
worksize
;
gpudata
*
workspace
;
gpudata
*
workspace
;
...
...
theano/sandbox/gpuarray/dnn_gw.c
浏览文件 @
4ad36ddc
...
@@ -130,15 +130,24 @@ APPLY_SPECIFIC(conv_gw)(PyGpuArrayObject *input, PyGpuArrayObject *output,
...
@@ -130,15 +130,24 @@ APPLY_SPECIFIC(conv_gw)(PyGpuArrayObject *input, PyGpuArrayObject *output,
#endif
#endif
#if CUDNN_VERSION > 3000
// The FFT implementation does not support strides, 1x1 filters or inputs
if
(
algo
==
CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT
)
{
// with a spatial dimension larger than 1024.
// If the chosen implementation is FFT, validate that it can
// be used on the current data and default to a safe implementation if it
// can't.
// The following code is 2d-specific but it is fine as FFT and tiled-FFT are
// defined only for 2d filters
if
(
algo
==
CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT
&&
PyGpuArray_NDIM
(
input
)
==
4
)
{
// Extract the properties of the convolution descriptor
int
nd
;
int
nd
;
int
pad
[
2
];
int
pad
[
2
];
int
stride
[
2
];
int
stride
[
2
];
int
upscale
[
2
];
int
upscale
[
2
];
cudnnConvolutionMode_t
mode
;
cudnnConvolutionMode_t
mode
;
err
=
cudnnGetConvolutionNdDescriptor
(
desc
,
2
,
&
nd
,
pad
,
stride
,
cudnnDataType_t
data_type
;
upscale
,
&
mode
);
err
=
cudnnGetConvolutionNdDescriptor_v3
(
desc
,
2
,
&
nd
,
pad
,
stride
,
upscale
,
&
mode
,
&
data_type
);
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
if
(
err
!=
CUDNN_STATUS_SUCCESS
)
{
PyErr_Format
(
PyExc_RuntimeError
,
PyErr_Format
(
PyExc_RuntimeError
,
"error getting convolution properties: %s"
,
"error getting convolution properties: %s"
,
...
@@ -153,7 +162,6 @@ APPLY_SPECIFIC(conv_gw)(PyGpuArrayObject *input, PyGpuArrayObject *output,
...
@@ -153,7 +162,6 @@ APPLY_SPECIFIC(conv_gw)(PyGpuArrayObject *input, PyGpuArrayObject *output,
algo
=
CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0
;
algo
=
CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0
;
}
}
}
}
#endif
size_t
worksize
;
size_t
worksize
;
gpudata
*
workspace
;
gpudata
*
workspace
;
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论