Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
49ffd9b8
提交
49ffd9b8
authored
6月 26, 2015
作者:
Arnaud Bergeron
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Adapt the tests for gpuarray.
上级
ccd25be8
显示空白字符变更
内嵌
并排
正在显示
1 个修改的文件
包含
36 行增加
和
51 行删除
+36
-51
test_dnn.py
theano/sandbox/gpuarray/tests/test_dnn.py
+36
-51
没有找到文件。
theano/sandbox/gpuarray/tests/test_dnn.py
浏览文件 @
49ffd9b8
...
@@ -11,26 +11,16 @@ import theano.tests.unittest_tools as utt
...
@@ -11,26 +11,16 @@ import theano.tests.unittest_tools as utt
from
theano.sandbox.neighbours
import
images2neibs
from
theano.sandbox.neighbours
import
images2neibs
from
theano.tensor.signal.downsample
import
max_pool_2d
from
theano.tensor.signal.downsample
import
max_pool_2d
from
theano.tensor.signal.downsample
import
DownsampleFactorMaxGrad
from
theano.tensor.signal.downsample
import
DownsampleFactorMaxGrad
import
theano.sandbox.cuda.dnn
as
dnn
from
theano.sandbox.cuda.basic_ops
import
GpuAllocEmpty
,
gpu_alloc_empty
# Skip test if cuda_ndarray is not available.
from
..
import
dnn
import
theano.sandbox.cuda
as
cuda
from
..basic_ops
import
GpuAllocEmpty
if
not
cuda
.
cuda_available
:
raise
SkipTest
(
'Optional package cuda disabled'
)
if
theano
.
config
.
mode
==
'FAST_COMPILE'
:
from
.test_basic_ops
import
mode_with_gpu
,
mode_without_gpu
mode_with_gpu
=
theano
.
compile
.
mode
.
get_mode
(
'FAST_RUN'
)
.
including
(
'gpu'
)
mode_without_gpu
=
theano
.
compile
.
mode
.
get_mode
(
'FAST_RUN'
)
.
excluding
(
'gpu'
)
else
:
mode_with_gpu
=
theano
.
compile
.
mode
.
get_default_mode
()
.
including
(
'gpu'
)
mode_without_gpu
=
theano
.
compile
.
mode
.
get_default_mode
()
.
excluding
(
'gpu'
)
def
test_dnn_conv_desc_merge
():
def
test_dnn_conv_desc_merge
():
if
not
cuda
.
dnn
.
dnn_available
():
if
not
dnn
.
dnn_available
():
raise
SkipTest
(
cuda
.
dnn
.
dnn_available
.
msg
)
raise
SkipTest
(
dnn
.
dnn_available
.
msg
)
img_shp
=
T
.
as_tensor_variable
(
img_shp
=
T
.
as_tensor_variable
(
numpy
.
asarray
([
2
,
1
,
8
,
8
])
.
astype
(
'int64'
))
numpy
.
asarray
([
2
,
1
,
8
,
8
])
.
astype
(
'int64'
))
kern_shp
=
T
.
as_tensor_variable
(
kern_shp
=
T
.
as_tensor_variable
(
...
@@ -51,14 +41,9 @@ def test_dnn_conv_desc_merge():
...
@@ -51,14 +41,9 @@ def test_dnn_conv_desc_merge():
def
test_dnn_conv_merge
():
def
test_dnn_conv_merge
():
"""This test that we merge correctly multiple dnn_conv.
# This test that we merge correctly multiple dnn_conv.
if
not
dnn
.
dnn_available
():
This can is more difficult due to GpuEmptyAlloc that aren't
raise
SkipTest
(
dnn
.
dnn_available
.
msg
)
merged.
"""
if
not
cuda
.
dnn
.
dnn_available
():
raise
SkipTest
(
cuda
.
dnn
.
dnn_available
.
msg
)
img_shp
=
[
2
,
5
,
6
,
8
]
img_shp
=
[
2
,
5
,
6
,
8
]
kern_shp
=
[
3
,
5
,
5
,
6
]
kern_shp
=
[
3
,
5
,
5
,
6
]
img
=
T
.
ftensor4
(
'img'
)
img
=
T
.
ftensor4
(
'img'
)
...
@@ -96,8 +81,8 @@ def test_dnn_conv_inplace():
...
@@ -96,8 +81,8 @@ def test_dnn_conv_inplace():
GpuAllocEmpty get merged together.
GpuAllocEmpty get merged together.
"""
"""
if
not
cuda
.
dnn
.
dnn_available
():
if
not
dnn
.
dnn_available
():
raise
SkipTest
(
cuda
.
dnn
.
dnn_available
.
msg
)
raise
SkipTest
(
dnn
.
dnn_available
.
msg
)
img_shp
=
[
2
,
5
,
6
,
8
]
img_shp
=
[
2
,
5
,
6
,
8
]
kern_shp
=
[
3
,
5
,
5
,
6
]
kern_shp
=
[
3
,
5
,
5
,
6
]
img
=
T
.
ftensor4
(
'img'
)
img
=
T
.
ftensor4
(
'img'
)
...
@@ -121,7 +106,7 @@ def test_dnn_conv_inplace():
...
@@ -121,7 +106,7 @@ def test_dnn_conv_inplace():
assert
len
([
n
for
n
in
topo
if
isinstance
(
n
.
op
,
GpuAllocEmpty
)])
==
2
assert
len
([
n
for
n
in
topo
if
isinstance
(
n
.
op
,
GpuAllocEmpty
)])
==
2
# Test grad w op
# Test grad w op
out
=
gpu_alloc_empty
(
*
kern
.
shape
)
out
=
GpuAllocEmpty
(
kern
.
dtype
)
(
*
kern
.
shape
)
o1
=
dnn
.
GpuDnnConvGradW
()(
img
,
kern
,
out
,
desc1
)
o1
=
dnn
.
GpuDnnConvGradW
()(
img
,
kern
,
out
,
desc1
)
o2
=
dnn
.
GpuDnnConvGradW
()(
img
,
kern
,
out
,
desc2
)
o2
=
dnn
.
GpuDnnConvGradW
()(
img
,
kern
,
out
,
desc2
)
f
=
theano
.
function
([
img
,
kern
],
[
o1
,
o2
],
mode
=
mode_with_gpu
)
f
=
theano
.
function
([
img
,
kern
],
[
o1
,
o2
],
mode
=
mode_with_gpu
)
...
@@ -132,7 +117,7 @@ def test_dnn_conv_inplace():
...
@@ -132,7 +117,7 @@ def test_dnn_conv_inplace():
assert
len
([
n
for
n
in
topo
if
isinstance
(
n
.
op
,
GpuAllocEmpty
)])
==
2
assert
len
([
n
for
n
in
topo
if
isinstance
(
n
.
op
,
GpuAllocEmpty
)])
==
2
# Test grad i op
# Test grad i op
out
=
gpu_alloc_empty
(
*
img
.
shape
)
out
=
GpuAllocEmpty
(
img
.
dtype
)
(
*
img
.
shape
)
o1
=
dnn
.
GpuDnnConvGradI
()(
img
,
kern
,
out
,
desc1
)
o1
=
dnn
.
GpuDnnConvGradI
()(
img
,
kern
,
out
,
desc1
)
o2
=
dnn
.
GpuDnnConvGradI
()(
img
,
kern
,
out
,
desc2
)
o2
=
dnn
.
GpuDnnConvGradI
()(
img
,
kern
,
out
,
desc2
)
f
=
theano
.
function
([
img
,
kern
],
[
o1
,
o2
],
mode
=
mode_with_gpu
)
f
=
theano
.
function
([
img
,
kern
],
[
o1
,
o2
],
mode
=
mode_with_gpu
)
...
@@ -179,8 +164,8 @@ def pool_2d_i2n(input, ds=(2, 2), strides=None,
...
@@ -179,8 +164,8 @@ def pool_2d_i2n(input, ds=(2, 2), strides=None,
def
test_pooling
():
def
test_pooling
():
if
not
cuda
.
dnn
.
dnn_available
():
if
not
dnn
.
dnn_available
():
raise
SkipTest
(
cuda
.
dnn
.
dnn_available
.
msg
)
raise
SkipTest
(
dnn
.
dnn_available
.
msg
)
x
=
T
.
ftensor4
()
x
=
T
.
ftensor4
()
for
mode
,
pad
in
product
((
'max'
,
'average_inc_pad'
,
'average_exc_pad'
),
for
mode
,
pad
in
product
((
'max'
,
'average_inc_pad'
,
'average_exc_pad'
),
...
@@ -189,7 +174,7 @@ def test_pooling():
...
@@ -189,7 +174,7 @@ def test_pooling():
func
=
T
.
max
func
=
T
.
max
else
:
else
:
func
=
T
.
mean
func
=
T
.
mean
if
pad
!=
(
0
,
0
)
and
cuda
.
dnn
.
version
()
==
-
1
:
if
pad
!=
(
0
,
0
)
and
dnn
.
version
()
==
-
1
:
continue
continue
if
pad
!=
(
0
,
0
)
and
func
is
T
.
mean
:
if
pad
!=
(
0
,
0
)
and
func
is
T
.
mean
:
...
@@ -213,10 +198,10 @@ def test_pooling():
...
@@ -213,10 +198,10 @@ def test_pooling():
mode_without_gpu2
=
mode_without_gpu
.
including
()
mode_without_gpu2
=
mode_without_gpu
.
including
()
mode_without_gpu2
.
check_isfinite
=
False
mode_without_gpu2
.
check_isfinite
=
False
f1
=
theano
.
function
([
x
],
out1
,
mode
=
mode_with_gpu
)
f1
=
theano
.
function
([
x
],
out1
,
mode
=
mode_with_gpu
)
assert
any
([
isinstance
(
node
.
op
,
cuda
.
dnn
.
GpuDnnPool
)
assert
any
([
isinstance
(
node
.
op
,
dnn
.
GpuDnnPool
)
for
node
in
f1
.
maker
.
fgraph
.
apply_nodes
])
for
node
in
f1
.
maker
.
fgraph
.
apply_nodes
])
f2
=
theano
.
function
([
x
],
out2
,
mode
=
mode_without_gpu2
)
f2
=
theano
.
function
([
x
],
out2
,
mode
=
mode_without_gpu2
)
assert
not
any
([
isinstance
(
node
.
op
,
cuda
.
dnn
.
GpuDnnPool
)
assert
not
any
([
isinstance
(
node
.
op
,
dnn
.
GpuDnnPool
)
for
node
in
f2
.
maker
.
fgraph
.
apply_nodes
])
for
node
in
f2
.
maker
.
fgraph
.
apply_nodes
])
for
shp
in
[(
1
,
10
,
100
,
100
),
for
shp
in
[(
1
,
10
,
100
,
100
),
(
1
,
3
,
99
,
99
),
(
1
,
3
,
99
,
99
),
...
@@ -250,12 +235,12 @@ def test_pooling():
...
@@ -250,12 +235,12 @@ def test_pooling():
# Confirm that the opt would have inserted it.
# Confirm that the opt would have inserted it.
fg
=
theano
.
function
([
x
],
theano
.
grad
(
fn
(
x
)
.
sum
(),
x
),
fg
=
theano
.
function
([
x
],
theano
.
grad
(
fn
(
x
)
.
sum
(),
x
),
mode
=
mode_with_gpu
)
mode
=
mode_with_gpu
)
assert
any
([
isinstance
(
node
.
op
,
cuda
.
dnn
.
GpuDnnPoolGrad
)
assert
any
([
isinstance
(
node
.
op
,
dnn
.
GpuDnnPoolGrad
)
for
node
in
fg
.
maker
.
fgraph
.
toposort
()])
for
node
in
fg
.
maker
.
fgraph
.
toposort
()])
# Test the GPU grad + GPU implementation
# Test the GPU grad + GPU implementation
def
fn
(
x
):
def
fn
(
x
):
dnn_op
=
cuda
.
dnn
.
dnn_pool
(
dnn_op
=
dnn
.
dnn_pool
(
x
,
ws
=
(
ws
,
ws
),
x
,
ws
=
(
ws
,
ws
),
stride
=
(
stride
,
stride
),
stride
=
(
stride
,
stride
),
pad
=
pad
,
pad
=
pad
,
...
@@ -268,7 +253,7 @@ def test_pooling():
...
@@ -268,7 +253,7 @@ def test_pooling():
# Confirm that we get the good op.
# Confirm that we get the good op.
fg
=
theano
.
function
([
x
],
theano
.
grad
(
fn
(
x
)
.
sum
(),
x
),
fg
=
theano
.
function
([
x
],
theano
.
grad
(
fn
(
x
)
.
sum
(),
x
),
mode
=
mode_with_gpu
)
mode
=
mode_with_gpu
)
assert
any
([
isinstance
(
node
.
op
,
cuda
.
dnn
.
GpuDnnPoolGrad
)
assert
any
([
isinstance
(
node
.
op
,
dnn
.
GpuDnnPoolGrad
)
for
node
in
fg
.
maker
.
fgraph
.
toposort
()])
for
node
in
fg
.
maker
.
fgraph
.
toposort
()])
g_out
=
fg
(
data
)
g_out
=
fg
(
data
)
...
@@ -285,8 +270,8 @@ def test_pooling():
...
@@ -285,8 +270,8 @@ def test_pooling():
def
test_pooling_opt
():
def
test_pooling_opt
():
if
not
cuda
.
dnn
.
dnn_available
():
if
not
dnn
.
dnn_available
():
raise
SkipTest
(
cuda
.
dnn
.
dnn_available
.
msg
)
raise
SkipTest
(
dnn
.
dnn_available
.
msg
)
x
=
T
.
ftensor4
()
x
=
T
.
ftensor4
()
...
@@ -295,7 +280,7 @@ def test_pooling_opt():
...
@@ -295,7 +280,7 @@ def test_pooling_opt():
max_pool_2d
(
x
,
ds
=
(
2
,
2
),
ignore_border
=
True
),
max_pool_2d
(
x
,
ds
=
(
2
,
2
),
ignore_border
=
True
),
mode
=
mode_with_gpu
)
mode
=
mode_with_gpu
)
assert
any
([
isinstance
(
n
.
op
,
cuda
.
dnn
.
GpuDnnPool
)
assert
any
([
isinstance
(
n
.
op
,
dnn
.
GpuDnnPool
)
for
n
in
f
.
maker
.
fgraph
.
toposort
()])
for
n
in
f
.
maker
.
fgraph
.
toposort
()])
f
=
theano
.
function
(
f
=
theano
.
function
(
...
@@ -303,7 +288,7 @@ def test_pooling_opt():
...
@@ -303,7 +288,7 @@ def test_pooling_opt():
T
.
grad
(
max_pool_2d
(
x
,
ds
=
(
2
,
2
),
ignore_border
=
True
)
.
sum
(),
x
),
T
.
grad
(
max_pool_2d
(
x
,
ds
=
(
2
,
2
),
ignore_border
=
True
)
.
sum
(),
x
),
mode
=
mode_with_gpu
.
including
(
"cudnn"
))
mode
=
mode_with_gpu
.
including
(
"cudnn"
))
assert
any
([
isinstance
(
n
.
op
,
cuda
.
dnn
.
GpuDnnPoolGrad
)
assert
any
([
isinstance
(
n
.
op
,
dnn
.
GpuDnnPoolGrad
)
for
n
in
f
.
maker
.
fgraph
.
toposort
()])
for
n
in
f
.
maker
.
fgraph
.
toposort
()])
...
@@ -327,7 +312,7 @@ def test_dnn_tag():
...
@@ -327,7 +312,7 @@ def test_dnn_tag():
max_pool_2d
(
x
,
ds
=
(
2
,
2
),
ignore_border
=
True
),
max_pool_2d
(
x
,
ds
=
(
2
,
2
),
ignore_border
=
True
),
mode
=
mode_with_gpu
.
including
(
"cudnn"
))
mode
=
mode_with_gpu
.
including
(
"cudnn"
))
except
(
AssertionError
,
RuntimeError
):
except
(
AssertionError
,
RuntimeError
):
assert
not
cuda
.
dnn
.
dnn_available
()
assert
not
dnn
.
dnn_available
()
raised
=
True
raised
=
True
finally
:
finally
:
theano
.
config
.
on_opt_error
=
old
theano
.
config
.
on_opt_error
=
old
...
@@ -336,8 +321,8 @@ def test_dnn_tag():
...
@@ -336,8 +321,8 @@ def test_dnn_tag():
logging
.
getLogger
(
'theano'
)
.
addHandler
(
theano
.
logging_default_handler
)
logging
.
getLogger
(
'theano'
)
.
addHandler
(
theano
.
logging_default_handler
)
if
not
raised
:
if
not
raised
:
assert
cuda
.
dnn
.
dnn_available
()
assert
dnn
.
dnn_available
()
assert
any
([
isinstance
(
n
.
op
,
cuda
.
dnn
.
GpuDnnPool
)
assert
any
([
isinstance
(
n
.
op
,
dnn
.
GpuDnnPool
)
for
n
in
f
.
maker
.
fgraph
.
toposort
()])
for
n
in
f
.
maker
.
fgraph
.
toposort
()])
...
@@ -579,8 +564,8 @@ class TestDnnInferShapes(utt.InferShapeTester):
...
@@ -579,8 +564,8 @@ class TestDnnInferShapes(utt.InferShapeTester):
# this has been a problem in the past
# this has been a problem in the past
def
test_dnn_conv_border_mode
():
def
test_dnn_conv_border_mode
():
if
not
cuda
.
dnn
.
dnn_available
():
if
not
dnn
.
dnn_available
():
raise
SkipTest
(
cuda
.
dnn
.
dnn_available
.
msg
)
raise
SkipTest
(
dnn
.
dnn_available
.
msg
)
img
=
T
.
ftensor4
()
img
=
T
.
ftensor4
()
kern
=
T
.
ftensor4
()
kern
=
T
.
ftensor4
()
...
@@ -591,8 +576,8 @@ def test_dnn_conv_border_mode():
...
@@ -591,8 +576,8 @@ def test_dnn_conv_border_mode():
def
test_dnn_conv_alpha_output_merge
():
def
test_dnn_conv_alpha_output_merge
():
if
not
cuda
.
dnn
.
dnn_available
():
if
not
dnn
.
dnn_available
():
raise
SkipTest
(
cuda
.
dnn
.
dnn_available
.
msg
)
raise
SkipTest
(
dnn
.
dnn_available
.
msg
)
img
=
T
.
ftensor4
()
img
=
T
.
ftensor4
()
kern
=
T
.
ftensor4
()
kern
=
T
.
ftensor4
()
out
=
T
.
ftensor4
()
out
=
T
.
ftensor4
()
...
@@ -615,7 +600,7 @@ def test_dnn_conv_alpha_output_merge():
...
@@ -615,7 +600,7 @@ def test_dnn_conv_alpha_output_merge():
lr
=
numpy
.
asarray
(
0.05
,
dtype
=
'float32'
)
lr
=
numpy
.
asarray
(
0.05
,
dtype
=
'float32'
)
if
cuda
.
dnn
.
version
()
==
-
1
:
if
dnn
.
version
()
==
-
1
:
# Can't merge alpha with cudnn v1
# Can't merge alpha with cudnn v1
fr
=
conv
+
out
fr
=
conv
+
out
wr
=
kern
+
gw
wr
=
kern
+
gw
...
@@ -660,7 +645,7 @@ def test_dnn_conv_alpha_output_merge():
...
@@ -660,7 +645,7 @@ def test_dnn_conv_alpha_output_merge():
def
test_dnn_conv_grad
():
def
test_dnn_conv_grad
():
if
not
cuda
.
dnn
.
dnn_available
()
or
dnn
.
version
()
==
-
1
:
if
not
dnn
.
dnn_available
()
or
dnn
.
version
()
==
-
1
:
raise
SkipTest
(
'alpha != 1.0 not supported in cudnn v1'
)
raise
SkipTest
(
'alpha != 1.0 not supported in cudnn v1'
)
b
=
1
b
=
1
...
@@ -698,6 +683,6 @@ def test_dnn_conv_grad():
...
@@ -698,6 +683,6 @@ def test_dnn_conv_grad():
def
test_version
():
def
test_version
():
if
not
cuda
.
dnn
.
dnn_available
():
if
not
dnn
.
dnn_available
():
raise
SkipTest
(
cuda
.
dnn
.
dnn_available
.
msg
)
raise
SkipTest
(
dnn
.
dnn_available
.
msg
)
assert
isinstance
(
cuda
.
dnn
.
version
(),
(
int
,
tuple
))
assert
isinstance
(
dnn
.
version
(),
(
int
,
tuple
))
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论