Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
a2f6c1c9
提交
a2f6c1c9
authored
7月 13, 2017
作者:
erakra
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fixing flake8 for test_blas.py
上级
c5cd87fa
隐藏空白字符变更
内嵌
并排
正在显示
1 个修改的文件
包含
221 行增加
和
229 行删除
+221
-229
test_blas.py
theano/tensor/tests/test_blas.py
+221
-229
没有找到文件。
theano/tensor/tests/test_blas.py
浏览文件 @
a2f6c1c9
...
@@ -5,7 +5,7 @@ from unittest import TestCase
...
@@ -5,7 +5,7 @@ from unittest import TestCase
import
numpy
as
np
import
numpy
as
np
from
numpy
import
(
arange
,
array
,
common_type
,
complex64
,
complex128
,
float32
,
from
numpy
import
(
arange
,
array
,
common_type
,
complex64
,
complex128
,
float32
,
float64
,
newaxis
,
shape
,
transpose
,
zeros
)
float64
,
newaxis
,
shape
,
transpose
,
zeros
)
from
numpy.testing
import
assert_array_almost_equal
from
numpy.testing
import
assert_array_almost_equal
from
six.moves
import
xrange
from
six.moves
import
xrange
...
@@ -22,7 +22,7 @@ from theano.tensor.blas import (_dot22, _dot22scalar, res_is_a, _as_scalar,
...
@@ -22,7 +22,7 @@ from theano.tensor.blas import (_dot22, _dot22scalar, res_is_a, _as_scalar,
InconsistencyError
,
Ger
,
ger
,
ger_destructive
)
InconsistencyError
,
Ger
,
ger
,
ger_destructive
)
from
theano.tests
import
unittest_tools
from
theano.tests
import
unittest_tools
from
.test_basic
import
(
as_tensor_variable
,
inplace_func
,
from
.test_basic
import
(
as_tensor_variable
,
inplace_func
,
compile
,
inplace
)
compile
,
inplace
)
import
theano.tensor.blas_scipy
import
theano.tensor.blas_scipy
from
theano.tests.unittest_tools
import
attr
from
theano.tests.unittest_tools
import
attr
...
@@ -81,7 +81,7 @@ class t_gemm(TestCase):
...
@@ -81,7 +81,7 @@ class t_gemm(TestCase):
f
=
inplace_func
([
tz
,
ta
,
tx
,
ty
,
tb
],
f
=
inplace_func
([
tz
,
ta
,
tx
,
ty
,
tb
],
gemm_inplace
(
tz
,
ta
,
tx
,
ty
,
tb
),
gemm_inplace
(
tz
,
ta
,
tx
,
ty
,
tb
),
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
new_z
=
f
(
z
,
a
,
x
,
y
,
b
)
f
(
z
,
a
,
x
,
y
,
b
)
z_after
=
self
.
_gemm
(
z_orig
,
a
,
x
,
y
,
b
)
z_after
=
self
.
_gemm
(
z_orig
,
a
,
x
,
y
,
b
)
# print z_orig, z_after, z, type(z_orig), type(z_after), type(z)
# print z_orig, z_after, z, type(z_orig), type(z_after), type(z)
...
@@ -96,8 +96,8 @@ class t_gemm(TestCase):
...
@@ -96,8 +96,8 @@ class t_gemm(TestCase):
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'c|py'
)
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'c|py'
)
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'py'
)
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'py'
)
if
(
not
dtype
.
startswith
(
"complex"
)
if
(
not
dtype
.
startswith
(
"complex"
)
and
and
theano
.
config
.
cxx
):
theano
.
config
.
cxx
):
# If theano.config.blas.ldflags is empty, Theano will use
# If theano.config.blas.ldflags is empty, Theano will use
# a NumPy C implementation of [sd]gemm_.
# a NumPy C implementation of [sd]gemm_.
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'c'
)
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'c'
)
...
@@ -105,7 +105,7 @@ class t_gemm(TestCase):
...
@@ -105,7 +105,7 @@ class t_gemm(TestCase):
def
test0a
(
self
):
def
test0a
(
self
):
Gemm
.
debug
=
True
Gemm
.
debug
=
True
try
:
try
:
g
=
g
emm_no_inplace
([
1.
],
1.
,
[
1.
],
[
1.
],
1.
)
gemm_no_inplace
([
1.
],
1.
,
[
1.
],
[
1.
],
1.
)
except
TypeError
as
e
:
except
TypeError
as
e
:
if
exc_message
(
e
)
is
Gemm
.
E_rank
:
if
exc_message
(
e
)
is
Gemm
.
E_rank
:
return
return
...
@@ -171,7 +171,6 @@ class t_gemm(TestCase):
...
@@ -171,7 +171,6 @@ class t_gemm(TestCase):
def
test_factorised_scalar
(
self
):
def
test_factorised_scalar
(
self
):
a
=
T
.
matrix
()
a
=
T
.
matrix
()
b
=
T
.
matrix
()
b
=
T
.
matrix
()
c
=
T
.
matrix
()
s
=
theano
.
shared
(
np
.
zeros
((
5
,
5
))
.
astype
(
config
.
floatX
))
s
=
theano
.
shared
(
np
.
zeros
((
5
,
5
))
.
astype
(
config
.
floatX
))
lr1
=
T
.
constant
(
0.01
)
.
astype
(
config
.
floatX
)
lr1
=
T
.
constant
(
0.01
)
.
astype
(
config
.
floatX
)
...
@@ -180,9 +179,9 @@ class t_gemm(TestCase):
...
@@ -180,9 +179,9 @@ class t_gemm(TestCase):
# test constant merge with gemm
# test constant merge with gemm
f
=
theano
.
function
([
a
,
b
],
updates
=
[(
s
,
lr1
*
T
.
dot
(
a
,
b
)
+
f
=
theano
.
function
([
a
,
b
],
updates
=
[(
s
,
lr1
*
T
.
dot
(
a
,
b
)
+
l2_reg
*
lr2
*
s
)],
l2_reg
*
lr2
*
s
)],
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
#[Gemm{inplace}(<TensorType(float64, matrix)>, 0.01,
#
[Gemm{inplace}(<TensorType(float64, matrix)>, 0.01,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# 2e-06)]
# 2e-06)]
assert
len
(
f
)
==
1
assert
len
(
f
)
==
1
...
@@ -192,7 +191,7 @@ class t_gemm(TestCase):
...
@@ -192,7 +191,7 @@ class t_gemm(TestCase):
f
=
theano
.
function
([
a
,
b
],
updates
=
[(
s
,
lr1
*
(
T
.
dot
(
a
,
b
)
-
f
=
theano
.
function
([
a
,
b
],
updates
=
[(
s
,
lr1
*
(
T
.
dot
(
a
,
b
)
-
l2_reg
*
s
))],
l2_reg
*
s
))],
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
#[Gemm{inplace}(<TensorType(float64, matrix)>, 0.01,
#
[Gemm{inplace}(<TensorType(float64, matrix)>, 0.01,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# -2e-06)]
# -2e-06)]
assert
len
(
f
)
==
1
assert
len
(
f
)
==
1
...
@@ -202,7 +201,7 @@ class t_gemm(TestCase):
...
@@ -202,7 +201,7 @@ class t_gemm(TestCase):
f
=
theano
.
function
([
a
,
b
],
f
=
theano
.
function
([
a
,
b
],
updates
=
[(
s
,
s
-
lr1
*
(
s
*
.
0002
+
T
.
dot
(
a
,
b
)))],
updates
=
[(
s
,
s
-
lr1
*
(
s
*
.
0002
+
T
.
dot
(
a
,
b
)))],
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
#[Gemm{inplace}(<TensorType(float64, matrix)>, -0.01,
#
[Gemm{inplace}(<TensorType(float64, matrix)>, -0.01,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# 0.999998)]
# 0.999998)]
assert
len
(
f
)
==
1
assert
len
(
f
)
==
1
...
@@ -270,14 +269,14 @@ class t_gemm(TestCase):
...
@@ -270,14 +269,14 @@ class t_gemm(TestCase):
def
t
(
z
,
x
,
y
,
a
=
1.0
,
b
=
0.0
,
l
=
'c|py'
,
dt
=
'float64'
):
def
t
(
z
,
x
,
y
,
a
=
1.0
,
b
=
0.0
,
l
=
'c|py'
,
dt
=
'float64'
):
z
,
a
,
x
,
y
,
b
=
[
theano
.
_asarray
(
p
,
dtype
=
dt
)
z
,
a
,
x
,
y
,
b
=
[
theano
.
_asarray
(
p
,
dtype
=
dt
)
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
z_orig
=
z
.
copy
()
#
z_orig = z.copy()
z_after
=
self
.
_gemm
(
z
,
a
,
x
,
y
,
b
)
z_after
=
self
.
_gemm
(
z
,
a
,
x
,
y
,
b
)
tz
,
ta
,
tx
,
ty
,
tb
=
[
shared
(
p
)
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
tz
,
ta
,
tx
,
ty
,
tb
=
[
shared
(
p
)
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
# f = inplace_func([tz,ta,tx,ty,tb], gemm_inplace(tz,ta,tx,ty,tb),
# f = inplace_func([tz,ta,tx,ty,tb], gemm_inplace(tz,ta,tx,ty,tb),
# mode = compile.Mode(optimizer = None, linker=l))
# mode = compile.Mode(optimizer = None, linker=l))
#f(z, a, x, y, b)
#
f(z, a, x, y, b)
f
=
inplace_func
([],
gemm_inplace
(
tz
,
ta
,
tx
,
ty
,
tb
),
f
=
inplace_func
([],
gemm_inplace
(
tz
,
ta
,
tx
,
ty
,
tb
),
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
f
()
f
()
...
@@ -339,9 +338,9 @@ class t_gemm(TestCase):
...
@@ -339,9 +338,9 @@ class t_gemm(TestCase):
tz
,
ta
,
tx
,
ty
,
tb
=
[
shared
(
p
)
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
tz
,
ta
,
tx
,
ty
,
tb
=
[
shared
(
p
)
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
for
i
in
xrange
(
3
):
for
i
in
xrange
(
3
):
f_i
=
inplace_func
([],
f_i
=
inplace_func
([],
gemm_inplace
(
tz
[:,
:,
i
],
gemm_inplace
(
tz
[:,
:,
i
],
ta
,
tx
[:,
:,
i
],
ty
[:,
:,
i
],
tb
),
ta
,
tx
[:,
:,
i
],
ty
[:,
:,
i
],
tb
),
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
for
j
in
xrange
(
3
):
for
j
in
xrange
(
3
):
# tz will not _always_ be overwritten,
# tz will not _always_ be overwritten,
# and adding update={...} in the call to function()
# and adding update={...} in the call to function()
...
@@ -355,9 +354,10 @@ class t_gemm(TestCase):
...
@@ -355,9 +354,10 @@ class t_gemm(TestCase):
tz_i
=
gemm_no_inplace
(
tz
[:,
:,
i
],
ta
,
tx
[
tz_i
=
gemm_no_inplace
(
tz
[:,
:,
i
],
ta
,
tx
[
:,
:,
i
],
ty
[:,
:,
i
],
tb
)
:,
:,
i
],
ty
[:,
:,
i
],
tb
)
g_i
=
theano
.
function
([],
tz_i
,
g_i
=
theano
.
function
(
updates
=
[(
tz
,
T
.
set_subtensor
(
tz
[:,
:,
i
],
tz_i
))],
[],
tz_i
,
updates
=
[(
tz
,
T
.
set_subtensor
(
tz
[:,
:,
i
],
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
tz_i
))],
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
for
j
in
xrange
(
3
):
for
j
in
xrange
(
3
):
g_i
()
g_i
()
unittest_tools
.
assert_allclose
(
z_after
[:,
:,
i
],
unittest_tools
.
assert_allclose
(
z_after
[:,
:,
i
],
...
@@ -404,8 +404,8 @@ class t_as_scalar(TestCase):
...
@@ -404,8 +404,8 @@ class t_as_scalar(TestCase):
def
test1
(
self
):
def
test1
(
self
):
"""Test that it fails on nonscalar constants"""
"""Test that it fails on nonscalar constants"""
a
=
T
.
constant
(
np
.
ones
(
5
))
a
=
T
.
constant
(
np
.
ones
(
5
))
self
.
assertTrue
(
None
==
_as_scalar
(
a
)
)
self
.
assertTrue
(
_as_scalar
(
a
)
is
None
)
self
.
assertTrue
(
None
==
_as_scalar
(
T
.
DimShuffle
([
False
],
[
0
,
'x'
])(
a
))
)
self
.
assertTrue
(
_as_scalar
(
T
.
DimShuffle
([
False
],
[
0
,
'x'
])(
a
))
is
None
)
def
test2
(
self
):
def
test2
(
self
):
"""Test that it works on scalar variables"""
"""Test that it works on scalar variables"""
...
@@ -420,9 +420,9 @@ class t_as_scalar(TestCase):
...
@@ -420,9 +420,9 @@ class t_as_scalar(TestCase):
def
test3
(
self
):
def
test3
(
self
):
"""Test that it fails on nonscalar variables"""
"""Test that it fails on nonscalar variables"""
a
=
T
.
matrix
()
a
=
T
.
matrix
()
self
.
assertTrue
(
None
==
_as_scalar
(
a
)
)
self
.
assertTrue
(
_as_scalar
(
a
)
is
None
)
self
.
assertTrue
(
None
==
_as_scalar
(
T
.
DimShuffle
([
False
,
False
],
self
.
assertTrue
(
_as_scalar
(
T
.
DimShuffle
([
False
,
False
],
[
0
,
'x'
,
1
])(
a
))
)
[
0
,
'x'
,
1
])(
a
))
is
None
)
class
T_real_matrix
(
TestCase
):
class
T_real_matrix
(
TestCase
):
...
@@ -458,10 +458,10 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
...
@@ -458,10 +458,10 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
max_graphlen
=
0
,
expected_nb_gemm
=
1
):
max_graphlen
=
0
,
expected_nb_gemm
=
1
):
try
:
try
:
f
=
inplace_func
(
f
=
inplace_func
(
[
In
(
ii
,
mutable
=
True
,
allow_downcast
=
True
)
for
ii
in
i
],
[
In
(
ii
,
mutable
=
True
,
allow_downcast
=
True
)
for
ii
in
i
],
o
,
o
,
mode
=
'FAST_RUN'
,
mode
=
'FAST_RUN'
,
on_unused_input
=
'ignore'
)
on_unused_input
=
'ignore'
)
nb_gemm
=
0
nb_gemm
=
0
for
node
in
f
.
maker
.
fgraph
.
apply_nodes
:
for
node
in
f
.
maker
.
fgraph
.
apply_nodes
:
if
isinstance
(
node
.
op
,
T
.
Dot
):
if
isinstance
(
node
.
op
,
T
.
Dot
):
...
@@ -472,7 +472,7 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
...
@@ -472,7 +472,7 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
nb_gemm
+=
1
nb_gemm
+=
1
assert
nb_gemm
==
expected_nb_gemm
,
(
nb_gemm
,
expected_nb_gemm
)
assert
nb_gemm
==
expected_nb_gemm
,
(
nb_gemm
,
expected_nb_gemm
)
g
=
inplace_func
(
i
,
o
,
mode
=
compile
.
Mode
(
linker
=
'py'
,
optimizer
=
None
),
g
=
inplace_func
(
i
,
o
,
mode
=
compile
.
Mode
(
linker
=
'py'
,
optimizer
=
None
),
allow_input_downcast
=
True
,
on_unused_input
=
'ignore'
)
allow_input_downcast
=
True
,
on_unused_input
=
'ignore'
)
for
node
in
g
.
maker
.
fgraph
.
apply_nodes
:
for
node
in
g
.
maker
.
fgraph
.
apply_nodes
:
if
node
.
op
==
gemm_inplace
:
if
node
.
op
==
gemm_inplace
:
raise
Exception
(
'gemm_inplace in original graph'
)
raise
Exception
(
'gemm_inplace in original graph'
)
...
@@ -492,7 +492,7 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
...
@@ -492,7 +492,7 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
eps
=
1.0e-8
eps
=
1.0e-8
if
config
.
floatX
==
'float32'
:
if
config
.
floatX
==
'float32'
:
eps
=
1.0e-6
eps
=
1.0e-6
if
max_abs_err
>
eps
:
if
max_abs_err
>
eps
:
raise
Failure
(
'GEMM is computing the wrong output. max_rel_err ='
,
raise
Failure
(
'GEMM is computing the wrong output. max_rel_err ='
,
max_abs_err
)
max_abs_err
)
except
Failure
:
except
Failure
:
...
@@ -516,7 +516,7 @@ def test_gemm_opt0():
...
@@ -516,7 +516,7 @@ def test_gemm_opt0():
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
.
T
-
a
*
T
.
dot
(
Y
.
T
,
X
.
T
)])
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
.
T
-
a
*
T
.
dot
(
Y
.
T
,
X
.
T
)])
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
.
T
+
a
*
b
*
T
.
dot
(
X
,
Y
)
.
T
])
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
.
T
+
a
*
b
*
T
.
dot
(
X
,
Y
)
.
T
])
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
+
a
*
T
.
dot
(
X
,
Y
)
.
T
],
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
+
a
*
T
.
dot
(
X
,
Y
)
.
T
],
ishapes
=
[(
5
,
3
),
(
3
,
4
),
(
4
,
5
),
(),
()])
ishapes
=
[(
5
,
3
),
(
3
,
4
),
(
4
,
5
),
(),
()])
# with N multiplications instead of just one
# with N multiplications instead of just one
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[(
b
*
b
)
*
Z
*
a
+
(
a
*
a
)
*
T
.
dot
(
X
,
Y
)
*
b
])
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[(
b
*
b
)
*
Z
*
a
+
(
a
*
a
)
*
T
.
dot
(
X
,
Y
)
*
b
])
...
@@ -541,32 +541,32 @@ def test_gemm_opt_double_gemm():
...
@@ -541,32 +541,32 @@ def test_gemm_opt_double_gemm():
ishapes
=
[(
4
,
3
),
(
3
,
5
),
(
4
,
5
),
(),
(),
(
5
,
9
),
(
9
,
4
),
()]
ishapes
=
[(
4
,
3
),
(
3
,
5
),
(
4
,
5
),
(),
(),
(
5
,
9
),
(
9
,
4
),
()]
i
=
[
X
,
Y
,
Z
,
a
,
b
,
R
,
S
,
c
]
i
=
[
X
,
Y
,
Z
,
a
,
b
,
R
,
S
,
c
]
o
=
[(
a
*
T
.
dot
(
X
,
Y
)
o
=
[(
a
*
T
.
dot
(
X
,
Y
)
+
+
gemm_inplace
(
Z
,
b
,
S
.
T
,
R
.
T
,
T
.
constant
(
1.0
)
.
astype
(
config
.
floatX
)))]
gemm_inplace
(
Z
,
b
,
S
.
T
,
R
.
T
,
T
.
constant
(
1.0
)
.
astype
(
config
.
floatX
)))]
try
:
try
:
f
=
inplace_func
([
In
(
ii
,
mutable
=
True
)
for
ii
in
i
],
o
,
f
=
inplace_func
([
In
(
ii
,
mutable
=
True
)
for
ii
in
i
],
o
,
mode
=
'FAST_RUN'
,
on_unused_input
=
'ignore'
)
mode
=
'FAST_RUN'
,
on_unused_input
=
'ignore'
)
for
node
in
f
.
maker
.
fgraph
.
apply_nodes
:
for
node
in
f
.
maker
.
fgraph
.
apply_nodes
:
if
isinstance
(
node
.
op
,
T
.
Dot
):
if
isinstance
(
node
.
op
,
T
.
Dot
):
raise
Failure
(
'dot in graph'
)
raise
Failure
(
'dot in graph'
)
if
node
.
op
==
_dot22
:
if
node
.
op
==
_dot22
:
raise
Failure
(
'_dot22 in graph'
)
raise
Failure
(
'_dot22 in graph'
)
g
=
inplace_func
(
i
,
o
,
mode
=
compile
.
Mode
(
linker
=
'py'
,
optimizer
=
None
),
g
=
inplace_func
(
i
,
o
,
mode
=
compile
.
Mode
(
linker
=
'py'
,
optimizer
=
None
),
on_unused_input
=
'ignore'
)
on_unused_input
=
'ignore'
)
# for node in g.maker.fgraph.apply_nodes:
# for node in g.maker.fgraph.apply_nodes:
# if node.op == gemm_inplace: raise Failure('gemm_inplace in graph')
# if node.op == gemm_inplace: raise Failure('gemm_inplace in graph')
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
(
234
))
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
(
234
))
r0
=
f
(
*
[
np
.
asarray
(
rng
.
randn
(
*
sh
),
config
.
floatX
)
r0
=
f
(
*
[
np
.
asarray
(
rng
.
randn
(
*
sh
),
config
.
floatX
)
for
sh
in
ishapes
])
for
sh
in
ishapes
])
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
(
234
))
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
(
234
))
r1
=
g
(
*
[
np
.
asarray
(
rng
.
randn
(
*
sh
),
config
.
floatX
)
r1
=
g
(
*
[
np
.
asarray
(
rng
.
randn
(
*
sh
),
config
.
floatX
)
for
sh
in
ishapes
])
for
sh
in
ishapes
])
max_abs_err
=
np
.
max
(
np
.
abs
(
r0
[
0
]
-
r1
[
0
]))
max_abs_err
=
np
.
max
(
np
.
abs
(
r0
[
0
]
-
r1
[
0
]))
eps
=
1.0e-8
eps
=
1.0e-8
if
config
.
floatX
==
'float32'
:
if
config
.
floatX
==
'float32'
:
eps
=
1.0e-6
eps
=
1.0e-6
if
max_abs_err
>
eps
:
if
max_abs_err
>
eps
:
raise
Failure
(
raise
Failure
(
'GEMM is computing the wrong output. max_rel_err ='
,
'GEMM is computing the wrong output. max_rel_err ='
,
max_abs_err
)
max_abs_err
)
...
@@ -579,8 +579,7 @@ def test_gemm_opt_double_gemm():
...
@@ -579,8 +579,7 @@ def test_gemm_opt_double_gemm():
def
test_gemm_canonicalize
():
def
test_gemm_canonicalize
():
X
,
Y
,
Z
,
a
,
b
=
T
.
matrix
(
'X'
),
T
.
matrix
(
'Y'
),
T
.
matrix
(
'Z'
),
T
.
scalar
(
X
,
Y
,
Z
,
a
,
b
=
T
.
matrix
(
'X'
),
T
.
matrix
(
'Y'
),
T
.
matrix
(
'Z'
),
T
.
scalar
(
'a'
),
T
.
scalar
(
'b'
)
'a'
),
T
.
scalar
(
'b'
)
R
,
S
,
U
,
c
,
d
=
T
.
matrix
(
'R'
),
T
.
matrix
(
'S'
),
T
.
matrix
(
'U'
),
T
.
scalar
(
c
,
d
=
T
.
scalar
(
'c'
),
T
.
scalar
(
'd'
)
'c'
),
T
.
scalar
(
'd'
)
u
=
T
.
row
(
'u'
)
u
=
T
.
row
(
'u'
)
v
=
T
.
vector
(
'v'
)
v
=
T
.
vector
(
'v'
)
w
=
T
.
col
(
'w'
)
w
=
T
.
col
(
'w'
)
...
@@ -631,10 +630,7 @@ def test_gemm_canonicalize():
...
@@ -631,10 +630,7 @@ def test_gemm_canonicalize():
def
test_gemm_factor
():
def
test_gemm_factor
():
X
,
Y
,
Z
,
a
,
b
=
T
.
matrix
(
'X'
),
T
.
matrix
(
'Y'
),
T
.
matrix
(
'Z'
),
T
.
scalar
(
X
,
Y
=
T
.
matrix
(
'X'
),
T
.
matrix
(
'Y'
)
'a'
),
T
.
scalar
(
'b'
)
R
,
S
,
U
,
c
,
d
=
T
.
matrix
(
'R'
),
T
.
matrix
(
'S'
),
T
.
matrix
(
'U'
),
T
.
scalar
(
'c'
),
T
.
scalar
(
'd'
)
assert
[(
1.0
,
X
),
(
1.0
,
Y
)]
==
_factor_canonicalized
([(
1.0
,
X
),
(
1.0
,
Y
)])
assert
[(
1.0
,
X
),
(
1.0
,
Y
)]
==
_factor_canonicalized
([(
1.0
,
X
),
(
1.0
,
Y
)])
assert
[(
2.0
,
X
)]
==
_factor_canonicalized
([(
1.0
,
X
),
(
1.0
,
X
)])
assert
[(
2.0
,
X
)]
==
_factor_canonicalized
([(
1.0
,
X
),
(
1.0
,
X
)])
...
@@ -653,7 +649,7 @@ def test_upcasting_scalar_nogemm():
...
@@ -653,7 +649,7 @@ def test_upcasting_scalar_nogemm():
f
=
theano
.
function
([
w
,
v
,
t
,
alpha
],
rval
)
f
=
theano
.
function
([
w
,
v
,
t
,
alpha
],
rval
)
t
=
f
.
maker
.
fgraph
.
toposort
()
t
=
f
.
maker
.
fgraph
.
toposort
()
assert
np
.
sum
([
isinstance
(
n
.
op
,
Gemm
)
for
n
in
t
])
==
0
assert
np
.
sum
([
isinstance
(
n
.
op
,
Gemm
)
for
n
in
t
])
==
0
#theano.printing.debugprint(f, print_type=True)
#
theano.printing.debugprint(f, print_type=True)
v
=
T
.
fmatrix
(
'v'
)
v
=
T
.
fmatrix
(
'v'
)
w
=
T
.
fmatrix
(
'w'
)
w
=
T
.
fmatrix
(
'w'
)
...
@@ -670,7 +666,7 @@ def test_upcasting_scalar_nogemm():
...
@@ -670,7 +666,7 @@ def test_upcasting_scalar_nogemm():
t
=
f
.
maker
.
fgraph
.
toposort
()
t
=
f
.
maker
.
fgraph
.
toposort
()
assert
np
.
sum
([
isinstance
(
n
.
op
,
Gemm
)
for
n
in
t
])
==
0
assert
np
.
sum
([
isinstance
(
n
.
op
,
Gemm
)
for
n
in
t
])
==
0
#theano.printing.debugprint(f, print_type=True)
#
theano.printing.debugprint(f, print_type=True)
def
test_gemm_nested
():
def
test_gemm_nested
():
...
@@ -680,22 +676,22 @@ def test_gemm_nested():
...
@@ -680,22 +676,22 @@ def test_gemm_nested():
'c'
),
T
.
scalar
(
'd'
)
'c'
),
T
.
scalar
(
'd'
)
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
)],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
2
,
4
),
(),
(),
(),
()],
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
1
)
max_graphlen
=
1
)
# print "---------------------"
# print "---------------------"
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
+
c
*
Z
)],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
+
c
*
Z
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
2
,
4
),
(),
(),
(),
()],
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
1
)
max_graphlen
=
1
)
# print "---------------------"
# print "---------------------"
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
+
c
*
U
)],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
+
c
*
U
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
2
,
4
),
(),
(),
(),
()],
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
3
)
max_graphlen
=
3
)
def
test_gemm_opt_wishlist
():
def
test_gemm_opt_wishlist
():
...
@@ -720,7 +716,7 @@ def test_gemm_with_vector():
...
@@ -720,7 +716,7 @@ def test_gemm_with_vector():
def
my_just_gemm
(
o
):
def
my_just_gemm
(
o
):
i
=
[
X
,
Y
,
Z
,
a
,
b
,
v
]
i
=
[
X
,
Y
,
Z
,
a
,
b
,
v
]
ishapes
=
[(
4
,
3
),
(
3
,
5
),
(
4
,
5
),
(),
(),
(
5
,
)]
ishapes
=
[(
4
,
3
),
(
3
,
5
),
(
4
,
5
),
(),
(),
(
5
,
)]
rval
=
just_gemm
(
i
,
o
,
ishapes
=
ishapes
)
just_gemm
(
i
,
o
,
ishapes
=
ishapes
)
my_just_gemm
([
v
+
T
.
dot
(
X
,
Y
)
*
a
+
Z
*
b
])
my_just_gemm
([
v
+
T
.
dot
(
X
,
Y
)
*
a
+
Z
*
b
])
my_just_gemm
([
v
+
a
*
T
.
dot
(
X
,
Y
)
+
b
*
Z
])
my_just_gemm
([
v
+
a
*
T
.
dot
(
X
,
Y
)
+
b
*
Z
])
...
@@ -741,7 +737,7 @@ def test_gemm_with_vector():
...
@@ -741,7 +737,7 @@ def test_gemm_with_vector():
def
test_gemm_opt_vector_stuff
():
def
test_gemm_opt_vector_stuff
():
X
,
Y
,
Z
,
a
,
b
=
T
.
matrix
(),
T
.
matrix
(),
T
.
matrix
(),
T
.
scalar
(),
T
.
scalar
()
X
,
Y
,
a
=
T
.
matrix
(),
T
.
matrix
(),
T
.
scalar
()
u
,
v
=
T
.
vector
(),
T
.
vector
()
u
,
v
=
T
.
vector
(),
T
.
vector
()
f
=
inplace_func
([
a
,
u
,
v
],
a
+
T
.
dot
(
u
,
v
),
mode
=
'FAST_RUN'
)
f
=
inplace_func
([
a
,
u
,
v
],
a
+
T
.
dot
(
u
,
v
),
mode
=
'FAST_RUN'
)
...
@@ -771,8 +767,6 @@ def test_gemm_unrolled():
...
@@ -771,8 +767,6 @@ def test_gemm_unrolled():
H
=
sharedX
(
np
.
zeros
((
batch_size
,
rep_size
)),
name
=
'H'
)
H
=
sharedX
(
np
.
zeros
((
batch_size
,
rep_size
)),
name
=
'H'
)
G
=
sharedX
(
np
.
zeros
((
batch_size
,
rep_size
)),
name
=
'G'
)
G
=
sharedX
(
np
.
zeros
((
batch_size
,
rep_size
)),
name
=
'G'
)
init_V
=
sharedX
(
rng
.
uniform
(
0
,
1
,
(
batch_size
,
rep_size
)),
name
=
'init_V'
)
init_H
=
sharedX
(
rng
.
uniform
(
0
,
1
,
(
batch_size
,
rep_size
)),
name
=
'init_H'
)
cur_V
=
V
cur_V
=
V
cur_H
=
H
cur_H
=
H
...
@@ -787,7 +781,7 @@ def test_gemm_unrolled():
...
@@ -787,7 +781,7 @@ def test_gemm_unrolled():
cur_H
=
update_H
(
cur_V
)
cur_H
=
update_H
(
cur_V
)
unrolled_theano
=
theano
.
function
([],
updates
=
[(
V
,
cur_V
),
(
H
,
cur_H
)],
unrolled_theano
=
theano
.
function
([],
updates
=
[(
V
,
cur_V
),
(
H
,
cur_H
)],
name
=
'unrolled_theano'
)
name
=
'unrolled_theano'
)
nb_dot
=
sum
([
1
for
node
in
unrolled_theano
.
maker
.
fgraph
.
toposort
()
nb_dot
=
sum
([
1
for
node
in
unrolled_theano
.
maker
.
fgraph
.
toposort
()
if
isinstance
(
node
.
op
,
(
theano
.
tensor
.
Dot
,
if
isinstance
(
node
.
op
,
(
theano
.
tensor
.
Dot
,
theano
.
tensor
.
blas
.
Dot22
,
theano
.
tensor
.
blas
.
Dot22
,
...
@@ -807,7 +801,7 @@ def test_inplace0():
...
@@ -807,7 +801,7 @@ def test_inplace0():
R
,
S
,
c
=
T
.
matrix
(
'R'
),
T
.
matrix
(
'S'
),
T
.
scalar
(
'c'
)
R
,
S
,
c
=
T
.
matrix
(
'R'
),
T
.
matrix
(
'S'
),
T
.
scalar
(
'c'
)
f
=
inplace_func
([
Z
,
b
,
R
,
S
],
f
=
inplace_func
([
Z
,
b
,
R
,
S
],
[
Z
*
(
Z
+
b
*
T
.
dot
(
R
,
S
)
.
T
)],
mode
=
'FAST_RUN'
)
[
Z
*
(
Z
+
b
*
T
.
dot
(
R
,
S
)
.
T
)],
mode
=
'FAST_RUN'
)
if
(
gemm_inplace
in
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]):
if
(
gemm_inplace
in
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]):
print
(
pp
(
f
.
maker
.
fgraph
.
outputs
[
0
]))
print
(
pp
(
f
.
maker
.
fgraph
.
outputs
[
0
]))
raise
Failure
(
'gemm_inplace in graph'
)
raise
Failure
(
'gemm_inplace in graph'
)
...
@@ -815,9 +809,9 @@ def test_inplace0():
...
@@ -815,9 +809,9 @@ def test_inplace0():
# gemm_inplace should be inserted here, to work in-place on Z*c
# gemm_inplace should be inserted here, to work in-place on Z*c
f
=
inplace_func
([
X
,
Y
,
Z
,
a
,
b
,
R
,
S
,
c
],
f
=
inplace_func
([
X
,
Y
,
Z
,
a
,
b
,
R
,
S
,
c
],
[
Z
*
(
c
*
Z
+
a
*
T
.
dot
(
X
,
Y
)
+
b
*
T
.
dot
(
R
,
S
)
.
T
)],
[
Z
*
(
c
*
Z
+
a
*
T
.
dot
(
X
,
Y
)
+
b
*
T
.
dot
(
R
,
S
)
.
T
)],
mode
=
'FAST_RUN'
)
mode
=
'FAST_RUN'
)
if
(
not
gemm_inplace
in
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]):
if
(
gemm_inplace
not
in
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]):
theano
.
printing
.
debugprint
(
f
)
theano
.
printing
.
debugprint
(
f
)
raise
Failure
(
'no gemm_inplace in graph'
)
raise
Failure
(
'no gemm_inplace in graph'
)
...
@@ -826,7 +820,7 @@ def test_inplace1():
...
@@ -826,7 +820,7 @@ def test_inplace1():
X
,
Y
,
Z
,
a
,
b
=
XYZab
()
X
,
Y
,
Z
,
a
,
b
=
XYZab
()
# with > 2 terms in the overall addition
# with > 2 terms in the overall addition
f
=
inplace_func
([
X
,
Y
,
Z
],
f
=
inplace_func
([
X
,
Y
,
Z
],
[
Z
+
Z
+
T
.
dot
(
X
,
Y
)],
mode
=
'FAST_RUN'
)
[
Z
+
Z
+
T
.
dot
(
X
,
Y
)],
mode
=
'FAST_RUN'
)
# theano.printing.debugprint(f)
# theano.printing.debugprint(f)
# it doesn't work inplace because we didn't mark Z as mutable input
# it doesn't work inplace because we didn't mark Z as mutable input
assert
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]
==
[
gemm_no_inplace
]
assert
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]
==
[
gemm_no_inplace
]
...
@@ -866,7 +860,7 @@ def test_dot22scalar():
...
@@ -866,7 +860,7 @@ def test_dot22scalar():
# TODO: exclude other optimizations in BlasOpt?
# TODO: exclude other optimizations in BlasOpt?
# m = theano.compile.get_default_mode().including('local_dot_to_dot22',
# m = theano.compile.get_default_mode().including('local_dot_to_dot22',
# 'local_dot22_to_dot22scalar','specialize')
# 'local_dot22_to_dot22scalar','specialize')
#m = theano.compile.get_default_mode().including('BlasOpt', 'specialize')
#
m = theano.compile.get_default_mode().including('BlasOpt', 'specialize')
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
for
dtype1
in
[
'complex64'
,
'complex128'
]:
for
dtype1
in
[
'complex64'
,
'complex128'
]:
a
=
T
.
matrix
(
'a'
,
dtype
=
dtype1
)
a
=
T
.
matrix
(
'a'
,
dtype
=
dtype1
)
...
@@ -1087,7 +1081,7 @@ def test_dot_w_self():
...
@@ -1087,7 +1081,7 @@ def test_dot_w_self():
class
TestGemv
(
TestCase
,
unittest_tools
.
TestOptimizationMixin
):
class
TestGemv
(
TestCase
,
unittest_tools
.
TestOptimizationMixin
):
def
test_dot_vv
(
self
):
def
test_dot_vv
(
self
):
''' Currently we generate a gemv for that case'''
# Currently we generate a gemv for that case
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
w
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
w
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
...
@@ -1101,11 +1095,10 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1101,11 +1095,10 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
w
.
get_value
()))
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
w
.
get_value
()))
def
test_dot_vm
(
self
):
def
test_dot_vm
(
self
):
''' Test vector dot matrix '''
# Test vector dot matrix
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,
3
)),
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,
3
)),
dtype
=
'float32'
))
dtype
=
'float32'
))
f
=
theano
.
function
([],
theano
.
dot
(
v
,
m
),
mode
=
mode_blas_opt
)
f
=
theano
.
function
([],
theano
.
dot
(
v
,
m
),
mode
=
mode_blas_opt
)
# Assert that the dot was optimized somehow
# Assert that the dot was optimized somehow
...
@@ -1115,17 +1108,14 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1115,17 +1108,14 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
# Assert they produce the same output
# Assert they produce the same output
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
m
.
get_value
()))
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
m
.
get_value
()))
# Assert it works when m has no contiguous dimension
# Assert it works when m has no contiguous dimension
m
.
set_value
(
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
m
.
get_value
()))
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
m
.
get_value
()))
def
test_dot_mv
(
self
):
def
test_dot_mv
(
self
):
''' Test matrix dot vector '''
# Test matrix dot vector
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
3
,
2
)),
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
3
,
2
)),
dtype
=
'float32'
))
dtype
=
'float32'
))
f
=
theano
.
function
([],
theano
.
dot
(
m
,
v
),
mode
=
mode_blas_opt
)
f
=
theano
.
function
([],
theano
.
dot
(
m
,
v
),
mode
=
mode_blas_opt
)
# Assert that the dot was optimized somehow
# Assert that the dot was optimized somehow
...
@@ -1135,31 +1125,27 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1135,31 +1125,27 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
# Assert they produce the same output
# Assert they produce the same output
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v
.
get_value
()))
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v
.
get_value
()))
# Assert it works when m has no contiguous dimension
# Assert it works when m has no contiguous dimension
m
.
set_value
(
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v
.
get_value
()))
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v
.
get_value
()))
@staticmethod
@staticmethod
def
t_gemv1
(
m_shp
):
def
t_gemv1
(
m_shp
):
''' test vector2+dot(matrix,vector1) '''
# test vector2+dot(matrix,vector1)
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
m_shp
[
1
],)
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
m_shp
[
1
],)
),
),
dtype
=
'float32'
))
dtype
=
'float32'
))
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
m_shp
[
0
],)),
dtype
=
'float32'
)
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
m_shp
[
0
],)),
dtype
=
'float32'
)
v2
=
theano
.
shared
(
v2_orig
)
v2
=
theano
.
shared
(
v2_orig
)
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
m_shp
),
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
m_shp
),
dtype
=
'float32'
))
dtype
=
'float32'
))
f
=
theano
.
function
([],
v2
+
theano
.
dot
(
m
,
v1
),
mode
=
mode_blas_opt
)
f
=
theano
.
function
([],
v2
+
theano
.
dot
(
m
,
v1
),
mode
=
mode_blas_opt
)
# Assert they produce the same output
# Assert they produce the same output
assert
np
.
allclose
(
f
(),
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
topo
=
f
.
maker
.
fgraph
.
toposort
()
topo
=
f
.
maker
.
fgraph
.
toposort
()
assert
len
(
topo
)
==
1
assert
len
(
topo
)
==
1
assert
isinstance
(
topo
[
0
]
.
op
,
Gemv
)
assert
isinstance
(
topo
[
0
]
.
op
,
Gemv
)
assert
topo
[
0
]
.
op
.
inplace
==
False
assert
topo
[
0
]
.
op
.
inplace
is
False
# test the inplace version
# test the inplace version
g
=
theano
.
function
([],
[],
updates
=
[(
v2
,
v2
+
theano
.
dot
(
m
,
v1
))],
g
=
theano
.
function
([],
[],
updates
=
[(
v2
,
v2
+
theano
.
dot
(
m
,
v1
))],
...
@@ -1167,24 +1153,23 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1167,24 +1153,23 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
# Assert they produce the same output
# Assert they produce the same output
g
()
g
()
assert
np
.
allclose
(
v2
.
get_value
(),
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
m
.
get_value
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
v1
.
get_value
())
+
v2_orig
)
topo
=
g
.
maker
.
fgraph
.
toposort
()
topo
=
g
.
maker
.
fgraph
.
toposort
()
assert
len
(
topo
)
==
1
assert
len
(
topo
)
==
1
assert
isinstance
(
topo
[
0
]
.
op
,
Gemv
)
assert
isinstance
(
topo
[
0
]
.
op
,
Gemv
)
if
config
.
mode
!=
'FAST_COMPILE'
:
if
config
.
mode
!=
'FAST_COMPILE'
:
assert
topo
[
0
]
.
op
.
inplace
==
True
assert
topo
[
0
]
.
op
.
inplace
is
True
# Do the same tests with a matrix with strides in both dimensions
# Do the same tests with a matrix with strides in both dimensions
m
.
set_value
(
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
borrow
=
True
)
v2
.
set_value
(
v2_orig
)
v2
.
set_value
(
v2_orig
)
assert
np
.
allclose
(
f
(),
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
g
()
g
()
assert
np
.
allclose
(
v2
.
get_value
(),
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
@attr
(
'slow'
)
@attr
(
'slow'
)
def
test_gemv1
(
self
):
def
test_gemv1
(
self
):
...
@@ -1194,23 +1179,24 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1194,23 +1179,24 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
self
.
t_gemv1
((
0
,
0
))
self
.
t_gemv1
((
0
,
0
))
def
test_gemv2
(
self
):
def
test_gemv2
(
self
):
''' test vector2+dot(vector1,matrix) '''
# test vector2+dot(vector1,matrix)
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
dtype
=
'float32'
))
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
3
,)),
dtype
=
'float32'
)
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
3
,)),
dtype
=
'float32'
)
v2
=
theano
.
shared
(
v2_orig
)
v2
=
theano
.
shared
(
v2_orig
)
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,
3
)),
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,
3
)),
dtype
=
'float32'
))
dtype
=
'float32'
))
f
=
theano
.
function
([],
v2
+
theano
.
dot
(
v1
,
m
),
mode
=
mode_blas_opt
)
f
=
theano
.
function
([],
v2
+
theano
.
dot
(
v1
,
m
),
mode
=
mode_blas_opt
)
# Assert they produce the same output
# Assert they produce the same output
assert
np
.
allclose
(
f
(),
assert
np
.
allclose
(
f
(),
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2
.
get_value
())
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2
.
get_value
())
topo
=
f
.
maker
.
fgraph
.
toposort
()
topo
=
f
.
maker
.
fgraph
.
toposort
()
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
assert
topo
[
-
1
]
.
op
.
inplace
==
False
assert
topo
[
-
1
]
.
op
.
inplace
is
False
# test the inplace version
# test the inplace version
g
=
theano
.
function
([],
[],
updates
=
[(
v2
,
v2
+
theano
.
dot
(
v1
,
m
))],
g
=
theano
.
function
([],
[],
updates
=
[(
v2
,
v2
+
theano
.
dot
(
v1
,
m
))],
...
@@ -1219,32 +1205,32 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1219,32 +1205,32 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
# Assert they produce the same output
# Assert they produce the same output
g
()
g
()
assert
np
.
allclose
(
v2
.
get_value
(),
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2_orig
)
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2_orig
)
topo
=
g
.
maker
.
fgraph
.
toposort
()
topo
=
g
.
maker
.
fgraph
.
toposort
()
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
if
config
.
mode
!=
'FAST_COMPILE'
:
if
config
.
mode
!=
'FAST_COMPILE'
:
assert
topo
[
-
1
]
.
op
.
inplace
==
True
assert
topo
[
-
1
]
.
op
.
inplace
is
True
# Do the same tests with a matrix with strides in both dimensions
# Do the same tests with a matrix with strides in both dimensions
m
.
set_value
(
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
borrow
=
True
)
v2
.
set_value
(
v2_orig
)
v2
.
set_value
(
v2_orig
)
assert
np
.
allclose
(
f
(),
assert
np
.
allclose
(
f
(),
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2
.
get_value
())
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2
.
get_value
())
g
()
g
()
assert
np
.
allclose
(
v2
.
get_value
(),
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2_orig
)
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2_orig
)
def
test_gemv_broadcast
(
self
):
def
test_gemv_broadcast
(
self
):
''' test gemv with some broadcasted input '''
# test gemv with some broadcasted input
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
dtype
=
'float32'
))
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
1
,)),
dtype
=
'float32'
)
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
1
,)),
dtype
=
'float32'
)
v2
=
theano
.
shared
(
v2_orig
)
v2
=
theano
.
shared
(
v2_orig
)
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
1
,
2
)),
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
1
,
2
)),
dtype
=
'float32'
),
dtype
=
'float32'
),
broadcastable
=
(
True
,
False
))
broadcastable
=
(
True
,
False
))
o
=
theano
.
dot
(
m
,
v1
)
o
=
theano
.
dot
(
m
,
v1
)
f
=
theano
.
function
([],
o
+
v2
,
mode
=
mode_blas_opt
)
f
=
theano
.
function
([],
o
+
v2
,
mode
=
mode_blas_opt
)
...
@@ -1261,7 +1247,7 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1261,7 +1247,7 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
f
=
theano
.
function
([],
o
,
mode
=
mode_blas_opt
)
f
=
theano
.
function
([],
o
,
mode
=
mode_blas_opt
)
assert
np
.
allclose
(
assert
np
.
allclose
(
f
(),
f
(),
0.5
*
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
0.25
*
v2
.
get_value
())
0.5
*
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
0.25
*
v2
.
get_value
())
topo
=
f
.
maker
.
fgraph
.
toposort
()
topo
=
f
.
maker
.
fgraph
.
toposort
()
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
...
@@ -1269,9 +1255,9 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1269,9 +1255,9 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
A
=
T
.
matrix
(
'A'
)
A
=
T
.
matrix
(
'A'
)
x
,
y
=
T
.
vectors
(
'x'
,
'y'
)
x
,
y
=
T
.
vectors
(
'x'
,
'y'
)
alpha
=
theano
.
shared
(
theano
.
_asarray
(
1.0
,
dtype
=
config
.
floatX
),
alpha
=
theano
.
shared
(
theano
.
_asarray
(
1.0
,
dtype
=
config
.
floatX
),
name
=
'alpha'
)
name
=
'alpha'
)
beta
=
theano
.
shared
(
theano
.
_asarray
(
1.0
,
dtype
=
config
.
floatX
),
beta
=
theano
.
shared
(
theano
.
_asarray
(
1.0
,
dtype
=
config
.
floatX
),
name
=
'beta'
)
name
=
'beta'
)
z
=
beta
*
y
+
alpha
*
T
.
dot
(
A
,
x
)
z
=
beta
*
y
+
alpha
*
T
.
dot
(
A
,
x
)
f
=
theano
.
function
([
A
,
x
,
y
],
z
)
f
=
theano
.
function
([
A
,
x
,
y
],
z
)
...
@@ -1335,9 +1321,9 @@ class BaseGemv(object):
...
@@ -1335,9 +1321,9 @@ class BaseGemv(object):
def
test_simple
(
self
):
def
test_simple
(
self
):
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
value
)
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
value
)
for
value
in
self
.
get_data
()]
for
value
in
self
.
get_data
()]
desired_oy
=
alpha
.
get_value
()
*
matrixmultiply
(
a
.
desired_oy
=
alpha
.
get_value
()
*
get_value
(),
x
.
get_value
())
+
beta
.
get_value
()
*
y
.
get_value
()
matrixmultiply
(
a
.
get_value
(),
x
.
get_value
())
+
beta
.
get_value
()
*
y
.
get_value
()
oy
=
alpha
*
T
.
dot
(
a
,
x
)
+
beta
*
y
oy
=
alpha
*
T
.
dot
(
a
,
x
)
+
beta
*
y
...
@@ -1406,8 +1392,8 @@ class BaseGemv(object):
...
@@ -1406,8 +1392,8 @@ class BaseGemv(object):
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
desired_oy
=
alpha_v
*
matrixmultiply
(
transpose
(
a_v
),
x_v
[::
desired_oy
=
alpha_v
*
matrixmultiply
(
transpose
(
a_v
),
x_v
[::
2
])
+
2
])
+
beta_v
*
y_v
beta_v
*
y_v
oy
=
alpha
*
T
.
dot
(
a
.
T
,
x
[::
2
])
+
beta
*
y
oy
=
alpha
*
T
.
dot
(
a
.
T
,
x
[::
2
])
+
beta
*
y
...
@@ -1456,10 +1442,9 @@ class BaseGemv(object):
...
@@ -1456,10 +1442,9 @@ class BaseGemv(object):
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
a_v
=
a_v
[::
-
1
,
::
-
1
]
a_v
=
a_v
[::
-
1
,
::
-
1
]
a
.
set_value
(
a
.
set_value
(
a
.
get_value
(
borrow
=
True
,
a
.
get_value
(
borrow
=
True
,
return_internal_type
=
True
)[::
-
1
,
::
-
1
],
return_internal_type
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
borrow
=
True
)
desired_oy
=
alpha_v
*
matrixmultiply
(
a_v
,
x_v
)
+
beta_v
*
y_v
desired_oy
=
alpha_v
*
matrixmultiply
(
a_v
,
x_v
)
+
beta_v
*
y_v
...
@@ -1477,10 +1462,9 @@ class BaseGemv(object):
...
@@ -1477,10 +1462,9 @@ class BaseGemv(object):
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
a_v
=
a_v
[::
-
1
,
::
-
1
]
a_v
=
a_v
[::
-
1
,
::
-
1
]
a
.
set_value
(
a
.
set_value
(
a
.
get_value
(
borrow
=
True
,
a
.
get_value
(
borrow
=
True
,
return_internal_type
=
True
)[::
-
1
,
::
-
1
],
return_internal_type
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
borrow
=
True
)
desired_oy
=
alpha_v
*
matrixmultiply
(
transpose
(
a_v
),
desired_oy
=
alpha_v
*
matrixmultiply
(
transpose
(
a_v
),
x_v
)
+
beta_v
*
y_v
x_v
)
+
beta_v
*
y_v
...
@@ -1517,7 +1501,7 @@ class BaseGemv(object):
...
@@ -1517,7 +1501,7 @@ class BaseGemv(object):
# done inplace on a temporarily allocated-buffer, which is
# done inplace on a temporarily allocated-buffer, which is
# then scaled by alpha and to t with a fused elemwise.
# then scaled by alpha and to t with a fused elemwise.
n_gemvs
=
0
n_gemvs
=
0
#theano.printing.debugprint(f, print_type=True)
#
theano.printing.debugprint(f, print_type=True)
for
node
in
f
.
maker
.
fgraph
.
toposort
():
for
node
in
f
.
maker
.
fgraph
.
toposort
():
if
node
.
op
==
self
.
gemv_inplace
:
if
node
.
op
==
self
.
gemv_inplace
:
n_gemvs
+=
1
n_gemvs
+=
1
...
@@ -1580,39 +1564,42 @@ class TestGer_make_node(TestCase):
...
@@ -1580,39 +1564,42 @@ class TestGer_make_node(TestCase):
def
test_works_on_all_valid_dtypes
(
self
):
def
test_works_on_all_valid_dtypes
(
self
):
self
.
assertEqual
(
self
.
fm
.
type
,
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
self
.
assertEqual
(
self
.
fm
.
type
,
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
self
.
assertEqual
(
self
.
fm
.
type
,
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
self
.
assertEqual
(
self
.
fm
.
type
,
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
def
test_fails_on_invalid_dtypes
(
self
):
def
test_fails_on_invalid_dtypes
(
self
):
self
.
assertRaises
(
TypeError
,
self
.
assertRaises
(
TypeError
,
ger
,
T
.
imatrix
(),
T
.
iscalar
(),
T
.
ivector
(),
ger
,
T
.
imatrix
(),
T
.
iscalar
(),
T
.
ivector
(),
T
.
ivector
())
T
.
ivector
())
def
test_fails_for_nonscalar_alpha
(
self
):
def
test_fails_for_nonscalar_alpha
(
self
):
self
.
assertRaises
(
TypeError
,
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fm
,
self
.
fm
,
self
.
fv
,
self
.
fv_2
)
ger
,
self
.
fm
,
self
.
fm
,
self
.
fv
,
self
.
fv_2
)
# boundary case - fv1 has the right dtype and could be dimshuffled to a
# boundary case - fv1 has the right dtype and could be dimshuffled to a
# scalar, but that's not make_node's job.
# scalar, but that's not make_node's job.
self
.
assertRaises
(
TypeError
,
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fm
,
self
.
fv1
,
self
.
fv
,
self
.
fv_2
)
ger
,
self
.
fm
,
self
.
fv1
,
self
.
fv
,
self
.
fv_2
)
# actually doing the aforementioned dimshuffle makes it work
# actually doing the aforementioned dimshuffle makes it work
self
.
assertEqual
(
self
.
fm
.
type
,
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fv1
.
dimshuffle
(),
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fv1
.
dimshuffle
(),
self
.
fv
,
self
.
fv_2
)
.
type
)
def
test_fails_for_nonmatrix_A
(
self
):
def
test_fails_for_nonmatrix_A
(
self
):
self
.
assertRaises
(
TypeError
,
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fv
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
ger
,
self
.
fv
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
def
test_fails_for_nonvector_x_or_y
(
self
):
def
test_fails_for_nonvector_x_or_y
(
self
):
self
.
assertRaises
(
TypeError
,
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fm
,
self
.
fa
,
self
.
fv
.
dimshuffle
(
'x'
,
0
),
self
.
fv_2
)
ger
,
self
.
fm
,
self
.
fa
,
self
.
fv
.
dimshuffle
(
'x'
,
0
),
self
.
fv_2
)
self
.
assertRaises
(
TypeError
,
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
.
dimshuffle
(
'x'
,
0
))
ger
,
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
.
dimshuffle
(
'x'
,
0
))
def
test_fails_for_mixed_dtypes
(
self
):
def
test_fails_for_mixed_dtypes
(
self
):
self
.
assertRaises
(
TypeError
,
ger
,
self
.
dm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
self
.
assertRaises
(
TypeError
,
ger
,
self
.
dm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
...
@@ -1657,30 +1644,26 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1657,30 +1644,26 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
def
test_b_0_triggers_ger
(
self
):
def
test_b_0_triggers_ger
(
self
):
""" test local_gemm_to_ger opt"""
""" test local_gemm_to_ger opt"""
assert
T
.
blas
.
local_gemm_to_ger
.
transform
(
assert
T
.
blas
.
local_gemm_to_ger
.
transform
(
gemm_no_inplace
(
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
0
))
.
owner
)
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
0
))
.
owner
)
def
test_b_1_triggers_ger
(
self
):
def
test_b_1_triggers_ger
(
self
):
""" test local_gemm_to_ger opt"""
""" test local_gemm_to_ger opt"""
assert
T
.
blas
.
local_gemm_to_ger
.
transform
(
assert
T
.
blas
.
local_gemm_to_ger
.
transform
(
gemm_no_inplace
(
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
1
))
.
owner
)
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
1
))
.
owner
)
def
test_b_other_does_not_triggers_ger
(
self
):
def
test_b_other_does_not_triggers_ger
(
self
):
""" test local_gemm_to_ger opt"""
""" test local_gemm_to_ger opt"""
assert
not
T
.
blas
.
local_gemm_to_ger
.
transform
(
assert
not
T
.
blas
.
local_gemm_to_ger
.
transform
(
gemm_no_inplace
(
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
1.5
))
.
owner
)
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
1.5
))
.
owner
)
def
test_b_nonconst_does_not_triggers_ger
(
self
):
def
test_b_nonconst_does_not_triggers_ger
(
self
):
""" test local_gemm_to_ger opt"""
""" test local_gemm_to_ger opt"""
assert
not
T
.
blas
.
local_gemm_to_ger
.
transform
(
assert
not
T
.
blas
.
local_gemm_to_ger
.
transform
(
gemm_no_inplace
(
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
a
)
.
owner
)
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
a
)
.
owner
)
def
test_outer
(
self
):
def
test_outer
(
self
):
f
=
self
.
function
([
self
.
x
,
self
.
y
],
T
.
outer
(
self
.
x
,
self
.
y
))
f
=
self
.
function
([
self
.
x
,
self
.
y
],
T
.
outer
(
self
.
x
,
self
.
y
))
...
@@ -1690,7 +1673,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1690,7 +1673,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
def
test_A_plus_outer
(
self
):
def
test_A_plus_outer
(
self
):
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
self
.
A
+
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
A
+
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
assertFunctionContains
(
f
,
self
.
ger
)
self
.
assertFunctionContains
(
f
,
self
.
ger
)
f
(
np
.
random
.
rand
(
5
,
4
)
.
astype
(
self
.
dtype
),
f
(
np
.
random
.
rand
(
5
,
4
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
5
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
5
)
.
astype
(
self
.
dtype
),
...
@@ -1701,7 +1684,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1701,7 +1684,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
def
test_A_plus_scaled_outer
(
self
):
def
test_A_plus_scaled_outer
(
self
):
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
self
.
A
+
0.1
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
A
+
0.1
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
assertFunctionContains
(
f
,
self
.
ger
)
self
.
assertFunctionContains
(
f
,
self
.
ger
)
f
(
np
.
random
.
rand
(
5
,
4
)
.
astype
(
self
.
dtype
),
f
(
np
.
random
.
rand
(
5
,
4
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
5
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
5
)
.
astype
(
self
.
dtype
),
...
@@ -1714,7 +1697,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1714,7 +1697,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
np
.
asarray
(
0.2
,
self
.
dtype
)
*
self
.
A
+
np
.
asarray
(
0.2
,
self
.
dtype
)
*
self
.
A
+
np
.
asarray
(
0.1
,
self
.
dtype
)
*
T
.
outer
(
np
.
asarray
(
0.1
,
self
.
dtype
)
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
x
,
self
.
y
))
# Why gemm? This make the graph simpler did we test that it
# Why gemm? This make the graph simpler did we test that it
# make it faster?
# make it faster?
self
.
assertFunctionContains
(
f
,
self
.
gemm
)
self
.
assertFunctionContains
(
f
,
self
.
gemm
)
...
@@ -1729,7 +1712,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
...
@@ -1729,7 +1712,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
""" test corner case shape and dtype"""
""" test corner case shape and dtype"""
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
self
.
A
+
0.1
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
A
+
0.1
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
assertFunctionContains
(
f
,
self
.
ger
)
self
.
assertFunctionContains
(
f
,
self
.
ger
)
f
(
np
.
random
.
rand
(
M
,
N
)
.
astype
(
self
.
dtype
),
f
(
np
.
random
.
rand
(
M
,
N
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
M
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
M
)
.
astype
(
self
.
dtype
),
...
@@ -1812,21 +1795,21 @@ class TestBlasStrides(TestCase):
...
@@ -1812,21 +1795,21 @@ class TestBlasStrides(TestCase):
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_nn
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b
,
c
))],
f_nn
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b
,
c
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
# print 'class name:', self.__class__.__name__
# print 'class name:', self.__class__.__name__
# theano.printing.debugprint(f_nn)
# theano.printing.debugprint(f_nn)
f_nt
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b
,
c_t
.
T
))],
f_nt
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b
,
c_t
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tn
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b_t
.
T
,
c
))],
f_tn
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b_t
.
T
,
c
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tt
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
))],
f_tt
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed pattern
# Try with all stride patterns, and all transposed pattern
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
4
):
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
4
):
for
step
in
(
1
,
2
):
for
step
in
(
1
,
2
):
b_step1
,
b_step2
,
c_step1
,
c_step2
=
(
s
*
step
b_step1
,
b_step2
,
c_step1
,
c_step2
=
(
s
*
step
for
s
in
step_signs
)
for
s
in
step_signs
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
...
@@ -1835,7 +1818,7 @@ class TestBlasStrides(TestCase):
...
@@ -1835,7 +1818,7 @@ class TestBlasStrides(TestCase):
# Numpy result
# Numpy result
a_n
=
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
a_n
=
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
])
cv
[::
c_step1
,
::
c_step2
])
f_nn
()
f_nn
()
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
...
@@ -1883,20 +1866,20 @@ class TestBlasStrides(TestCase):
...
@@ -1883,20 +1866,20 @@ class TestBlasStrides(TestCase):
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_nn
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b
,
c
))],
f_nn
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b
,
c
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_nt
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b
,
c_t
.
T
))],
f_nt
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b
,
c_t
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tn
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b_t
.
T
,
c
))],
f_tn
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b_t
.
T
,
c
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tt
=
theano
.
function
([],
[],
f_tt
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
))],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed pattern
# Try with all stride patterns, and all transposed pattern
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
4
):
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
4
):
for
step
in
(
1
,
2
):
for
step
in
(
1
,
2
):
b_step1
,
b_step2
,
c_step1
,
c_step2
=
(
s
*
step
b_step1
,
b_step2
,
c_step1
,
c_step2
=
(
s
*
step
for
s
in
step_signs
)
for
s
in
step_signs
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
...
@@ -1905,7 +1888,7 @@ class TestBlasStrides(TestCase):
...
@@ -1905,7 +1888,7 @@ class TestBlasStrides(TestCase):
# Numpy result
# Numpy result
a_n
=
l
*
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
a_n
=
l
*
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
])
cv
[::
c_step1
,
::
c_step2
])
f_nn
()
f_nn
()
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
...
@@ -1954,36 +1937,44 @@ class TestBlasStrides(TestCase):
...
@@ -1954,36 +1937,44 @@ class TestBlasStrides(TestCase):
bt_dev
=
b_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
bt_dev
=
b_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_nnn
=
theano
.
function
([],
[],
f_nnn
=
theano
.
function
(
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b
,
c
)))],
[],
[],
mode
=
self
.
mode
)
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b
,
c
)))],
f_nnt
=
theano
.
function
([],
[],
mode
=
self
.
mode
)
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b
,
c_t
.
T
)))],
f_nnt
=
theano
.
function
(
mode
=
self
.
mode
)
[],
[],
f_ntn
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b
,
c_t
.
T
)))],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b_t
.
T
,
c
)))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_ntn
=
theano
.
function
(
f_ntt
=
theano
.
function
([],
[],
[],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
)))],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b_t
.
T
,
c
)))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tnn
=
theano
.
function
([],
[],
f_ntt
=
theano
.
function
(
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b
,
c
)
.
T
))],
[],
[],
mode
=
self
.
mode
)
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
)))],
f_tnt
=
theano
.
function
([],
[],
mode
=
self
.
mode
)
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b
,
c_t
.
T
)
.
T
))],
f_tnn
=
theano
.
function
(
mode
=
self
.
mode
)
[],
[],
f_ttn
=
theano
.
function
([],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b
,
c
)
.
T
))],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b_t
.
T
,
c
)
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tnt
=
theano
.
function
(
f_ttt
=
theano
.
function
([],
[],
[],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
)
.
T
))],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b
,
c_t
.
T
)
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_ttn
=
theano
.
function
(
[],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b_t
.
T
,
c
)
.
T
))],
mode
=
self
.
mode
)
f_ttt
=
theano
.
function
(
[],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
)
.
T
))],
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed pattern
# Try with all stride patterns, and all transposed pattern
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
6
):
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
6
):
for
step
in
(
1
,
2
):
for
step
in
(
1
,
2
):
a_step1
,
a_step2
,
b_step1
,
b_step2
,
c_step1
,
c_step2
=
\
a_step1
,
a_step2
,
b_step1
,
b_step2
,
c_step1
,
c_step2
=
\
(
s
*
step
for
s
in
step_signs
)
(
s
*
step
for
s
in
step_signs
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
...
@@ -1991,12 +1982,12 @@ class TestBlasStrides(TestCase):
...
@@ -1991,12 +1982,12 @@ class TestBlasStrides(TestCase):
c_t
.
set_value
(
ct_dev
.
copy
()[::
c_step2
,
::
c_step1
],
borrow
=
True
)
c_t
.
set_value
(
ct_dev
.
copy
()[::
c_step2
,
::
c_step1
],
borrow
=
True
)
# Numpy results
# Numpy results
a_n
=
(
l
*
av
[::
a_step1
,
::
a_step2
]
a_n
=
(
l
*
av
[::
a_step1
,
::
a_step2
]
+
+
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
]))
cv
[::
c_step1
,
::
c_step2
]))
at_n
=
(
l
*
av
[::
a_step1
,
::
a_step2
]
.
T
at_n
=
(
l
*
av
[::
a_step1
,
::
a_step2
]
.
T
+
+
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
])
.
T
)
cv
[::
c_step1
,
::
c_step2
])
.
T
)
# a's value is updated, so we need to reinitialize it each time
# a's value is updated, so we need to reinitialize it each time
a
.
set_value
(
a_dev
.
copy
()[::
a_step1
,
::
a_step2
],
borrow
=
True
)
a
.
set_value
(
a_dev
.
copy
()[::
a_step1
,
::
a_step2
],
borrow
=
True
)
...
@@ -2016,22 +2007,22 @@ class TestBlasStrides(TestCase):
...
@@ -2016,22 +2007,22 @@ class TestBlasStrides(TestCase):
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
borrow
=
True
)
borrow
=
True
)
f_tnn
()
f_tnn
()
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
borrow
=
True
)
borrow
=
True
)
f_tnt
()
f_tnt
()
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
borrow
=
True
)
borrow
=
True
)
f_ttn
()
f_ttn
()
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
borrow
=
True
)
borrow
=
True
)
f_ttt
()
f_ttt
()
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
...
@@ -2066,28 +2057,28 @@ class TestBlasStrides(TestCase):
...
@@ -2066,28 +2057,28 @@ class TestBlasStrides(TestCase):
c_dev
=
c
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
c_dev
=
c
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_n
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
a
+
l
*
tensor
.
dot
(
b
,
c
)))],
f_n
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
a
+
l
*
tensor
.
dot
(
b
,
c
)))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_t
=
theano
.
function
([],
[],
f_t
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
a
+
l
*
tensor
.
dot
(
b_t
.
T
,
c
)))],
updates
=
[(
a
,
(
a
+
l
*
tensor
.
dot
(
b_t
.
T
,
c
)))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed pattern
# Try with all stride patterns, and all transposed pattern
for
step_signs
in
itertools_product
((
1
,
-
1
),
repeat
=
4
):
for
step_signs
in
itertools_product
((
1
,
-
1
),
repeat
=
4
):
for
step
in
(
1
,
2
):
for
step
in
(
1
,
2
):
a_step
,
b_step1
,
b_step2
,
c_step
=
(
s
*
step
a_step
,
b_step1
,
b_step2
,
c_step
=
(
s
*
step
for
s
in
step_signs
)
for
s
in
step_signs
)
a
.
set_value
(
a_dev
.
copy
()[::
a_step
],
borrow
=
True
)
a
.
set_value
(
a_dev
.
copy
()[::
a_step
],
borrow
=
True
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
borrow
=
True
)
b_t
.
set_value
(
transpose
(
b_dev
.
copy
())[::
b_step2
,
::
b_step1
],
b_t
.
set_value
(
transpose
(
b_dev
.
copy
())[::
b_step2
,
::
b_step1
],
borrow
=
True
)
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step
],
borrow
=
True
)
a_n
=
(
av
[::
a_step
]
a_n
=
(
av
[::
a_step
]
+
+
l
*
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
l
*
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step
]))
cv
[::
c_step
]))
f_n
()
f_n
()
assert
np
.
allclose
(
a
.
get_value
(),
a_n
),
(
a
.
get_value
(),
a_n
)
assert
np
.
allclose
(
a
.
get_value
(),
a_n
),
(
a
.
get_value
(),
a_n
)
...
@@ -2120,36 +2111,37 @@ class TestBlasStrides(TestCase):
...
@@ -2120,36 +2111,37 @@ class TestBlasStrides(TestCase):
b_dev
=
b
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
b_dev
=
b
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
c_dev
=
c
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
c_dev
=
c
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_n
=
theano
.
function
([],
[],
f_n
=
theano
.
function
(
updates
=
[(
a
,
(
a
+
l
*
tensor
.
outer
(
b
,
c
)))],
[],
[],
mode
=
self
.
mode
)
updates
=
[(
a
,
(
a
+
l
*
tensor
.
outer
(
b
,
c
)))],
mode
=
self
.
mode
)
f_t
=
theano
.
function
([],
[],
f_t
=
theano
.
function
(
updates
=
[(
a_t
,
(
a_t
+
l
*
tensor
.
outer
(
b
,
c
)
.
T
))],
[],
[],
mode
=
self
.
mode
)
updates
=
[(
a_t
,
(
a_t
+
l
*
tensor
.
outer
(
b
,
c
)
.
T
))],
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed patterns
# Try with all stride patterns, and all transposed patterns
for
step_signs
in
itertools_product
((
1
,
-
1
),
repeat
=
4
):
for
step_signs
in
itertools_product
((
1
,
-
1
),
repeat
=
4
):
for
step
in
(
1
,
2
):
for
step
in
(
1
,
2
):
a_step1
,
a_step2
,
b_step
,
c_step
=
(
s
*
step
a_step1
,
a_step2
,
b_step
,
c_step
=
(
s
*
step
for
s
in
step_signs
)
for
s
in
step_signs
)
a
.
set_value
(
a_dev
.
copy
()[::
a_step1
,
::
a_step2
],
borrow
=
True
)
a
.
set_value
(
a_dev
.
copy
()[::
a_step1
,
::
a_step2
],
borrow
=
True
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step1
,
::
a_step2
],
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step1
,
::
a_step2
],
borrow
=
True
)
borrow
=
True
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step
],
borrow
=
True
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step
],
borrow
=
True
)
f_n
()
f_n
()
n_n
=
(
av
[::
a_step1
,
::
a_step2
]
n_n
=
(
av
[::
a_step1
,
::
a_step2
]
+
+
l
*
np
.
outer
(
bv
[::
b_step
],
cv
[::
c_step
]))
l
*
np
.
outer
(
bv
[::
b_step
],
cv
[::
c_step
]))
assert
np
.
allclose
(
a
.
get_value
(),
n_n
),
(
a
.
get_value
(),
n_n
)
assert
np
.
allclose
(
a
.
get_value
(),
n_n
),
(
a
.
get_value
(),
n_n
)
f_t
()
f_t
()
n_t
=
(
av
.
T
[::
a_step1
,
::
a_step2
]
n_t
=
(
av
.
T
[::
a_step1
,
::
a_step2
]
+
+
l
*
np
.
outer
(
bv
[::
b_step
],
cv
[::
c_step
])
.
T
)
l
*
np
.
outer
(
bv
[::
b_step
],
cv
[::
c_step
])
.
T
)
assert
np
.
allclose
(
a_t
.
get_value
(),
n_t
),
\
assert
np
.
allclose
(
a_t
.
get_value
(),
n_t
),
(
a_t
.
get_value
(),
n_t
)
(
a_t
.
get_value
(),
n_t
)
def
test_ger_strides
(
self
):
def
test_ger_strides
(
self
):
self
.
cmp_ger
((
3
,
5
),
3
,
5
)
self
.
cmp_ger
((
3
,
5
),
3
,
5
)
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论