Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
a2f6c1c9
提交
a2f6c1c9
authored
7月 13, 2017
作者:
erakra
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fixing flake8 for test_blas.py
上级
c5cd87fa
隐藏空白字符变更
内嵌
并排
正在显示
1 个修改的文件
包含
221 行增加
和
229 行删除
+221
-229
test_blas.py
theano/tensor/tests/test_blas.py
+221
-229
没有找到文件。
theano/tensor/tests/test_blas.py
浏览文件 @
a2f6c1c9
...
...
@@ -5,7 +5,7 @@ from unittest import TestCase
import
numpy
as
np
from
numpy
import
(
arange
,
array
,
common_type
,
complex64
,
complex128
,
float32
,
float64
,
newaxis
,
shape
,
transpose
,
zeros
)
float64
,
newaxis
,
shape
,
transpose
,
zeros
)
from
numpy.testing
import
assert_array_almost_equal
from
six.moves
import
xrange
...
...
@@ -22,7 +22,7 @@ from theano.tensor.blas import (_dot22, _dot22scalar, res_is_a, _as_scalar,
InconsistencyError
,
Ger
,
ger
,
ger_destructive
)
from
theano.tests
import
unittest_tools
from
.test_basic
import
(
as_tensor_variable
,
inplace_func
,
compile
,
inplace
)
compile
,
inplace
)
import
theano.tensor.blas_scipy
from
theano.tests.unittest_tools
import
attr
...
...
@@ -81,7 +81,7 @@ class t_gemm(TestCase):
f
=
inplace_func
([
tz
,
ta
,
tx
,
ty
,
tb
],
gemm_inplace
(
tz
,
ta
,
tx
,
ty
,
tb
),
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
new_z
=
f
(
z
,
a
,
x
,
y
,
b
)
f
(
z
,
a
,
x
,
y
,
b
)
z_after
=
self
.
_gemm
(
z_orig
,
a
,
x
,
y
,
b
)
# print z_orig, z_after, z, type(z_orig), type(z_after), type(z)
...
...
@@ -96,8 +96,8 @@ class t_gemm(TestCase):
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'c|py'
)
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'py'
)
if
(
not
dtype
.
startswith
(
"complex"
)
and
theano
.
config
.
cxx
):
if
(
not
dtype
.
startswith
(
"complex"
)
and
theano
.
config
.
cxx
):
# If theano.config.blas.ldflags is empty, Theano will use
# a NumPy C implementation of [sd]gemm_.
cmp_linker
(
copy
(
z
),
a
,
x
,
y
,
b
,
'c'
)
...
...
@@ -105,7 +105,7 @@ class t_gemm(TestCase):
def
test0a
(
self
):
Gemm
.
debug
=
True
try
:
g
=
g
emm_no_inplace
([
1.
],
1.
,
[
1.
],
[
1.
],
1.
)
gemm_no_inplace
([
1.
],
1.
,
[
1.
],
[
1.
],
1.
)
except
TypeError
as
e
:
if
exc_message
(
e
)
is
Gemm
.
E_rank
:
return
...
...
@@ -171,7 +171,6 @@ class t_gemm(TestCase):
def
test_factorised_scalar
(
self
):
a
=
T
.
matrix
()
b
=
T
.
matrix
()
c
=
T
.
matrix
()
s
=
theano
.
shared
(
np
.
zeros
((
5
,
5
))
.
astype
(
config
.
floatX
))
lr1
=
T
.
constant
(
0.01
)
.
astype
(
config
.
floatX
)
...
...
@@ -180,9 +179,9 @@ class t_gemm(TestCase):
# test constant merge with gemm
f
=
theano
.
function
([
a
,
b
],
updates
=
[(
s
,
lr1
*
T
.
dot
(
a
,
b
)
+
l2_reg
*
lr2
*
s
)],
l2_reg
*
lr2
*
s
)],
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
#[Gemm{inplace}(<TensorType(float64, matrix)>, 0.01,
#
[Gemm{inplace}(<TensorType(float64, matrix)>, 0.01,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# 2e-06)]
assert
len
(
f
)
==
1
...
...
@@ -192,7 +191,7 @@ class t_gemm(TestCase):
f
=
theano
.
function
([
a
,
b
],
updates
=
[(
s
,
lr1
*
(
T
.
dot
(
a
,
b
)
-
l2_reg
*
s
))],
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
#[Gemm{inplace}(<TensorType(float64, matrix)>, 0.01,
#
[Gemm{inplace}(<TensorType(float64, matrix)>, 0.01,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# -2e-06)]
assert
len
(
f
)
==
1
...
...
@@ -202,7 +201,7 @@ class t_gemm(TestCase):
f
=
theano
.
function
([
a
,
b
],
updates
=
[(
s
,
s
-
lr1
*
(
s
*
.
0002
+
T
.
dot
(
a
,
b
)))],
mode
=
mode_not_fast_compile
)
.
maker
.
fgraph
.
toposort
()
#[Gemm{inplace}(<TensorType(float64, matrix)>, -0.01,
#
[Gemm{inplace}(<TensorType(float64, matrix)>, -0.01,
# <TensorType(float64, matrix)>, <TensorType(float64, matrix)>,
# 0.999998)]
assert
len
(
f
)
==
1
...
...
@@ -270,14 +269,14 @@ class t_gemm(TestCase):
def
t
(
z
,
x
,
y
,
a
=
1.0
,
b
=
0.0
,
l
=
'c|py'
,
dt
=
'float64'
):
z
,
a
,
x
,
y
,
b
=
[
theano
.
_asarray
(
p
,
dtype
=
dt
)
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
z_orig
=
z
.
copy
()
#
z_orig = z.copy()
z_after
=
self
.
_gemm
(
z
,
a
,
x
,
y
,
b
)
tz
,
ta
,
tx
,
ty
,
tb
=
[
shared
(
p
)
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
# f = inplace_func([tz,ta,tx,ty,tb], gemm_inplace(tz,ta,tx,ty,tb),
# mode = compile.Mode(optimizer = None, linker=l))
#f(z, a, x, y, b)
#
f(z, a, x, y, b)
f
=
inplace_func
([],
gemm_inplace
(
tz
,
ta
,
tx
,
ty
,
tb
),
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
f
()
...
...
@@ -339,9 +338,9 @@ class t_gemm(TestCase):
tz
,
ta
,
tx
,
ty
,
tb
=
[
shared
(
p
)
for
p
in
(
z
,
a
,
x
,
y
,
b
)]
for
i
in
xrange
(
3
):
f_i
=
inplace_func
([],
gemm_inplace
(
tz
[:,
:,
i
],
ta
,
tx
[:,
:,
i
],
ty
[:,
:,
i
],
tb
),
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
gemm_inplace
(
tz
[:,
:,
i
],
ta
,
tx
[:,
:,
i
],
ty
[:,
:,
i
],
tb
),
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
for
j
in
xrange
(
3
):
# tz will not _always_ be overwritten,
# and adding update={...} in the call to function()
...
...
@@ -355,9 +354,10 @@ class t_gemm(TestCase):
tz_i
=
gemm_no_inplace
(
tz
[:,
:,
i
],
ta
,
tx
[
:,
:,
i
],
ty
[:,
:,
i
],
tb
)
g_i
=
theano
.
function
([],
tz_i
,
updates
=
[(
tz
,
T
.
set_subtensor
(
tz
[:,
:,
i
],
tz_i
))],
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
g_i
=
theano
.
function
(
[],
tz_i
,
updates
=
[(
tz
,
T
.
set_subtensor
(
tz
[:,
:,
i
],
tz_i
))],
mode
=
compile
.
Mode
(
optimizer
=
None
,
linker
=
l
))
for
j
in
xrange
(
3
):
g_i
()
unittest_tools
.
assert_allclose
(
z_after
[:,
:,
i
],
...
...
@@ -404,8 +404,8 @@ class t_as_scalar(TestCase):
def
test1
(
self
):
"""Test that it fails on nonscalar constants"""
a
=
T
.
constant
(
np
.
ones
(
5
))
self
.
assertTrue
(
None
==
_as_scalar
(
a
)
)
self
.
assertTrue
(
None
==
_as_scalar
(
T
.
DimShuffle
([
False
],
[
0
,
'x'
])(
a
))
)
self
.
assertTrue
(
_as_scalar
(
a
)
is
None
)
self
.
assertTrue
(
_as_scalar
(
T
.
DimShuffle
([
False
],
[
0
,
'x'
])(
a
))
is
None
)
def
test2
(
self
):
"""Test that it works on scalar variables"""
...
...
@@ -420,9 +420,9 @@ class t_as_scalar(TestCase):
def
test3
(
self
):
"""Test that it fails on nonscalar variables"""
a
=
T
.
matrix
()
self
.
assertTrue
(
None
==
_as_scalar
(
a
)
)
self
.
assertTrue
(
None
==
_as_scalar
(
T
.
DimShuffle
([
False
,
False
],
[
0
,
'x'
,
1
])(
a
))
)
self
.
assertTrue
(
_as_scalar
(
a
)
is
None
)
self
.
assertTrue
(
_as_scalar
(
T
.
DimShuffle
([
False
,
False
],
[
0
,
'x'
,
1
])(
a
))
is
None
)
class
T_real_matrix
(
TestCase
):
...
...
@@ -458,10 +458,10 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
max_graphlen
=
0
,
expected_nb_gemm
=
1
):
try
:
f
=
inplace_func
(
[
In
(
ii
,
mutable
=
True
,
allow_downcast
=
True
)
for
ii
in
i
],
o
,
mode
=
'FAST_RUN'
,
on_unused_input
=
'ignore'
)
[
In
(
ii
,
mutable
=
True
,
allow_downcast
=
True
)
for
ii
in
i
],
o
,
mode
=
'FAST_RUN'
,
on_unused_input
=
'ignore'
)
nb_gemm
=
0
for
node
in
f
.
maker
.
fgraph
.
apply_nodes
:
if
isinstance
(
node
.
op
,
T
.
Dot
):
...
...
@@ -472,7 +472,7 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
nb_gemm
+=
1
assert
nb_gemm
==
expected_nb_gemm
,
(
nb_gemm
,
expected_nb_gemm
)
g
=
inplace_func
(
i
,
o
,
mode
=
compile
.
Mode
(
linker
=
'py'
,
optimizer
=
None
),
allow_input_downcast
=
True
,
on_unused_input
=
'ignore'
)
allow_input_downcast
=
True
,
on_unused_input
=
'ignore'
)
for
node
in
g
.
maker
.
fgraph
.
apply_nodes
:
if
node
.
op
==
gemm_inplace
:
raise
Exception
(
'gemm_inplace in original graph'
)
...
...
@@ -492,7 +492,7 @@ def just_gemm(i, o, ishapes=[(4, 3), (3, 5), (4, 5), (), ()],
eps
=
1.0e-8
if
config
.
floatX
==
'float32'
:
eps
=
1.0e-6
if
max_abs_err
>
eps
:
if
max_abs_err
>
eps
:
raise
Failure
(
'GEMM is computing the wrong output. max_rel_err ='
,
max_abs_err
)
except
Failure
:
...
...
@@ -516,7 +516,7 @@ def test_gemm_opt0():
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
.
T
-
a
*
T
.
dot
(
Y
.
T
,
X
.
T
)])
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
.
T
+
a
*
b
*
T
.
dot
(
X
,
Y
)
.
T
])
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[
b
*
Z
+
a
*
T
.
dot
(
X
,
Y
)
.
T
],
ishapes
=
[(
5
,
3
),
(
3
,
4
),
(
4
,
5
),
(),
()])
ishapes
=
[(
5
,
3
),
(
3
,
4
),
(
4
,
5
),
(),
()])
# with N multiplications instead of just one
just_gemm
([
X
,
Y
,
Z
,
a
,
b
],
[(
b
*
b
)
*
Z
*
a
+
(
a
*
a
)
*
T
.
dot
(
X
,
Y
)
*
b
])
...
...
@@ -541,32 +541,32 @@ def test_gemm_opt_double_gemm():
ishapes
=
[(
4
,
3
),
(
3
,
5
),
(
4
,
5
),
(),
(),
(
5
,
9
),
(
9
,
4
),
()]
i
=
[
X
,
Y
,
Z
,
a
,
b
,
R
,
S
,
c
]
o
=
[(
a
*
T
.
dot
(
X
,
Y
)
+
gemm_inplace
(
Z
,
b
,
S
.
T
,
R
.
T
,
T
.
constant
(
1.0
)
.
astype
(
config
.
floatX
)))]
o
=
[(
a
*
T
.
dot
(
X
,
Y
)
+
gemm_inplace
(
Z
,
b
,
S
.
T
,
R
.
T
,
T
.
constant
(
1.0
)
.
astype
(
config
.
floatX
)))]
try
:
f
=
inplace_func
([
In
(
ii
,
mutable
=
True
)
for
ii
in
i
],
o
,
mode
=
'FAST_RUN'
,
on_unused_input
=
'ignore'
)
mode
=
'FAST_RUN'
,
on_unused_input
=
'ignore'
)
for
node
in
f
.
maker
.
fgraph
.
apply_nodes
:
if
isinstance
(
node
.
op
,
T
.
Dot
):
raise
Failure
(
'dot in graph'
)
if
node
.
op
==
_dot22
:
raise
Failure
(
'_dot22 in graph'
)
g
=
inplace_func
(
i
,
o
,
mode
=
compile
.
Mode
(
linker
=
'py'
,
optimizer
=
None
),
on_unused_input
=
'ignore'
)
on_unused_input
=
'ignore'
)
# for node in g.maker.fgraph.apply_nodes:
# if node.op == gemm_inplace: raise Failure('gemm_inplace in graph')
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
(
234
))
r0
=
f
(
*
[
np
.
asarray
(
rng
.
randn
(
*
sh
),
config
.
floatX
)
for
sh
in
ishapes
])
for
sh
in
ishapes
])
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
(
234
))
r1
=
g
(
*
[
np
.
asarray
(
rng
.
randn
(
*
sh
),
config
.
floatX
)
for
sh
in
ishapes
])
for
sh
in
ishapes
])
max_abs_err
=
np
.
max
(
np
.
abs
(
r0
[
0
]
-
r1
[
0
]))
eps
=
1.0e-8
if
config
.
floatX
==
'float32'
:
eps
=
1.0e-6
if
max_abs_err
>
eps
:
if
max_abs_err
>
eps
:
raise
Failure
(
'GEMM is computing the wrong output. max_rel_err ='
,
max_abs_err
)
...
...
@@ -579,8 +579,7 @@ def test_gemm_opt_double_gemm():
def
test_gemm_canonicalize
():
X
,
Y
,
Z
,
a
,
b
=
T
.
matrix
(
'X'
),
T
.
matrix
(
'Y'
),
T
.
matrix
(
'Z'
),
T
.
scalar
(
'a'
),
T
.
scalar
(
'b'
)
R
,
S
,
U
,
c
,
d
=
T
.
matrix
(
'R'
),
T
.
matrix
(
'S'
),
T
.
matrix
(
'U'
),
T
.
scalar
(
'c'
),
T
.
scalar
(
'd'
)
c
,
d
=
T
.
scalar
(
'c'
),
T
.
scalar
(
'd'
)
u
=
T
.
row
(
'u'
)
v
=
T
.
vector
(
'v'
)
w
=
T
.
col
(
'w'
)
...
...
@@ -631,10 +630,7 @@ def test_gemm_canonicalize():
def
test_gemm_factor
():
X
,
Y
,
Z
,
a
,
b
=
T
.
matrix
(
'X'
),
T
.
matrix
(
'Y'
),
T
.
matrix
(
'Z'
),
T
.
scalar
(
'a'
),
T
.
scalar
(
'b'
)
R
,
S
,
U
,
c
,
d
=
T
.
matrix
(
'R'
),
T
.
matrix
(
'S'
),
T
.
matrix
(
'U'
),
T
.
scalar
(
'c'
),
T
.
scalar
(
'd'
)
X
,
Y
=
T
.
matrix
(
'X'
),
T
.
matrix
(
'Y'
)
assert
[(
1.0
,
X
),
(
1.0
,
Y
)]
==
_factor_canonicalized
([(
1.0
,
X
),
(
1.0
,
Y
)])
assert
[(
2.0
,
X
)]
==
_factor_canonicalized
([(
1.0
,
X
),
(
1.0
,
X
)])
...
...
@@ -653,7 +649,7 @@ def test_upcasting_scalar_nogemm():
f
=
theano
.
function
([
w
,
v
,
t
,
alpha
],
rval
)
t
=
f
.
maker
.
fgraph
.
toposort
()
assert
np
.
sum
([
isinstance
(
n
.
op
,
Gemm
)
for
n
in
t
])
==
0
#theano.printing.debugprint(f, print_type=True)
#
theano.printing.debugprint(f, print_type=True)
v
=
T
.
fmatrix
(
'v'
)
w
=
T
.
fmatrix
(
'w'
)
...
...
@@ -670,7 +666,7 @@ def test_upcasting_scalar_nogemm():
t
=
f
.
maker
.
fgraph
.
toposort
()
assert
np
.
sum
([
isinstance
(
n
.
op
,
Gemm
)
for
n
in
t
])
==
0
#theano.printing.debugprint(f, print_type=True)
#
theano.printing.debugprint(f, print_type=True)
def
test_gemm_nested
():
...
...
@@ -680,22 +676,22 @@ def test_gemm_nested():
'c'
),
T
.
scalar
(
'd'
)
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
1
)
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
1
)
# print "---------------------"
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
+
c
*
Z
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
1
)
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
+
c
*
Z
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
1
)
# print "---------------------"
just_gemm
([
X
,
Y
,
Z
,
R
,
S
,
U
,
a
,
b
,
c
,
d
],
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
+
c
*
U
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
3
)
[
a
*
Z
-
b
*
(
c
*
T
.
dot
(
X
,
Y
)
+
d
*
Z
+
c
*
U
)],
ishapes
=
[(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(
2
,
3
),
(
3
,
4
),
(
2
,
4
),
(),
(),
(),
()],
max_graphlen
=
3
)
def
test_gemm_opt_wishlist
():
...
...
@@ -720,7 +716,7 @@ def test_gemm_with_vector():
def
my_just_gemm
(
o
):
i
=
[
X
,
Y
,
Z
,
a
,
b
,
v
]
ishapes
=
[(
4
,
3
),
(
3
,
5
),
(
4
,
5
),
(),
(),
(
5
,
)]
rval
=
just_gemm
(
i
,
o
,
ishapes
=
ishapes
)
just_gemm
(
i
,
o
,
ishapes
=
ishapes
)
my_just_gemm
([
v
+
T
.
dot
(
X
,
Y
)
*
a
+
Z
*
b
])
my_just_gemm
([
v
+
a
*
T
.
dot
(
X
,
Y
)
+
b
*
Z
])
...
...
@@ -741,7 +737,7 @@ def test_gemm_with_vector():
def
test_gemm_opt_vector_stuff
():
X
,
Y
,
Z
,
a
,
b
=
T
.
matrix
(),
T
.
matrix
(),
T
.
matrix
(),
T
.
scalar
(),
T
.
scalar
()
X
,
Y
,
a
=
T
.
matrix
(),
T
.
matrix
(),
T
.
scalar
()
u
,
v
=
T
.
vector
(),
T
.
vector
()
f
=
inplace_func
([
a
,
u
,
v
],
a
+
T
.
dot
(
u
,
v
),
mode
=
'FAST_RUN'
)
...
...
@@ -771,8 +767,6 @@ def test_gemm_unrolled():
H
=
sharedX
(
np
.
zeros
((
batch_size
,
rep_size
)),
name
=
'H'
)
G
=
sharedX
(
np
.
zeros
((
batch_size
,
rep_size
)),
name
=
'G'
)
init_V
=
sharedX
(
rng
.
uniform
(
0
,
1
,
(
batch_size
,
rep_size
)),
name
=
'init_V'
)
init_H
=
sharedX
(
rng
.
uniform
(
0
,
1
,
(
batch_size
,
rep_size
)),
name
=
'init_H'
)
cur_V
=
V
cur_H
=
H
...
...
@@ -787,7 +781,7 @@ def test_gemm_unrolled():
cur_H
=
update_H
(
cur_V
)
unrolled_theano
=
theano
.
function
([],
updates
=
[(
V
,
cur_V
),
(
H
,
cur_H
)],
name
=
'unrolled_theano'
)
name
=
'unrolled_theano'
)
nb_dot
=
sum
([
1
for
node
in
unrolled_theano
.
maker
.
fgraph
.
toposort
()
if
isinstance
(
node
.
op
,
(
theano
.
tensor
.
Dot
,
theano
.
tensor
.
blas
.
Dot22
,
...
...
@@ -807,7 +801,7 @@ def test_inplace0():
R
,
S
,
c
=
T
.
matrix
(
'R'
),
T
.
matrix
(
'S'
),
T
.
scalar
(
'c'
)
f
=
inplace_func
([
Z
,
b
,
R
,
S
],
[
Z
*
(
Z
+
b
*
T
.
dot
(
R
,
S
)
.
T
)],
mode
=
'FAST_RUN'
)
[
Z
*
(
Z
+
b
*
T
.
dot
(
R
,
S
)
.
T
)],
mode
=
'FAST_RUN'
)
if
(
gemm_inplace
in
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]):
print
(
pp
(
f
.
maker
.
fgraph
.
outputs
[
0
]))
raise
Failure
(
'gemm_inplace in graph'
)
...
...
@@ -815,9 +809,9 @@ def test_inplace0():
# gemm_inplace should be inserted here, to work in-place on Z*c
f
=
inplace_func
([
X
,
Y
,
Z
,
a
,
b
,
R
,
S
,
c
],
[
Z
*
(
c
*
Z
+
a
*
T
.
dot
(
X
,
Y
)
+
b
*
T
.
dot
(
R
,
S
)
.
T
)],
mode
=
'FAST_RUN'
)
if
(
not
gemm_inplace
in
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]):
[
Z
*
(
c
*
Z
+
a
*
T
.
dot
(
X
,
Y
)
+
b
*
T
.
dot
(
R
,
S
)
.
T
)],
mode
=
'FAST_RUN'
)
if
(
gemm_inplace
not
in
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]):
theano
.
printing
.
debugprint
(
f
)
raise
Failure
(
'no gemm_inplace in graph'
)
...
...
@@ -826,7 +820,7 @@ def test_inplace1():
X
,
Y
,
Z
,
a
,
b
=
XYZab
()
# with > 2 terms in the overall addition
f
=
inplace_func
([
X
,
Y
,
Z
],
[
Z
+
Z
+
T
.
dot
(
X
,
Y
)],
mode
=
'FAST_RUN'
)
[
Z
+
Z
+
T
.
dot
(
X
,
Y
)],
mode
=
'FAST_RUN'
)
# theano.printing.debugprint(f)
# it doesn't work inplace because we didn't mark Z as mutable input
assert
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
apply_nodes
]
==
[
gemm_no_inplace
]
...
...
@@ -866,7 +860,7 @@ def test_dot22scalar():
# TODO: exclude other optimizations in BlasOpt?
# m = theano.compile.get_default_mode().including('local_dot_to_dot22',
# 'local_dot22_to_dot22scalar','specialize')
#m = theano.compile.get_default_mode().including('BlasOpt', 'specialize')
#
m = theano.compile.get_default_mode().including('BlasOpt', 'specialize')
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
for
dtype1
in
[
'complex64'
,
'complex128'
]:
a
=
T
.
matrix
(
'a'
,
dtype
=
dtype1
)
...
...
@@ -1087,7 +1081,7 @@ def test_dot_w_self():
class
TestGemv
(
TestCase
,
unittest_tools
.
TestOptimizationMixin
):
def
test_dot_vv
(
self
):
''' Currently we generate a gemv for that case'''
# Currently we generate a gemv for that case
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
w
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
...
...
@@ -1101,11 +1095,10 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
w
.
get_value
()))
def
test_dot_vm
(
self
):
''' Test vector dot matrix '''
# Test vector dot matrix
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,
3
)),
dtype
=
'float32'
))
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,
3
)),
dtype
=
'float32'
))
f
=
theano
.
function
([],
theano
.
dot
(
v
,
m
),
mode
=
mode_blas_opt
)
# Assert that the dot was optimized somehow
...
...
@@ -1115,17 +1108,14 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
# Assert they produce the same output
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
m
.
get_value
()))
# Assert it works when m has no contiguous dimension
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
assert
np
.
allclose
(
f
(),
np
.
dot
(
v
.
get_value
(),
m
.
get_value
()))
def
test_dot_mv
(
self
):
''' Test matrix dot vector '''
# Test matrix dot vector
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
3
,
2
)),
dtype
=
'float32'
))
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
3
,
2
)),
dtype
=
'float32'
))
f
=
theano
.
function
([],
theano
.
dot
(
m
,
v
),
mode
=
mode_blas_opt
)
# Assert that the dot was optimized somehow
...
...
@@ -1135,31 +1125,27 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
# Assert they produce the same output
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v
.
get_value
()))
# Assert it works when m has no contiguous dimension
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v
.
get_value
()))
@staticmethod
def
t_gemv1
(
m_shp
):
''' test vector2+dot(matrix,vector1) '''
# test vector2+dot(matrix,vector1)
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
m_shp
[
1
],)
),
dtype
=
'float32'
))
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
m_shp
[
1
],)
),
dtype
=
'float32'
))
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
m_shp
[
0
],)),
dtype
=
'float32'
)
v2
=
theano
.
shared
(
v2_orig
)
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
m_shp
),
dtype
=
'float32'
))
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
m_shp
),
dtype
=
'float32'
))
f
=
theano
.
function
([],
v2
+
theano
.
dot
(
m
,
v1
),
mode
=
mode_blas_opt
)
# Assert they produce the same output
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
topo
=
f
.
maker
.
fgraph
.
toposort
()
assert
len
(
topo
)
==
1
assert
isinstance
(
topo
[
0
]
.
op
,
Gemv
)
assert
topo
[
0
]
.
op
.
inplace
==
False
assert
topo
[
0
]
.
op
.
inplace
is
False
# test the inplace version
g
=
theano
.
function
([],
[],
updates
=
[(
v2
,
v2
+
theano
.
dot
(
m
,
v1
))],
...
...
@@ -1167,24 +1153,23 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
# Assert they produce the same output
g
()
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
topo
=
g
.
maker
.
fgraph
.
toposort
()
assert
len
(
topo
)
==
1
assert
isinstance
(
topo
[
0
]
.
op
,
Gemv
)
if
config
.
mode
!=
'FAST_COMPILE'
:
assert
topo
[
0
]
.
op
.
inplace
==
True
assert
topo
[
0
]
.
op
.
inplace
is
True
# Do the same tests with a matrix with strides in both dimensions
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
v2
.
set_value
(
v2_orig
)
assert
np
.
allclose
(
f
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
g
()
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
v2_orig
)
@attr
(
'slow'
)
def
test_gemv1
(
self
):
...
...
@@ -1194,23 +1179,24 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
self
.
t_gemv1
((
0
,
0
))
def
test_gemv2
(
self
):
''' test vector2+dot(vector1,matrix) '''
# test vector2+dot(vector1,matrix)
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
dtype
=
'float32'
))
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
3
,)),
dtype
=
'float32'
)
v2
=
theano
.
shared
(
v2_orig
)
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,
3
)),
dtype
=
'float32'
))
dtype
=
'float32'
))
f
=
theano
.
function
([],
v2
+
theano
.
dot
(
v1
,
m
),
mode
=
mode_blas_opt
)
# Assert they produce the same output
assert
np
.
allclose
(
f
(),
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2
.
get_value
())
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2
.
get_value
())
topo
=
f
.
maker
.
fgraph
.
toposort
()
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
assert
topo
[
-
1
]
.
op
.
inplace
==
False
assert
topo
[
-
1
]
.
op
.
inplace
is
False
# test the inplace version
g
=
theano
.
function
([],
[],
updates
=
[(
v2
,
v2
+
theano
.
dot
(
v1
,
m
))],
...
...
@@ -1219,32 +1205,32 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
# Assert they produce the same output
g
()
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2_orig
)
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2_orig
)
topo
=
g
.
maker
.
fgraph
.
toposort
()
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
if
config
.
mode
!=
'FAST_COMPILE'
:
assert
topo
[
-
1
]
.
op
.
inplace
==
True
assert
topo
[
-
1
]
.
op
.
inplace
is
True
# Do the same tests with a matrix with strides in both dimensions
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
m
.
set_value
(
m
.
get_value
(
borrow
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
v2
.
set_value
(
v2_orig
)
assert
np
.
allclose
(
f
(),
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2
.
get_value
())
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2
.
get_value
())
g
()
assert
np
.
allclose
(
v2
.
get_value
(),
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2_orig
)
np
.
dot
(
v1
.
get_value
(),
m
.
get_value
())
+
v2_orig
)
def
test_gemv_broadcast
(
self
):
''' test gemv with some broadcasted input '''
# test gemv with some broadcasted input
rng
=
np
.
random
.
RandomState
(
unittest_tools
.
fetch_seed
())
v1
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
2
,)),
dtype
=
'float32'
))
dtype
=
'float32'
))
v2_orig
=
np
.
array
(
rng
.
uniform
(
size
=
(
1
,)),
dtype
=
'float32'
)
v2
=
theano
.
shared
(
v2_orig
)
m
=
theano
.
shared
(
np
.
array
(
rng
.
uniform
(
size
=
(
1
,
2
)),
dtype
=
'float32'
),
dtype
=
'float32'
),
broadcastable
=
(
True
,
False
))
o
=
theano
.
dot
(
m
,
v1
)
f
=
theano
.
function
([],
o
+
v2
,
mode
=
mode_blas_opt
)
...
...
@@ -1261,7 +1247,7 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
f
=
theano
.
function
([],
o
,
mode
=
mode_blas_opt
)
assert
np
.
allclose
(
f
(),
0.5
*
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
0.25
*
v2
.
get_value
())
0.5
*
np
.
dot
(
m
.
get_value
(),
v1
.
get_value
())
+
0.25
*
v2
.
get_value
())
topo
=
f
.
maker
.
fgraph
.
toposort
()
assert
sum
(
isinstance
(
node
.
op
,
Gemv
)
for
node
in
topo
)
==
1
...
...
@@ -1269,9 +1255,9 @@ class TestGemv(TestCase, unittest_tools.TestOptimizationMixin):
A
=
T
.
matrix
(
'A'
)
x
,
y
=
T
.
vectors
(
'x'
,
'y'
)
alpha
=
theano
.
shared
(
theano
.
_asarray
(
1.0
,
dtype
=
config
.
floatX
),
name
=
'alpha'
)
name
=
'alpha'
)
beta
=
theano
.
shared
(
theano
.
_asarray
(
1.0
,
dtype
=
config
.
floatX
),
name
=
'beta'
)
name
=
'beta'
)
z
=
beta
*
y
+
alpha
*
T
.
dot
(
A
,
x
)
f
=
theano
.
function
([
A
,
x
,
y
],
z
)
...
...
@@ -1335,9 +1321,9 @@ class BaseGemv(object):
def
test_simple
(
self
):
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
value
)
for
value
in
self
.
get_data
()]
desired_oy
=
alpha
.
get_value
()
*
matrixmultiply
(
a
.
get_value
(),
x
.
get_value
())
+
beta
.
get_value
()
*
y
.
get_value
()
for
value
in
self
.
get_data
()]
desired_oy
=
alpha
.
get_value
()
*
matrixmultiply
(
a
.
get_value
(),
x
.
get_value
())
+
beta
.
get_value
()
*
y
.
get_value
()
oy
=
alpha
*
T
.
dot
(
a
,
x
)
+
beta
*
y
...
...
@@ -1406,8 +1392,8 @@ class BaseGemv(object):
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
desired_oy
=
alpha_v
*
matrixmultiply
(
transpose
(
a_v
),
x_v
[::
2
])
+
beta_v
*
y_v
desired_oy
=
alpha_v
*
matrixmultiply
(
transpose
(
a_v
),
x_v
[::
2
])
+
beta_v
*
y_v
oy
=
alpha
*
T
.
dot
(
a
.
T
,
x
[::
2
])
+
beta
*
y
...
...
@@ -1456,10 +1442,9 @@ class BaseGemv(object):
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
a_v
=
a_v
[::
-
1
,
::
-
1
]
a
.
set_value
(
a
.
get_value
(
borrow
=
True
,
return_internal_type
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
a
.
set_value
(
a
.
get_value
(
borrow
=
True
,
return_internal_type
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
desired_oy
=
alpha_v
*
matrixmultiply
(
a_v
,
x_v
)
+
beta_v
*
y_v
...
...
@@ -1477,10 +1462,9 @@ class BaseGemv(object):
alpha_v
,
beta_v
,
a_v
,
x_v
,
y_v
=
vs
alpha
,
beta
,
a
,
x
,
y
=
[
self
.
shared
(
v
)
for
v
in
vs
]
a_v
=
a_v
[::
-
1
,
::
-
1
]
a
.
set_value
(
a
.
get_value
(
borrow
=
True
,
return_internal_type
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
a
.
set_value
(
a
.
get_value
(
borrow
=
True
,
return_internal_type
=
True
)[::
-
1
,
::
-
1
],
borrow
=
True
)
desired_oy
=
alpha_v
*
matrixmultiply
(
transpose
(
a_v
),
x_v
)
+
beta_v
*
y_v
...
...
@@ -1517,7 +1501,7 @@ class BaseGemv(object):
# done inplace on a temporarily allocated-buffer, which is
# then scaled by alpha and to t with a fused elemwise.
n_gemvs
=
0
#theano.printing.debugprint(f, print_type=True)
#
theano.printing.debugprint(f, print_type=True)
for
node
in
f
.
maker
.
fgraph
.
toposort
():
if
node
.
op
==
self
.
gemv_inplace
:
n_gemvs
+=
1
...
...
@@ -1580,39 +1564,42 @@ class TestGer_make_node(TestCase):
def
test_works_on_all_valid_dtypes
(
self
):
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
.
type
)
def
test_fails_on_invalid_dtypes
(
self
):
self
.
assertRaises
(
TypeError
,
ger
,
T
.
imatrix
(),
T
.
iscalar
(),
T
.
ivector
(),
T
.
ivector
())
ger
,
T
.
imatrix
(),
T
.
iscalar
(),
T
.
ivector
(),
T
.
ivector
())
def
test_fails_for_nonscalar_alpha
(
self
):
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fm
,
self
.
fm
,
self
.
fv
,
self
.
fv_2
)
ger
,
self
.
fm
,
self
.
fm
,
self
.
fv
,
self
.
fv_2
)
# boundary case - fv1 has the right dtype and could be dimshuffled to a
# scalar, but that's not make_node's job.
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fm
,
self
.
fv1
,
self
.
fv
,
self
.
fv_2
)
ger
,
self
.
fm
,
self
.
fv1
,
self
.
fv
,
self
.
fv_2
)
# actually doing the aforementioned dimshuffle makes it work
self
.
assertEqual
(
self
.
fm
.
type
,
ger
(
self
.
fm
,
self
.
fv1
.
dimshuffle
(),
self
.
fv
,
self
.
fv_2
)
.
type
)
ger
(
self
.
fm
,
self
.
fv1
.
dimshuffle
(),
self
.
fv
,
self
.
fv_2
)
.
type
)
def
test_fails_for_nonmatrix_A
(
self
):
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fv
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
ger
,
self
.
fv
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
def
test_fails_for_nonvector_x_or_y
(
self
):
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fm
,
self
.
fa
,
self
.
fv
.
dimshuffle
(
'x'
,
0
),
self
.
fv_2
)
ger
,
self
.
fm
,
self
.
fa
,
self
.
fv
.
dimshuffle
(
'x'
,
0
),
self
.
fv_2
)
self
.
assertRaises
(
TypeError
,
ger
,
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
.
dimshuffle
(
'x'
,
0
))
ger
,
self
.
fm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
.
dimshuffle
(
'x'
,
0
))
def
test_fails_for_mixed_dtypes
(
self
):
self
.
assertRaises
(
TypeError
,
ger
,
self
.
dm
,
self
.
fa
,
self
.
fv
,
self
.
fv_2
)
...
...
@@ -1657,30 +1644,26 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
def
test_b_0_triggers_ger
(
self
):
""" test local_gemm_to_ger opt"""
assert
T
.
blas
.
local_gemm_to_ger
.
transform
(
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
0
))
.
owner
)
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
0
))
.
owner
)
def
test_b_1_triggers_ger
(
self
):
""" test local_gemm_to_ger opt"""
assert
T
.
blas
.
local_gemm_to_ger
.
transform
(
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
1
))
.
owner
)
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
1
))
.
owner
)
def
test_b_other_does_not_triggers_ger
(
self
):
""" test local_gemm_to_ger opt"""
assert
not
T
.
blas
.
local_gemm_to_ger
.
transform
(
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
1.5
))
.
owner
)
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
b
(
1.5
))
.
owner
)
def
test_b_nonconst_does_not_triggers_ger
(
self
):
""" test local_gemm_to_ger opt"""
assert
not
T
.
blas
.
local_gemm_to_ger
.
transform
(
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
a
)
.
owner
)
gemm_no_inplace
(
self
.
A
,
self
.
a
,
self
.
x
.
dimshuffle
(
0
,
'x'
),
self
.
y
.
dimshuffle
(
'x'
,
0
),
self
.
a
)
.
owner
)
def
test_outer
(
self
):
f
=
self
.
function
([
self
.
x
,
self
.
y
],
T
.
outer
(
self
.
x
,
self
.
y
))
...
...
@@ -1690,7 +1673,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
def
test_A_plus_outer
(
self
):
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
self
.
A
+
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
A
+
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
assertFunctionContains
(
f
,
self
.
ger
)
f
(
np
.
random
.
rand
(
5
,
4
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
5
)
.
astype
(
self
.
dtype
),
...
...
@@ -1701,7 +1684,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
def
test_A_plus_scaled_outer
(
self
):
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
self
.
A
+
0.1
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
A
+
0.1
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
assertFunctionContains
(
f
,
self
.
ger
)
f
(
np
.
random
.
rand
(
5
,
4
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
5
)
.
astype
(
self
.
dtype
),
...
...
@@ -1714,7 +1697,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
np
.
asarray
(
0.2
,
self
.
dtype
)
*
self
.
A
+
np
.
asarray
(
0.1
,
self
.
dtype
)
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
x
,
self
.
y
))
# Why gemm? This make the graph simpler did we test that it
# make it faster?
self
.
assertFunctionContains
(
f
,
self
.
gemm
)
...
...
@@ -1729,7 +1712,7 @@ class TestGer(TestCase, unittest_tools.TestOptimizationMixin):
""" test corner case shape and dtype"""
f
=
self
.
function
([
self
.
A
,
self
.
x
,
self
.
y
],
self
.
A
+
0.1
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
A
+
0.1
*
T
.
outer
(
self
.
x
,
self
.
y
))
self
.
assertFunctionContains
(
f
,
self
.
ger
)
f
(
np
.
random
.
rand
(
M
,
N
)
.
astype
(
self
.
dtype
),
np
.
random
.
rand
(
M
)
.
astype
(
self
.
dtype
),
...
...
@@ -1812,21 +1795,21 @@ class TestBlasStrides(TestCase):
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_nn
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b
,
c
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
# print 'class name:', self.__class__.__name__
# theano.printing.debugprint(f_nn)
f_nt
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b
,
c_t
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tn
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b_t
.
T
,
c
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tt
=
theano
.
function
([],
[],
updates
=
[(
a
,
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed pattern
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
4
):
for
step
in
(
1
,
2
):
b_step1
,
b_step2
,
c_step1
,
c_step2
=
(
s
*
step
for
s
in
step_signs
)
for
s
in
step_signs
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
...
...
@@ -1835,7 +1818,7 @@ class TestBlasStrides(TestCase):
# Numpy result
a_n
=
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
])
cv
[::
c_step1
,
::
c_step2
])
f_nn
()
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
...
...
@@ -1883,20 +1866,20 @@ class TestBlasStrides(TestCase):
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_nn
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b
,
c
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_nt
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b
,
c_t
.
T
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tn
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b_t
.
T
,
c
))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_tt
=
theano
.
function
([],
[],
updates
=
[(
a
,
l
*
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
))],
mode
=
self
.
mode
)
updates
=
[(
a
,
l
*
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
))],
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed pattern
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
4
):
for
step
in
(
1
,
2
):
b_step1
,
b_step2
,
c_step1
,
c_step2
=
(
s
*
step
for
s
in
step_signs
)
for
s
in
step_signs
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
...
...
@@ -1905,7 +1888,7 @@ class TestBlasStrides(TestCase):
# Numpy result
a_n
=
l
*
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
])
cv
[::
c_step1
,
::
c_step2
])
f_nn
()
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
...
...
@@ -1954,36 +1937,44 @@ class TestBlasStrides(TestCase):
bt_dev
=
b_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
ct_dev
=
c_t
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_nnn
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b
,
c
)))],
mode
=
self
.
mode
)
f_nnt
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b
,
c_t
.
T
)))],
mode
=
self
.
mode
)
f_ntn
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b_t
.
T
,
c
)))],
mode
=
self
.
mode
)
f_ntt
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
)))],
mode
=
self
.
mode
)
f_tnn
=
theano
.
function
([],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b
,
c
)
.
T
))],
mode
=
self
.
mode
)
f_tnt
=
theano
.
function
([],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b
,
c_t
.
T
)
.
T
))],
mode
=
self
.
mode
)
f_ttn
=
theano
.
function
([],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b_t
.
T
,
c
)
.
T
))],
mode
=
self
.
mode
)
f_ttt
=
theano
.
function
([],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
)
.
T
))],
mode
=
self
.
mode
)
f_nnn
=
theano
.
function
(
[],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b
,
c
)))],
mode
=
self
.
mode
)
f_nnt
=
theano
.
function
(
[],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b
,
c_t
.
T
)))],
mode
=
self
.
mode
)
f_ntn
=
theano
.
function
(
[],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b_t
.
T
,
c
)))],
mode
=
self
.
mode
)
f_ntt
=
theano
.
function
(
[],
[],
updates
=
[(
a
,
(
l
*
a
+
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
)))],
mode
=
self
.
mode
)
f_tnn
=
theano
.
function
(
[],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b
,
c
)
.
T
))],
mode
=
self
.
mode
)
f_tnt
=
theano
.
function
(
[],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b
,
c_t
.
T
)
.
T
))],
mode
=
self
.
mode
)
f_ttn
=
theano
.
function
(
[],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b_t
.
T
,
c
)
.
T
))],
mode
=
self
.
mode
)
f_ttt
=
theano
.
function
(
[],
[],
updates
=
[(
a_t
,
(
l
*
a_t
+
tensor
.
dot
(
b_t
.
T
,
c_t
.
T
)
.
T
))],
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed pattern
for
step_signs
in
itertools_product
((
-
1
,
1
),
repeat
=
6
):
for
step
in
(
1
,
2
):
a_step1
,
a_step2
,
b_step1
,
b_step2
,
c_step1
,
c_step2
=
\
(
s
*
step
for
s
in
step_signs
)
(
s
*
step
for
s
in
step_signs
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step1
,
::
c_step2
],
borrow
=
True
)
...
...
@@ -1991,12 +1982,12 @@ class TestBlasStrides(TestCase):
c_t
.
set_value
(
ct_dev
.
copy
()[::
c_step2
,
::
c_step1
],
borrow
=
True
)
# Numpy results
a_n
=
(
l
*
av
[::
a_step1
,
::
a_step2
]
+
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
]))
at_n
=
(
l
*
av
[::
a_step1
,
::
a_step2
]
.
T
+
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
])
.
T
)
a_n
=
(
l
*
av
[::
a_step1
,
::
a_step2
]
+
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
]))
at_n
=
(
l
*
av
[::
a_step1
,
::
a_step2
]
.
T
+
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step1
,
::
c_step2
])
.
T
)
# a's value is updated, so we need to reinitialize it each time
a
.
set_value
(
a_dev
.
copy
()[::
a_step1
,
::
a_step2
],
borrow
=
True
)
...
...
@@ -2016,22 +2007,22 @@ class TestBlasStrides(TestCase):
assert
np
.
allclose
(
a
.
get_value
(),
a_n
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
borrow
=
True
)
borrow
=
True
)
f_tnn
()
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
borrow
=
True
)
borrow
=
True
)
f_tnt
()
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
borrow
=
True
)
borrow
=
True
)
f_ttn
()
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step2
,
::
a_step1
],
borrow
=
True
)
borrow
=
True
)
f_ttt
()
assert
np
.
allclose
(
a_t
.
get_value
(),
at_n
)
...
...
@@ -2066,28 +2057,28 @@ class TestBlasStrides(TestCase):
c_dev
=
c
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_n
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
a
+
l
*
tensor
.
dot
(
b
,
c
)))],
mode
=
self
.
mode
)
mode
=
self
.
mode
)
f_t
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
a
+
l
*
tensor
.
dot
(
b_t
.
T
,
c
)))],
mode
=
self
.
mode
)
updates
=
[(
a
,
(
a
+
l
*
tensor
.
dot
(
b_t
.
T
,
c
)))],
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed pattern
for
step_signs
in
itertools_product
((
1
,
-
1
),
repeat
=
4
):
for
step
in
(
1
,
2
):
a_step
,
b_step1
,
b_step2
,
c_step
=
(
s
*
step
for
s
in
step_signs
)
for
s
in
step_signs
)
a
.
set_value
(
a_dev
.
copy
()[::
a_step
],
borrow
=
True
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step1
,
::
b_step2
],
borrow
=
True
)
borrow
=
True
)
b_t
.
set_value
(
transpose
(
b_dev
.
copy
())[::
b_step2
,
::
b_step1
],
borrow
=
True
)
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step
],
borrow
=
True
)
a_n
=
(
av
[::
a_step
]
+
l
*
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step
]))
a_n
=
(
av
[::
a_step
]
+
l
*
np
.
dot
(
bv
[::
b_step1
,
::
b_step2
],
cv
[::
c_step
]))
f_n
()
assert
np
.
allclose
(
a
.
get_value
(),
a_n
),
(
a
.
get_value
(),
a_n
)
...
...
@@ -2120,36 +2111,37 @@ class TestBlasStrides(TestCase):
b_dev
=
b
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
c_dev
=
c
.
get_value
(
borrow
=
False
,
return_internal_type
=
True
)
f_n
=
theano
.
function
([],
[],
updates
=
[(
a
,
(
a
+
l
*
tensor
.
outer
(
b
,
c
)))],
mode
=
self
.
mode
)
f_n
=
theano
.
function
(
[],
[],
updates
=
[(
a
,
(
a
+
l
*
tensor
.
outer
(
b
,
c
)))],
mode
=
self
.
mode
)
f_t
=
theano
.
function
([],
[],
updates
=
[(
a_t
,
(
a_t
+
l
*
tensor
.
outer
(
b
,
c
)
.
T
))],
mode
=
self
.
mode
)
f_t
=
theano
.
function
(
[],
[],
updates
=
[(
a_t
,
(
a_t
+
l
*
tensor
.
outer
(
b
,
c
)
.
T
))],
mode
=
self
.
mode
)
# Try with all stride patterns, and all transposed patterns
for
step_signs
in
itertools_product
((
1
,
-
1
),
repeat
=
4
):
for
step
in
(
1
,
2
):
a_step1
,
a_step2
,
b_step
,
c_step
=
(
s
*
step
for
s
in
step_signs
)
for
s
in
step_signs
)
a
.
set_value
(
a_dev
.
copy
()[::
a_step1
,
::
a_step2
],
borrow
=
True
)
a_t
.
set_value
(
transpose
(
a_dev
.
copy
())[::
a_step1
,
::
a_step2
],
borrow
=
True
)
borrow
=
True
)
b
.
set_value
(
b_dev
.
copy
()[::
b_step
],
borrow
=
True
)
c
.
set_value
(
c_dev
.
copy
()[::
c_step
],
borrow
=
True
)
f_n
()
n_n
=
(
av
[::
a_step1
,
::
a_step2
]
+
l
*
np
.
outer
(
bv
[::
b_step
],
cv
[::
c_step
]))
n_n
=
(
av
[::
a_step1
,
::
a_step2
]
+
l
*
np
.
outer
(
bv
[::
b_step
],
cv
[::
c_step
]))
assert
np
.
allclose
(
a
.
get_value
(),
n_n
),
(
a
.
get_value
(),
n_n
)
f_t
()
n_t
=
(
av
.
T
[::
a_step1
,
::
a_step2
]
+
l
*
np
.
outer
(
bv
[::
b_step
],
cv
[::
c_step
])
.
T
)
assert
np
.
allclose
(
a_t
.
get_value
(),
n_t
),
\
(
a_t
.
get_value
(),
n_t
)
n_t
=
(
av
.
T
[::
a_step1
,
::
a_step2
]
+
l
*
np
.
outer
(
bv
[::
b_step
],
cv
[::
c_step
])
.
T
)
assert
np
.
allclose
(
a_t
.
get_value
(),
n_t
),
(
a_t
.
get_value
(),
n_t
)
def
test_ger_strides
(
self
):
self
.
cmp_ger
((
3
,
5
),
3
,
5
)
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论