Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
ce1eeab9
提交
ce1eeab9
authored
4月 09, 2014
作者:
abergeron
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #1797 from nouiz/fast_opt
Make the slow scan test fast!
上级
4a77221b
be6c8bc0
显示空白字符变更
内嵌
并排
正在显示
4 个修改的文件
包含
71 行增加
和
55 行删除
+71
-55
function_module.py
theano/compile/function_module.py
+2
-0
basic.py
theano/tensor/basic.py
+35
-28
opt.py
theano/tensor/opt.py
+9
-5
test_opt.py
theano/tensor/tests/test_opt.py
+25
-22
没有找到文件。
theano/compile/function_module.py
浏览文件 @
ce1eeab9
...
@@ -1077,6 +1077,7 @@ class FunctionMaker(object):
...
@@ -1077,6 +1077,7 @@ class FunctionMaker(object):
self
.
mode
=
mode
self
.
mode
=
mode
self
.
accept_inplace
=
accept_inplace
self
.
accept_inplace
=
accept_inplace
self
.
function_builder
=
function_builder
self
.
function_builder
=
function_builder
self
.
on_unused_input
=
on_unused_input
# Used only for the pickling
self
.
required
=
[(
i
.
value
is
None
)
for
i
in
self
.
inputs
]
self
.
required
=
[(
i
.
value
is
None
)
for
i
in
self
.
inputs
]
self
.
refeed
=
[
self
.
refeed
=
[
...
@@ -1215,6 +1216,7 @@ def _pickle_FunctionMaker(self):
...
@@ -1215,6 +1216,7 @@ def _pickle_FunctionMaker(self):
accept_inplace
=
self
.
accept_inplace
,
accept_inplace
=
self
.
accept_inplace
,
function_builder
=
self
.
function_builder
,
function_builder
=
self
.
function_builder
,
profile
=
self
.
profile
,
profile
=
self
.
profile
,
on_unused_input
=
self
.
on_unused_input
,
)
)
return
(
_constructor_FunctionMaker
,
(
kwargs
,))
return
(
_constructor_FunctionMaker
,
(
kwargs
,))
...
...
theano/tensor/basic.py
浏览文件 @
ce1eeab9
...
@@ -508,13 +508,31 @@ class EmptyConstantError(NotScalarConstantError):
...
@@ -508,13 +508,31 @@ class EmptyConstantError(NotScalarConstantError):
"""
"""
def
numpy_scalar
(
data
):
""" Return a scalar stored in a numpy ndarray, or raise
NotScalarConstantError if the numpy ndarray is not a scalar
"""
# handle case where data is numpy.array([])
if
data
.
ndim
>
0
and
(
len
(
data
.
shape
)
==
0
or
__builtins__
[
'max'
](
data
.
shape
)
==
0
):
assert
numpy
.
all
(
numpy
.
array
([])
==
data
)
raise
EmptyConstantError
()
try
:
numpy
.
complex
(
data
)
# works for all numeric scalars
return
data
except
Exception
:
raise
NotScalarConstantError
(
'v.data is non-numeric, non-scalar, or has more than one'
' unique value'
,
data
)
get_scalar_constant_value_elemwises
=
(
get_scalar_constant_value_elemwises
=
(
scal
.
Cast
,
scal
.
Switch
,
scal
.
Cast
,
scal
.
Switch
,
scal
.
NEQ
,
scal
.
EQ
,
scal
.
NEQ
,
scal
.
EQ
,
scal
.
LT
,
scal
.
GT
,
scal
.
LE
,
scal
.
GE
,
scal
.
LT
,
scal
.
GT
,
scal
.
LE
,
scal
.
GE
,
scal
.
Sub
,
scal
.
Add
,
scal
.
Mod
,
scal
.
Mul
,
scal
.
Sub
,
scal
.
Add
,
scal
.
Mod
,
scal
.
Mul
,
scal
.
IntDiv
,
scal
.
TrueDiv
)
scal
.
IntDiv
,
scal
.
TrueDiv
)
def
get_scalar_constant_value
(
v
):
def
get_scalar_constant_value
(
orig_v
,
elemwise
=
True
):
"""return the constant scalar(0-D) value underlying variable `v`
"""return the constant scalar(0-D) value underlying variable `v`
If v is the output of dimshuffles, fills, allocs, rebroadcasts, cast
If v is the output of dimshuffles, fills, allocs, rebroadcasts, cast
...
@@ -523,10 +541,14 @@ def get_scalar_constant_value(v):
...
@@ -523,10 +541,14 @@ def get_scalar_constant_value(v):
If `v` is not some view of constant scalar data, then raise a
If `v` is not some view of constant scalar data, then raise a
NotScalarConstantError.
NotScalarConstantError.
:param elemwise: If False, we won't try to go into elemwise.
So this call is faster.
:note: There may be another function similar to this one in the
:note: There may be another function similar to this one in the
code, but I'm not sure where it is.
code, but I'm not sure where it is.
"""
"""
v
=
orig_v
while
True
:
if
v
is
None
:
if
v
is
None
:
# None is not a scalar (and many uses of this function seem to depend
# None is not a scalar (and many uses of this function seem to depend
# on passing it None)
# on passing it None)
...
@@ -535,24 +557,6 @@ def get_scalar_constant_value(v):
...
@@ -535,24 +557,6 @@ def get_scalar_constant_value(v):
if
isinstance
(
v
,
(
numpy
.
integer
,
int
,
float
)):
if
isinstance
(
v
,
(
numpy
.
integer
,
int
,
float
)):
return
numpy
.
asarray
(
v
)
return
numpy
.
asarray
(
v
)
def
numpy_scalar
(
data
):
""" Return a scalar stored in a numpy ndarray, or raise
NotScalarConstantError if the numpy ndarray is not a scalar
"""
# handle case where data is numpy.array([])
if
data
.
ndim
>
0
and
(
len
(
data
.
shape
)
==
0
or
__builtins__
[
'max'
](
data
.
shape
)
==
0
):
assert
numpy
.
all
(
numpy
.
array
([])
==
data
)
raise
EmptyConstantError
()
try
:
numpy
.
complex
(
data
)
# works for all numeric scalars
return
data
except
Exception
:
raise
NotScalarConstantError
(
'v.data is non-numeric, non-scalar, or has more than one'
' unique value'
,
data
)
if
isinstance
(
v
,
numpy
.
ndarray
):
if
isinstance
(
v
,
numpy
.
ndarray
):
return
numpy_scalar
(
v
)
return
numpy_scalar
(
v
)
...
@@ -567,9 +571,10 @@ def get_scalar_constant_value(v):
...
@@ -567,9 +571,10 @@ def get_scalar_constant_value(v):
if
isinstance
(
v
.
owner
.
op
,
(
Alloc
,
DimShuffle
,
Rebroadcast
,
if
isinstance
(
v
.
owner
.
op
,
(
Alloc
,
DimShuffle
,
Rebroadcast
,
compile
.
ops
.
OutputGuard
,
compile
.
ops
.
OutputGuard
,
compile
.
DeepCopyOp
)):
compile
.
DeepCopyOp
)):
return
get_scalar_constant_value
(
v
.
owner
.
inputs
[
0
])
v
=
v
.
owner
.
inputs
[
0
]
elif
(
isinstance
(
v
.
owner
.
op
,
theano
.
compile
.
ops
.
Shape_i
)
and
continue
isinstance
(
v
.
owner
.
inputs
[
0
],
Constant
)):
elif
isinstance
(
v
.
owner
.
op
,
theano
.
compile
.
ops
.
Shape_i
):
if
isinstance
(
v
.
owner
.
inputs
[
0
],
Constant
):
return
v
.
owner
.
inputs
[
0
]
.
data
.
shape
[
v
.
owner
.
op
.
i
]
return
v
.
owner
.
inputs
[
0
]
.
data
.
shape
[
v
.
owner
.
op
.
i
]
# Don't act as the constant_folding optimization here as this
# Don't act as the constant_folding optimization here as this
# fct is used too early in the optimization phase. This would
# fct is used too early in the optimization phase. This would
...
@@ -580,18 +585,20 @@ def get_scalar_constant_value(v):
...
@@ -580,18 +585,20 @@ def get_scalar_constant_value(v):
if
isinstance
(
v
.
owner
.
op
,
scal
.
Second
):
if
isinstance
(
v
.
owner
.
op
,
scal
.
Second
):
# We don't need both input to be constant for second
# We don't need both input to be constant for second
shape
,
val
=
v
.
owner
.
inputs
shape
,
val
=
v
.
owner
.
inputs
return
get_scalar_constant_value
(
val
)
v
=
val
continue
if
isinstance
(
v
.
owner
.
op
,
get_scalar_constant_value_elemwises
):
if
isinstance
(
v
.
owner
.
op
,
get_scalar_constant_value_elemwises
):
const
=
[
get_scalar_constant_value
(
i
)
const
=
[
get_scalar_constant_value
(
i
)
for
i
in
v
.
owner
.
inputs
]
for
i
in
v
.
owner
.
inputs
]
ret
=
[[
None
]]
ret
=
[[
None
]]
v
.
owner
.
op
.
perform
(
v
.
owner
,
const
,
ret
)
v
.
owner
.
op
.
perform
(
v
.
owner
,
const
,
ret
)
return
ret
[
0
][
0
]
return
ret
[
0
][
0
]
elif
isinstance
(
v
.
owner
.
op
,
Elemwise
):
elif
elemwise
and
isinstance
(
v
.
owner
.
op
,
Elemwise
):
if
isinstance
(
v
.
owner
.
op
.
scalar_op
,
scal
.
Second
):
if
isinstance
(
v
.
owner
.
op
.
scalar_op
,
scal
.
Second
):
# We don't need both input to be constant for second
# We don't need both input to be constant for second
shape
,
val
=
v
.
owner
.
inputs
shape
,
val
=
v
.
owner
.
inputs
return
get_scalar_constant_value
(
val
)
v
=
val
continue
elif
isinstance
(
v
.
owner
.
op
.
scalar_op
,
elif
isinstance
(
v
.
owner
.
op
.
scalar_op
,
get_scalar_constant_value_elemwises
):
get_scalar_constant_value_elemwises
):
const
=
[
get_scalar_constant_value
(
i
)
for
i
in
v
.
owner
.
inputs
]
const
=
[
get_scalar_constant_value
(
i
)
for
i
in
v
.
owner
.
inputs
]
...
@@ -3075,7 +3082,7 @@ pprint.assign(pow, printing.OperatorPrinter('**', 1, 'right'))
...
@@ -3075,7 +3082,7 @@ pprint.assign(pow, printing.OperatorPrinter('**', 1, 'right'))
##########################
##########################
def
extract_constant
(
x
):
def
extract_constant
(
x
,
elemwise
=
True
):
'''
'''
This function is basically a call to tensor.get_scalar_constant_value. The
This function is basically a call to tensor.get_scalar_constant_value. The
main difference is the behaviour in case of failure. While
main difference is the behaviour in case of failure. While
...
@@ -3085,7 +3092,7 @@ def extract_constant(x):
...
@@ -3085,7 +3092,7 @@ def extract_constant(x):
ScalarVariable, we convert it to a tensor with tensor_from_scalar.
ScalarVariable, we convert it to a tensor with tensor_from_scalar.
'''
'''
try
:
try
:
x
=
get_scalar_constant_value
(
x
)
x
=
get_scalar_constant_value
(
x
,
elemwise
=
elemwise
)
except
NotScalarConstantError
:
except
NotScalarConstantError
:
pass
pass
if
(
isinstance
(
x
,
scal
.
ScalarVariable
)
or
if
(
isinstance
(
x
,
scal
.
ScalarVariable
)
or
...
...
theano/tensor/opt.py
浏览文件 @
ce1eeab9
...
@@ -1581,7 +1581,7 @@ def local_upcast_elemwise_constant_inputs(node):
...
@@ -1581,7 +1581,7 @@ def local_upcast_elemwise_constant_inputs(node):
else
:
else
:
try
:
try
:
# works only for scalars
# works only for scalars
cval_i
=
get_scalar_constant_value
(
i
)
cval_i
=
get_scalar_constant_value
(
i
,
elemwise
=
False
)
if
all
(
i
.
broadcastable
):
if
all
(
i
.
broadcastable
):
new_inputs
.
append
(
T
.
shape_padleft
(
new_inputs
.
append
(
T
.
shape_padleft
(
T
.
cast
(
cval_i
,
output_dtype
),
T
.
cast
(
cval_i
,
output_dtype
),
...
@@ -2327,7 +2327,7 @@ def local_remove_switch_const_cond(node):
...
@@ -2327,7 +2327,7 @@ def local_remove_switch_const_cond(node):
"""
"""
if
(
isinstance
(
node
.
op
,
T
.
Elemwise
)
and
if
(
isinstance
(
node
.
op
,
T
.
Elemwise
)
and
isinstance
(
node
.
op
.
scalar_op
,
scalar
.
basic
.
Switch
)):
isinstance
(
node
.
op
.
scalar_op
,
scalar
.
basic
.
Switch
)):
cond
=
T
.
extract_constant
(
node
.
inputs
[
0
])
cond
=
T
.
extract_constant
(
node
.
inputs
[
0
]
,
elemwise
=
False
)
if
type
(
cond
)
is
numpy
.
ndarray
and
cond
.
ndim
==
0
:
if
type
(
cond
)
is
numpy
.
ndarray
and
cond
.
ndim
==
0
:
if
cond
==
0
:
if
cond
==
0
:
out
=
node
.
inputs
[
2
]
out
=
node
.
inputs
[
2
]
...
@@ -2377,7 +2377,8 @@ def local_mul_switch_sink(node):
...
@@ -2377,7 +2377,8 @@ def local_mul_switch_sink(node):
if
i
.
owner
and
i
.
owner
.
op
==
T
.
switch
:
if
i
.
owner
and
i
.
owner
.
op
==
T
.
switch
:
switch
=
i
.
owner
switch
=
i
.
owner
try
:
try
:
if
get_scalar_constant_value
(
switch
.
inputs
[
1
])
==
0.
:
if
(
isinstance
(
switch
.
inputs
[
0
],
Constant
)
and
get_scalar_constant_value
(
switch
.
inputs
[
1
])
==
0.
):
listmul
=
node
.
inputs
[:
idx
]
+
node
.
inputs
[
idx
+
1
:]
listmul
=
node
.
inputs
[:
idx
]
+
node
.
inputs
[
idx
+
1
:]
fct
=
[
T
.
switch
(
switch
.
inputs
[
0
],
0
,
fct
=
[
T
.
switch
(
switch
.
inputs
[
0
],
0
,
T
.
mul
(
*
(
listmul
+
[
switch
.
inputs
[
2
]])))]
T
.
mul
(
*
(
listmul
+
[
switch
.
inputs
[
2
]])))]
...
@@ -2387,7 +2388,8 @@ def local_mul_switch_sink(node):
...
@@ -2387,7 +2388,8 @@ def local_mul_switch_sink(node):
except
NotScalarConstantError
:
except
NotScalarConstantError
:
pass
pass
try
:
try
:
if
get_scalar_constant_value
(
switch
.
inputs
[
2
])
==
0.
:
if
(
isinstance
(
switch
.
inputs
[
2
],
Constant
)
and
get_scalar_constant_value
(
switch
.
inputs
[
2
])
==
0.
):
listmul
=
node
.
inputs
[:
idx
]
+
node
.
inputs
[
idx
+
1
:]
listmul
=
node
.
inputs
[:
idx
]
+
node
.
inputs
[
idx
+
1
:]
fct
=
[
T
.
switch
(
switch
.
inputs
[
0
],
fct
=
[
T
.
switch
(
switch
.
inputs
[
0
],
T
.
mul
(
*
(
listmul
+
[
switch
.
inputs
[
1
]])),
0
)]
T
.
mul
(
*
(
listmul
+
[
switch
.
inputs
[
1
]])),
0
)]
...
@@ -3784,7 +3786,7 @@ def local_abs_merge(node):
...
@@ -3784,7 +3786,7 @@ def local_abs_merge(node):
for
i
in
node
.
inputs
:
for
i
in
node
.
inputs
:
if
i
.
owner
and
i
.
owner
.
op
==
T
.
abs_
:
if
i
.
owner
and
i
.
owner
.
op
==
T
.
abs_
:
inputs
.
append
(
i
.
owner
.
inputs
[
0
])
inputs
.
append
(
i
.
owner
.
inputs
[
0
])
el
se
:
el
if
isinstance
(
i
,
Constant
)
:
try
:
try
:
const
=
get_scalar_constant_value
(
i
)
const
=
get_scalar_constant_value
(
i
)
except
NotScalarConstantError
:
except
NotScalarConstantError
:
...
@@ -3792,6 +3794,8 @@ def local_abs_merge(node):
...
@@ -3792,6 +3794,8 @@ def local_abs_merge(node):
if
not
(
const
>=
0
)
.
all
():
if
not
(
const
>=
0
)
.
all
():
return
False
return
False
inputs
.
append
(
i
)
inputs
.
append
(
i
)
else
:
return
False
return
[
T
.
abs_
(
T
.
mul
(
*
inputs
))]
return
[
T
.
abs_
(
T
.
mul
(
*
inputs
))]
if
node
.
op
==
T
.
true_div
and
sum
([
i
.
owner
.
op
==
T
.
abs_
for
i
in
if
node
.
op
==
T
.
true_div
and
sum
([
i
.
owner
.
op
==
T
.
abs_
for
i
in
node
.
inputs
if
i
.
owner
])
==
2
:
node
.
inputs
if
i
.
owner
])
==
2
:
...
...
theano/tensor/tests/test_opt.py
浏览文件 @
ce1eeab9
...
@@ -1495,11 +1495,11 @@ def test_log1p():
...
@@ -1495,11 +1495,11 @@ def test_log1p():
f
=
function
([
x
],
T
.
log
(
1
+
(
x
)),
mode
=
m
)
f
=
function
([
x
],
T
.
log
(
1
+
(
x
)),
mode
=
m
)
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
T
.
log1p
]
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
T
.
log1p
]
f
=
function
([
x
],
T
.
log
(
1
+
(
-
x
)),
mode
=
m
)
f
=
function
([
x
],
T
.
log
(
1
+
(
-
x
)),
mode
=
m
)
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
T
.
neg
,
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
inplace
.
log1p_inplace
]
T
.
neg
,
inplace
.
log1p_inplace
]
f
=
function
([
x
],
-
T
.
log
(
1
+
(
-
x
)),
mode
=
m
)
f
=
function
([
x
],
-
T
.
log
(
1
+
(
-
x
)),
mode
=
m
)
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
T
.
neg
,
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
inplace
.
log1p_inplace
,
inplace
.
neg_inplace
]
T
.
neg
,
inplace
.
log1p_inplace
,
inplace
.
neg_inplace
]
# check trickier cases (and use different dtype)
# check trickier cases (and use different dtype)
y
=
fmatrix
()
y
=
fmatrix
()
...
@@ -1507,12 +1507,12 @@ def test_log1p():
...
@@ -1507,12 +1507,12 @@ def test_log1p():
print
f
.
maker
.
fgraph
.
toposort
()
print
f
.
maker
.
fgraph
.
toposort
()
# the first three ops are Shape_i, Shape_i, and Dimshuffle
# the first three ops are Shape_i, Shape_i, and Dimshuffle
theano
.
printing
.
debugprint
(
f
)
theano
.
printing
.
debugprint
(
f
)
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()][
3
:]
\
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()][
3
:]
==
[
==
[
T
.
log1p
,
tensor
.
alloc
]
T
.
log1p
,
tensor
.
alloc
]
f
=
function
([
x
,
y
],
T
.
log
(
0
+
(
x
)
+
tensor
.
fill
(
y
,
1.0
)),
mode
=
m
)
f
=
function
([
x
,
y
],
T
.
log
(
0
+
(
x
)
+
tensor
.
fill
(
y
,
1.0
)),
mode
=
m
)
theano
.
printing
.
debugprint
(
f
)
theano
.
printing
.
debugprint
(
f
)
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()][
3
:]
\
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()][
3
:]
==
[
==
[
T
.
log1p
,
tensor
.
alloc
]
T
.
log1p
,
tensor
.
alloc
]
f
=
function
([
x
,
y
],
T
.
log
(
2
+
(
x
)
-
tensor
.
fill
(
y
,
1.0
)),
mode
=
m
)
f
=
function
([
x
,
y
],
T
.
log
(
2
+
(
x
)
-
tensor
.
fill
(
y
,
1.0
)),
mode
=
m
)
theano
.
printing
.
debugprint
(
f
)
theano
.
printing
.
debugprint
(
f
)
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()][
3
:]
\
assert
[
node
.
op
for
node
in
f
.
maker
.
fgraph
.
toposort
()][
3
:]
\
...
@@ -2333,7 +2333,8 @@ class Test_alloc_zero(unittest.TestCase):
...
@@ -2333,7 +2333,8 @@ class Test_alloc_zero(unittest.TestCase):
def
setUp
(
self
):
def
setUp
(
self
):
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
self
.
mode
=
mode
.
including
(
"local_incsubtensor_of_allocs"
,
self
.
mode
=
mode
.
including
(
"local_incsubtensor_of_allocs"
,
"local_setsubtensor_of_allocs"
,
"local_0_dot_x"
)
"local_setsubtensor_of_allocs"
,
"local_0_dot_x"
)
def
test_setsubtensor_allocs0
(
self
):
def
test_setsubtensor_allocs0
(
self
):
x
=
tensor
.
matrix
()
x
=
tensor
.
matrix
()
...
@@ -2427,7 +2428,7 @@ class Test_alloc_zero(unittest.TestCase):
...
@@ -2427,7 +2428,7 @@ class Test_alloc_zero(unittest.TestCase):
f
(
_e1
[
1
],
_e2
[
1
])
f
(
_e1
[
1
],
_e2
[
1
])
f
(
_e1
[
2
],
_e2
[
2
])
f
(
_e1
[
2
],
_e2
[
2
])
assert
numpy
.
all
([
not
isinstance
(
x
.
op
,
tensor
.
Dot
)
for
x
in
assert
numpy
.
all
([
not
isinstance
(
x
.
op
,
tensor
.
Dot
)
for
x
in
f
.
maker
.
fgraph
.
toposort
()
])
f
.
maker
.
fgraph
.
toposort
()])
#test that we don't remove shape errors
#test that we don't remove shape errors
self
.
assertRaises
((
ValueError
,
AssertionError
),
f
,
self
.
assertRaises
((
ValueError
,
AssertionError
),
f
,
...
@@ -2809,8 +2810,8 @@ class test_assert(utt.InferShapeTester):
...
@@ -2809,8 +2810,8 @@ class test_assert(utt.InferShapeTester):
x
=
T
.
scalar
()
x
=
T
.
scalar
()
y
=
T
.
scalar
()
y
=
T
.
scalar
()
f
=
theano
.
function
([
x
,
y
],
theano
.
tensor
.
opt
.
assert_
(
x
,
y
,
f
=
theano
.
function
([
x
,
y
],
theano
.
tensor
.
opt
.
assert_
(
x
,
y
,
1
),
1
),
mode
=
mode
)
mode
=
mode
)
assert
f
(
1
,
1
)
==
1
assert
f
(
1
,
1
)
==
1
assert
f
(
5
,
1
)
==
5
assert
f
(
5
,
1
)
==
5
topo
=
f
.
maker
.
fgraph
.
toposort
()
topo
=
f
.
maker
.
fgraph
.
toposort
()
...
@@ -2827,8 +2828,8 @@ class test_assert(utt.InferShapeTester):
...
@@ -2827,8 +2828,8 @@ class test_assert(utt.InferShapeTester):
x
=
T
.
scalar
()
x
=
T
.
scalar
()
y
=
T
.
scalar
()
y
=
T
.
scalar
()
f
=
theano
.
function
([
x
,
y
],
theano
.
tensor
.
opt
.
assert_
(
x
,
y
,
f
=
theano
.
function
([
x
,
y
],
theano
.
tensor
.
opt
.
assert_
(
x
,
y
,
0
),
0
),
mode
=
mode
)
mode
=
mode
)
self
.
assertRaises
(
AssertionError
,
f
,
1
,
0
)
self
.
assertRaises
(
AssertionError
,
f
,
1
,
0
)
topo
=
f
.
maker
.
fgraph
.
toposort
()
topo
=
f
.
maker
.
fgraph
.
toposort
()
assert
len
(
topo
)
==
2
assert
len
(
topo
)
==
2
...
@@ -3177,8 +3178,9 @@ def test_constant_get_stabilized():
...
@@ -3177,8 +3178,9 @@ def test_constant_get_stabilized():
class
T_local_switch_sink
(
unittest
.
TestCase
):
class
T_local_switch_sink
(
unittest
.
TestCase
):
def
setUp
(
self
):
def
setUp
(
self
):
# condition values
# condition values
self
.
condm
=
numpy
.
asarray
([[
0.1
,
0
,
1
,
-
1
],
[
0.
,
0.
,
0.
,
self
.
condm
=
numpy
.
asarray
([[
0.1
,
0
,
1
,
-
1
],
0.
],
[
1
,
1
,
1
,
1
]])
[
0.
,
0.
,
0.
,
0.
],
[
1
,
1
,
1
,
1
]])
self
.
condv
=
numpy
.
asarray
([
0.1
,
0
,
1
,
-
1
])
self
.
condv
=
numpy
.
asarray
([
0.1
,
0
,
1
,
-
1
])
self
.
conds
=
[
0.1
,
0
,
1
,
-
1
]
self
.
conds
=
[
0.1
,
0
,
1
,
-
1
]
...
@@ -3256,14 +3258,14 @@ class T_local_erf(unittest.TestCase):
...
@@ -3256,14 +3258,14 @@ class T_local_erf(unittest.TestCase):
f
=
theano
.
function
([
x
],
1
+
T
.
erf
(
x
),
mode
=
self
.
mode
)
f
=
theano
.
function
([
x
],
1
+
T
.
erf
(
x
),
mode
=
self
.
mode
)
print
f
.
maker
.
fgraph
.
toposort
()
print
f
.
maker
.
fgraph
.
toposort
()
assert
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
T
.
mul
,
T
.
assert
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
erfc
],
f
.
maker
.
fgraph
.
toposort
()
T
.
mul
,
T
.
erfc
],
f
.
maker
.
fgraph
.
toposort
()
f
(
val
)
f
(
val
)
f
=
theano
.
function
([
x
],
T
.
erf
(
x
)
+
1
,
mode
=
self
.
mode
)
f
=
theano
.
function
([
x
],
T
.
erf
(
x
)
+
1
,
mode
=
self
.
mode
)
print
f
.
maker
.
fgraph
.
toposort
()
print
f
.
maker
.
fgraph
.
toposort
()
assert
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
T
.
mul
,
T
.
assert
[
n
.
op
for
n
in
f
.
maker
.
fgraph
.
toposort
()]
==
[
erfc
],
f
.
maker
.
fgraph
.
toposort
()
T
.
mul
,
T
.
erfc
],
f
.
maker
.
fgraph
.
toposort
()
f
(
val
)
f
(
val
)
f
=
theano
.
function
([
x
],
T
.
erf
(
x
)
+
2
,
mode
=
self
.
mode
)
f
=
theano
.
function
([
x
],
T
.
erf
(
x
)
+
2
,
mode
=
self
.
mode
)
...
@@ -3305,7 +3307,7 @@ class T_local_erf(unittest.TestCase):
...
@@ -3305,7 +3307,7 @@ class T_local_erf(unittest.TestCase):
assert
topo
[
0
]
.
op
==
T
.
erf
,
f
.
maker
.
fgraph
.
toposort
()
assert
topo
[
0
]
.
op
==
T
.
erf
,
f
.
maker
.
fgraph
.
toposort
()
assert
isinstance
(
topo
[
1
]
.
op
,
T
.
Elemwise
),
f
.
maker
.
fgraph
.
toposort
()
assert
isinstance
(
topo
[
1
]
.
op
,
T
.
Elemwise
),
f
.
maker
.
fgraph
.
toposort
()
assert
isinstance
(
topo
[
1
]
.
op
.
scalar_op
,
scal
.
Add
)
\
assert
isinstance
(
topo
[
1
]
.
op
.
scalar_op
,
scal
.
Add
)
\
or
isinstance
(
topo
[
1
]
.
op
.
scalar_op
,
scal
.
Sub
),
f
.
maker
.
fgraph
.
toposort
()
or
isinstance
(
topo
[
1
]
.
op
.
scalar_op
,
scal
.
Sub
),
f
.
maker
.
fgraph
.
toposort
()
print
f
(
val
)
print
f
(
val
)
def
test_local_erf_minus_one
(
self
):
def
test_local_erf_minus_one
(
self
):
...
@@ -3345,7 +3347,8 @@ class T_local_erfc(unittest.TestCase):
...
@@ -3345,7 +3347,8 @@ class T_local_erfc(unittest.TestCase):
'canonicalize'
)
.
including
(
'fast_run'
)
.
excluding
(
'gpu'
)
'canonicalize'
)
.
including
(
'fast_run'
)
.
excluding
(
'gpu'
)
self
.
mode
=
self
.
mode_fusion
.
excluding
(
'fusion'
)
self
.
mode
=
self
.
mode_fusion
.
excluding
(
'fusion'
)
self
.
mode
.
_optimizer
.
position_cutoff
=
1.50001
self
.
mode
.
_optimizer
.
position_cutoff
=
1.50001
if
theano
.
config
.
cxx
==
''
and
not
theano
.
scalar
.
basic_scipy
.
imported_scipy_special
:
if
(
theano
.
config
.
cxx
==
''
and
not
theano
.
scalar
.
basic_scipy
.
imported_scipy_special
):
raise
SkipTest
(
"erfc need a c++ compiler or scipy"
)
raise
SkipTest
(
"erfc need a c++ compiler or scipy"
)
def
test_local_one_minus_erfc
(
self
):
def
test_local_one_minus_erfc
(
self
):
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论