Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
b031dfe0
提交
b031dfe0
authored
2月 27, 2009
作者:
James Bergstra
浏览文件
操作
浏览文件
下载
差异文件
merge
上级
f7a67132
4f79b6bc
隐藏空白字符变更
内嵌
并排
正在显示
9 个修改的文件
包含
196 行增加
和
37 行删除
+196
-37
debugmode.py
theano/compile/debugmode.py
+146
-23
link.py
theano/gof/link.py
+1
-1
opt.py
theano/gof/opt.py
+2
-6
type.py
theano/gof/type.py
+8
-0
basic.py
theano/scalar/basic.py
+8
-0
basic.py
theano/sparse/basic.py
+9
-0
basic.py
theano/tensor/basic.py
+11
-0
test_basic.py
theano/tensor/tests/test_basic.py
+7
-4
test_blas.py
theano/tensor/tests/test_blas.py
+4
-3
没有找到文件。
theano/compile/debugmode.py
浏览文件 @
b031dfe0
...
@@ -103,7 +103,25 @@ class BadOptimization(DebugModeError):
...
@@ -103,7 +103,25 @@ class BadOptimization(DebugModeError):
class
BadDestroyMap
(
DebugModeError
):
class
BadDestroyMap
(
DebugModeError
):
"""TODO #318"""
"""TODO #318"""
pass
def
__init__
(
self
,
node
,
idx
,
old_val
,
new_val
):
super
(
BadDestroyMap
,
self
)
.
__init__
()
self
.
node
=
node
self
.
idx
=
idx
self
.
old_val
=
old_val
self
.
new_val
=
new_val
def
__str__
(
self
):
sio
=
StringIO
()
print
>>
sio
,
" node:"
,
self
.
node
print
>>
sio
,
" node.inputs:"
,
[(
str
(
i
),
id
(
i
))
for
i
in
self
.
node
.
inputs
]
print
>>
sio
,
" destroy_map:"
,
getattr
(
self
.
node
.
op
,
'destroy_map'
,
{})
print
>>
sio
,
" changed input idx:"
,
self
.
idx
print
>>
sio
,
" changed input type:"
,
self
.
node
.
inputs
[
self
.
idx
]
.
type
print
>>
sio
,
" old val:"
,
self
.
old_val
print
>>
sio
,
" new val:"
,
self
.
new_val
print
>>
sio
,
""
print
>>
sio
,
" Hint: this can be caused by a deficient values_eq_enough() or __eq__() implementation that compares node input values"
return
sio
.
getvalue
()
class
StochasticOrder
(
DebugModeError
):
class
StochasticOrder
(
DebugModeError
):
"""TODO #319"""
"""TODO #319"""
...
@@ -113,6 +131,17 @@ class FloatError(DebugModeError):
...
@@ -113,6 +131,17 @@ class FloatError(DebugModeError):
"""TODO #320"""
"""TODO #320"""
pass
pass
class
InvalidValueError
(
DebugModeError
):
"""Exception: some Op an output value that is inconsistent with the Type of that output"""
def
__init__
(
self
,
r
,
v
):
super
(
InvalidValueError
,
self
)
.
__init__
()
self
.
r
=
r
self
.
v
=
v
def
__str__
(
self
):
r
,
v
=
self
.
r
,
self
.
v
return
"InvalidValueError: Result
%
s, Type
%
s, type(Value)
%
s, Value
%
s"
\
%
(
str
(
r
),
str
(
r
.
type
),
str
(
type
(
v
)),
str
(
v
)[
0
:
100
])
def
_debugprint
(
r
,
prefix
=
''
,
depth
=-
1
,
done
=
None
,
file
=
sys
.
stdout
):
def
_debugprint
(
r
,
prefix
=
''
,
depth
=-
1
,
done
=
None
,
file
=
sys
.
stdout
):
"""Print the graph leading to `r` to given depth.
"""Print the graph leading to `r` to given depth.
...
@@ -174,6 +203,29 @@ def _optcheck_env(input_specs, output_specs, accept_inplace = False):
...
@@ -174,6 +203,29 @@ def _optcheck_env(input_specs, output_specs, accept_inplace = False):
env
.
extend
(
Supervisor
(
input
for
spec
,
input
in
zip
(
input_specs
,
inputs
)
if
not
(
spec
.
mutable
or
(
hasattr
(
env
,
'destroyers'
)
and
env
.
destroyers
(
input
)))))
env
.
extend
(
Supervisor
(
input
for
spec
,
input
in
zip
(
input_specs
,
inputs
)
if
not
(
spec
.
mutable
or
(
hasattr
(
env
,
'destroyers'
)
and
env
.
destroyers
(
input
)))))
return
env
,
map
(
SymbolicOutput
,
updates
),
equivalence_tracker
return
env
,
map
(
SymbolicOutput
,
updates
),
equivalence_tracker
def
_check_inputs
(
node
,
storage_map
,
r_vals
,
dr_vals
,
active_nodes
):
"""Raise BadDestroyMap if necessary, update dr_vals"""
destroyed_idx_list
=
[]
destroy_map
=
getattr
(
node
.
op
,
'destroy_map'
,
{})
for
o_pos
,
i_pos_list
in
destroy_map
.
iteritems
():
destroyed_idx_list
.
extend
(
i_pos_list
)
destroyed_res_list
=
[
node
.
inputs
[
i
]
for
i
in
destroyed_idx_list
]
for
r_idx
,
r
in
enumerate
(
node
.
inputs
):
if
not
r
.
type
.
values_eq_enough
(
r_vals
[
r
],
storage_map
[
r
][
0
]):
# some input node 'r' got changed by running the node
# this may or may not be ok...
if
r
in
destroyed_res_list
:
# ok, we expected r to be destroyed
if
node
in
active_nodes
:
if
dr_vals
.
get
(
r
,
(
0
,
node
))[
1
]
is
not
node
:
# bad: there should only be one active node that destroys any result
raise
Exception
(
'failure in topological ordering'
)
dr_vals
[
r
]
=
(
storage_map
[
r
][
0
],
node
)
#no copy, this is the last use of this variable
else
:
raise
BadDestroyMap
(
node
,
r_idx
,
r_vals
[
r
],
storage_map
[
r
][
0
])
class
_EnvEvent
(
object
):
class
_EnvEvent
(
object
):
"""A record of an event in the life of an Env.
"""A record of an event in the life of an Env.
...
@@ -383,6 +435,9 @@ class _Linker(gof.link.LocalLinker):
...
@@ -383,6 +435,9 @@ class _Linker(gof.link.LocalLinker):
order_outputs
.
reverse
()
order_outputs
.
reverse
()
order
=
graph
.
io_toposort
(
env
.
inputs
,
order_outputs
)
order
=
graph
.
io_toposort
(
env
.
inputs
,
order_outputs
)
active_order
=
env
.
toposort
()
#an ordering of just the active nodes
active_order_set
=
set
(
active_order
)
no_recycling
=
self
.
no_recycling
no_recycling
=
self
.
no_recycling
input_storage
,
output_storage
,
storage_map
=
link
.
map_storage
(
env
,
order
,
input_storage
,
output_storage
)
input_storage
,
output_storage
,
storage_map
=
link
.
map_storage
(
env
,
order
,
input_storage
,
output_storage
)
...
@@ -459,61 +514,78 @@ class _Linker(gof.link.LocalLinker):
...
@@ -459,61 +514,78 @@ class _Linker(gof.link.LocalLinker):
#
#
# This dictionary is used to populate the storage_map as necessary
# This dictionary is used to populate the storage_map as necessary
r_vals
=
{}
r_vals
=
{}
# dr_vals are the values taken by results after being destroyed
dr_vals
=
{}
assert
len
(
thunks_py
)
==
len
(
order
)
assert
len
(
thunks_py
)
==
len
(
order
)
#
put the initial values in
to the r_vals
#
transfer the initial values from the storage_map
to the r_vals
for
r
in
storage_map
:
for
r
in
storage_map
:
if
storage_map
[
r
][
0
]
is
not
None
:
if
storage_map
[
r
][
0
]
is
not
None
:
r_vals
[
r
]
=
copy
.
copy
(
storage_map
[
r
][
0
])
if
r
.
owner
is
not
None
:
# DEBUG
print
r
,
storage_map
[
r
],
type
(
storage_map
[
r
]),
id
(
storage_map
[
r
])
assert
r
.
owner
is
None
r_vals
[
r
]
=
storage_map
[
r
][
0
]
storage_map
[
r
][
0
]
=
None
storage_map
[
r
][
0
]
=
None
#####
# Precondition: the storage map is empty, transferred completely to r_vals
#####
for
r
,
s
in
storage_map
.
iteritems
():
assert
s
[
0
]
is
None
try
:
try
:
# compute the value of all results
# compute the value of all results
for
i
,
(
thunk_py
,
thunk_c
,
node
)
in
enumerate
(
zip
(
thunks_py
,
thunks_c
,
order
)):
for
i
,
(
thunk_py
,
thunk_c
,
node
)
in
enumerate
(
zip
(
thunks_py
,
thunks_c
,
order
)):
this_node_destroyed_results
=
set
()
#put a copy of each input into the storage_map
#
put a copy of each input into the storage_map
for
r
in
node
.
inputs
:
for
r
in
node
.
inputs
:
storage_map
[
r
][
0
]
=
copy
.
copy
(
r_vals
[
r
])
storage_map
[
r
][
0
]
=
copy
.
copy
(
r_vals
[
r
])
thunk_py
()
thunk_py
()
_check_inputs
(
node
,
storage_map
,
r_vals
,
dr_vals
,
active_order_set
)
#retrieve a copy of each output from the storage_map
#retrieve a copy of each output from the storage_map
for
r
in
node
.
outputs
:
for
r
in
node
.
outputs
:
if
not
r
.
type
.
is_valid_value
(
storage_map
[
r
][
0
]):
raise
InvalidValueError
(
r
,
storage_map
[
r
][
0
])
if
r
in
r_vals
:
if
r
in
r_vals
:
# r has been constant-folded
print
>>
sys
.
stderr
,
'OUTPUT'
,
r
,
'ALREADY HAS_VALUE!'
,
r_vals
[
r
],
'WHAT ABOUT'
,
storage_map
[
r
][
0
]
if
not
r
.
type
.
values_eq_enough
(
r_vals
[
r
],
storage_map
[
r
][
0
]):
assert
r
not
in
r_vals
raise
DebugModeError
(
'BadConstantFold'
,
(
r
,
r_vals
[
r
],
r_vals
[
r
]
=
storage_map
[
r
][
0
]
storage_map
[
r
][
0
]))
#TODO: make a proper exception class for this
storage_map
[
r
][
0
]
=
None
#clear the storage_map for the thunk_c
else
:
r_vals
[
r
]
=
copy
.
copy
(
storage_map
[
r
][
0
])
if
thunk_c
:
if
thunk_c
:
for
r
in
node
.
outputs
:
storage_map
[
r
][
0
]
=
None
#clear the storage_map for the thunk_c
if
0
:
for
r
in
node
.
inputs
:
# TODO: check that Op didn't change any inputs that it wasn't allowed to
# TODO: we only need to overwrite the non-destroyed inputs
# (Hint: use the destroy_map attribute)
storage_map
[
r
][
0
]
=
copy
.
copy
(
r_vals
[
r
])
raise
NotImplementedError
()
else
:
for
r
in
node
.
inputs
:
storage_map
[
r
][
0
]
=
copy
.
copy
(
r_vals
[
r
])
thunk_c
()
thunk_c
()
_check_inputs
(
node
,
storage_map
,
r_vals
,
dr_vals
,
active_order_set
)
for
r
in
node
.
outputs
:
for
r
in
node
.
outputs
:
if
not
r
.
type
.
is_valid_value
(
storage_map
[
r
][
0
]):
raise
InvalidValueError
(
r
,
storage_map
[
r
][
0
])
# compares the version from thunk_py (in r_vals)
# compares the version from thunk_py (in r_vals)
# to the version produced by thunk_c (in storage_map)
# to the version produced by thunk_c (in storage_map)
if
not
r
.
type
.
values_eq_enough
(
r_vals
[
r
],
storage_map
[
r
][
0
]):
if
not
r
.
type
.
values_eq_enough
(
r_vals
[
r
],
storage_map
[
r
][
0
]):
raise
BadClinkerOutput
(
r
,
val_py
=
r_vals
[
r
],
val_c
=
storage_map
[
r
][
0
])
raise
BadClinkerOutput
(
r
,
val_py
=
r_vals
[
r
],
val_c
=
storage_map
[
r
][
0
])
storage_map
[
r
][
0
]
=
None
#clear the storage_map for the thunk_c
# we're done with this thunk
# clear everything out of the storage_map
for
r
in
node
.
inputs
:
storage_map
[
r
][
0
]
=
None
except
:
except
:
raise_with_op
(
node
)
raise_with_op
(
node
)
# iterate over results looking for values that don't match the values of the
# iterate over results looking for values that don't match the values of the
# results they replaced. This is the sign of a broken optimization.
# results they replaced. This is the sign of a broken optimization.
for
i
,
node
in
enumerate
(
order
):
for
i
,
node
in
enumerate
(
order
):
for
new_r
in
node
.
outputs
:
for
new_r
in
node
.
outputs
:
for
reason
,
r
,
old_graph_str
,
new_graph_str
in
env
.
equivalence_tracker
.
reasons
[
new_r
]:
for
reason
,
r
,
old_graph_str
,
new_graph_str
in
env
.
equivalence_tracker
.
reasons
[
new_r
]:
...
@@ -532,10 +604,58 @@ class _Linker(gof.link.LocalLinker):
...
@@ -532,10 +604,58 @@ class _Linker(gof.link.LocalLinker):
reason
=
reason
,
reason
=
reason
,
old_graph
=
old_graph_str
,
old_graph
=
old_graph_str
,
new_graph
=
new_graph_str
)
new_graph
=
new_graph_str
)
#####
# Postcondition: the input and output results are in the storage map, nothing more
#####
# Nothing should be in storage map after evaluating each the thunk (specifically the
# last one)
for
r
,
s
in
storage_map
.
iteritems
():
assert
type
(
s
)
is
list
assert
s
[
0
]
is
None
# store our output results to their respective storage lists
for
output
,
storage
in
zip
(
env
.
outputs
,
output_storage
):
storage
[
0
]
=
r_vals
[
output
]
# transfer all inputs back to their respective storage lists
for
r
in
r_vals
:
if
r
.
owner
is
None
:
if
r
in
env
.
inputs
:
assert
storage_map
[
r
]
is
input_storage
[
env
.
inputs
.
index
(
r
)]
storage_map
[
r
][
0
]
=
r_vals
[
r
]
# if an input was destroyed, the destroyed value should be returned
for
r
in
dr_vals
:
if
r
.
owner
is
None
:
assert
r
in
env
.
inputs
#HACK TO LOOK LIKE A REAL DESTRUCTIVE ACTION TOOK PLACE
if
type
(
dr_vals
[
r
][
0
])
is
numpy
.
ndarray
\
and
dr_vals
[
r
][
0
]
.
dtype
==
storage_map
[
r
][
0
]
.
dtype
\
and
dr_vals
[
r
][
0
]
.
shape
==
storage_map
[
r
][
0
]
.
shape
:
if
len
(
dr_vals
[
r
][
0
]
.
shape
):
storage_map
[
r
][
0
][:]
=
dr_vals
[
r
][
0
]
else
:
storage_map
[
r
][
0
]
.
itemset
(
dr_vals
[
r
][
0
])
else
:
storage_map
[
r
][
0
]
=
dr_vals
[
r
][
0
]
#print ""
#print output_storage
#print dr_vals
#print storage_map
###############
# Done f
##############
f
.
allow_gc
=
True
f
.
allow_gc
=
True
return
f
,
[
link
.
Container
(
input
,
storage
)
for
input
,
storage
in
zip
(
env
.
inputs
,
input_storage
)],
\
assert
len
(
env
.
inputs
)
==
len
(
input_storage
)
[
link
.
Container
(
output
,
storage
,
True
)
for
output
,
storage
in
zip
(
env
.
outputs
,
output_storage
)],
\
assert
len
(
env
.
outputs
)
==
len
(
output_storage
)
#print 'make_all returning output', [id(z) for z in output_storage]
return
f
,
[
link
.
Container
(
input
,
storage
,
readonly
=
False
)
for
input
,
storage
in
zip
(
env
.
inputs
,
input_storage
)],
\
[
link
.
Container
(
output
,
storage
,
readonly
=
True
)
for
output
,
storage
in
zip
(
env
.
outputs
,
output_storage
)],
\
thunks_py
,
order
thunks_py
,
order
_NODEFAULT
=
[
'NODEFAULT'
]
_NODEFAULT
=
[
'NODEFAULT'
]
...
@@ -737,6 +857,9 @@ class DebugMode(Mode):
...
@@ -737,6 +857,9 @@ class DebugMode(Mode):
- incomplete `destroy_map` specification (raises `BadDestroyMap`)
- incomplete `destroy_map` specification (raises `BadDestroyMap`)
- an op that returns an illegal value not matching the output Result Type (raises
InvalidValueError)
Each of these exceptions inherits from the more generic `DebugModeError`.
Each of these exceptions inherits from the more generic `DebugModeError`.
If there are no internal errors, this mode behaves like FAST_RUN or FAST_COMPILE, but takes
If there are no internal errors, this mode behaves like FAST_RUN or FAST_COMPILE, but takes
...
...
theano/gof/link.py
浏览文件 @
b031dfe0
...
@@ -163,7 +163,7 @@ class Container(object):
...
@@ -163,7 +163,7 @@ class Container(object):
def
map_storage
(
env
,
order
,
input_storage
,
output_storage
):
def
map_storage
(
env
,
order
,
input_storage
,
output_storage
):
"""Ensure there is storage for inputs, outputs, and interior nodes.
"""Ensure there is storage
(a length-1 list)
for inputs, outputs, and interior nodes.
:param env: The current env. This function uses the inputs and outputs attributes.
:param env: The current env. This function uses the inputs and outputs attributes.
:param order: an iterable over Apply instances (in program running order)
:param order: an iterable over Apply instances (in program running order)
...
...
theano/gof/opt.py
浏览文件 @
b031dfe0
...
@@ -431,7 +431,7 @@ class OpSub(LocalOptimizer):
...
@@ -431,7 +431,7 @@ class OpSub(LocalOptimizer):
new_output
.
tag
=
copy
(
output
.
tag
)
new_output
.
tag
=
copy
(
output
.
tag
)
return
repl
.
outputs
return
repl
.
outputs
def
str
(
self
):
def
__str__
(
self
):
return
"
%
s ->
%
s"
%
(
self
.
op1
,
self
.
op2
)
return
"
%
s ->
%
s"
%
(
self
.
op1
,
self
.
op2
)
...
@@ -444,10 +444,6 @@ class OpRemove(LocalOptimizer):
...
@@ -444,10 +444,6 @@ class OpRemove(LocalOptimizer):
reentrant
=
False
# no nodes are added at all
reentrant
=
False
# no nodes are added at all
def
__init__
(
self
,
op
):
def
__init__
(
self
,
op
):
"""
op1.make_node and op2.make_node must take the same number of
inputs and have the same number of outputs.
"""
self
.
op
=
op
self
.
op
=
op
def
op_key
(
self
):
def
op_key
(
self
):
...
@@ -461,7 +457,7 @@ class OpRemove(LocalOptimizer):
...
@@ -461,7 +457,7 @@ class OpRemove(LocalOptimizer):
return
False
return
False
return
node
.
inputs
return
node
.
inputs
def
str
(
self
):
def
__str__
(
self
):
return
"
%
s(x) -> x"
%
(
self
.
op
)
return
"
%
s(x) -> x"
%
(
self
.
op
)
...
...
theano/gof/type.py
浏览文件 @
b031dfe0
...
@@ -218,6 +218,10 @@ class PureType(object):
...
@@ -218,6 +218,10 @@ class PureType(object):
"""
"""
raise
AbstractFunctionError
()
raise
AbstractFunctionError
()
def
is_valid_value
(
self
,
a
):
"""Required: Return True for any python object `a` that would be a legal value for a Result of this Type"""
raise
AbstractFunctionError
()
def
make_result
(
self
,
name
=
None
):
def
make_result
(
self
,
name
=
None
):
"""Return a new `Result` instance of Type `self`.
"""Return a new `Result` instance of Type `self`.
...
@@ -325,6 +329,9 @@ class Generic(SingletonType):
...
@@ -325,6 +329,9 @@ class Generic(SingletonType):
def
filter
(
self
,
data
,
strict
=
False
):
def
filter
(
self
,
data
,
strict
=
False
):
return
data
return
data
def
is_valid_value
(
self
,
a
):
return
True
def
c_declare
(
self
,
name
,
sub
):
def
c_declare
(
self
,
name
,
sub
):
return
"""
return
"""
PyObject*
%(name)
s;
PyObject*
%(name)
s;
...
@@ -348,6 +355,7 @@ class Generic(SingletonType):
...
@@ -348,6 +355,7 @@ class Generic(SingletonType):
Py_XINCREF(py_
%(name)
s);
Py_XINCREF(py_
%(name)
s);
"""
%
locals
()
"""
%
locals
()
generic
=
Generic
()
generic
=
Generic
()
...
...
theano/scalar/basic.py
浏览文件 @
b031dfe0
...
@@ -52,6 +52,14 @@ class Scalar(Type):
...
@@ -52,6 +52,14 @@ class Scalar(Type):
except
Exception
,
e
:
except
Exception
,
e
:
raise
TypeError
(
"Could not convert
%
s (value=
%
s) to
%
s"
%
(
type
(
data
),
data
,
self
.
dtype
),
e
)
raise
TypeError
(
"Could not convert
%
s (value=
%
s) to
%
s"
%
(
type
(
data
),
data
,
self
.
dtype
),
e
)
def
values_eq_enough
(
self
,
a
,
b
):
return
abs
(
a
-
b
)
/
(
a
+
b
)
<
1e-4
def
is_valid_value
(
self
,
a
):
_a
=
numpy
.
asarray
(
a
)
rval
=
(
_a
.
ndim
==
0
)
and
(
str
(
_a
.
dtype
)
==
self
.
dtype
)
return
rval
def
__eq__
(
self
,
other
):
def
__eq__
(
self
,
other
):
return
type
(
self
)
==
type
(
other
)
and
other
.
dtype
==
self
.
dtype
return
type
(
self
)
==
type
(
other
)
and
other
.
dtype
==
self
.
dtype
...
...
theano/sparse/basic.py
浏览文件 @
b031dfe0
...
@@ -9,6 +9,7 @@ To read about different sparse formats, see U{http://www-users.cs.umn.edu/~saad/
...
@@ -9,6 +9,7 @@ To read about different sparse formats, see U{http://www-users.cs.umn.edu/~saad/
import
sys
,
operator
import
sys
,
operator
import
numpy
import
numpy
from
scipy
import
sparse
from
scipy
import
sparse
import
scipy.sparse
from
..
import
gof
from
..
import
gof
from
..
import
tensor
from
..
import
tensor
...
@@ -185,6 +186,14 @@ class Sparse(gof.Type):
...
@@ -185,6 +186,14 @@ class Sparse(gof.Type):
def
__repr__
(
self
):
def
__repr__
(
self
):
return
"Sparse[
%
s,
%
s]"
%
(
str
(
self
.
dtype
),
str
(
self
.
format
))
return
"Sparse[
%
s,
%
s]"
%
(
str
(
self
.
dtype
),
str
(
self
.
format
))
def
values_eq_enough
(
self
,
a
,
b
,
eps
=
1e-6
):
return
scipy
.
sparse
.
issparse
(
a
)
\
and
scipy
.
sparse
.
issparse
(
b
)
\
and
abs
(
a
-
b
)
.
sum
()
<
(
1e-6
*
a
.
nnz
)
def
is_valid_value
(
self
,
a
):
return
scipy
.
sparse
.
issparse
(
a
)
and
(
a
.
format
==
self
.
format
)
csc_matrix
=
Sparse
(
format
=
'csc'
)
csc_matrix
=
Sparse
(
format
=
'csc'
)
csr_matrix
=
Sparse
(
format
=
'csr'
)
csr_matrix
=
Sparse
(
format
=
'csr'
)
...
...
theano/tensor/basic.py
浏览文件 @
b031dfe0
...
@@ -226,6 +226,17 @@ class Tensor(Type):
...
@@ -226,6 +226,17 @@ class Tensor(Type):
return
type
(
a
)
is
numpy
.
ndarray
and
type
(
b
)
is
numpy
.
ndarray
\
return
type
(
a
)
is
numpy
.
ndarray
and
type
(
b
)
is
numpy
.
ndarray
\
and
(
a
.
shape
==
b
.
shape
)
and
numpy
.
allclose
(
a
,
b
)
and
(
a
.
shape
==
b
.
shape
)
and
numpy
.
allclose
(
a
,
b
)
def
is_valid_value
(
self
,
a
):
rval
=
(
type
(
a
)
is
numpy
.
ndarray
)
and
(
self
.
ndim
==
a
.
ndim
)
\
and
(
str
(
a
.
dtype
)
==
self
.
dtype
)
\
and
all
([((
si
==
1
)
or
not
bi
)
for
si
,
bi
in
zip
(
a
.
shape
,
self
.
broadcastable
)])
if
not
rval
:
print
type
(
a
),(
type
(
a
)
is
numpy
.
ndarray
)
print
a
.
ndim
,
(
self
.
ndim
==
a
.
ndim
)
print
a
.
dtype
,
(
str
(
a
.
dtype
)
==
self
.
dtype
)
print
a
.
shape
,
self
.
broadcastable
,
([(
shp_i
==
1
)
for
shp_i
in
a
.
shape
]
==
self
.
broadcastable
)
return
rval
def
__hash__
(
self
):
def
__hash__
(
self
):
"""Hash equal for same kinds of Tensor"""
"""Hash equal for same kinds of Tensor"""
return
hash
(
self
.
dtype
)
^
hash
(
self
.
broadcastable
)
return
hash
(
self
.
dtype
)
^
hash
(
self
.
broadcastable
)
...
...
theano/tensor/tests/test_basic.py
浏览文件 @
b031dfe0
...
@@ -114,8 +114,8 @@ def make_restet(name, op, expected, checks = {}, good = {}, bad_build = {}, bad_
...
@@ -114,8 +114,8 @@ def make_restet(name, op, expected, checks = {}, good = {}, bad_build = {}, bad_
for
description
,
check
in
self
.
checks
.
items
():
for
description
,
check
in
self
.
checks
.
items
():
if
not
check
(
inputs
,
results
):
if
not
check
(
inputs
,
results
):
self
.
fail
(
"Test
%
s::
%
s: Failed check:
%
s (inputs were
%
s)"
self
.
fail
(
"Test
%
s::
%
s: Failed check:
%
s (inputs were
%
s
, outputs were
%
s
)"
%
(
self
.
op
,
testname
,
description
,
inputs
))
%
(
self
.
op
,
testname
,
description
,
inputs
,
results
))
def
test_bad_build
(
self
):
def
test_bad_build
(
self
):
for
testname
,
inputs
in
self
.
bad_build
.
items
():
for
testname
,
inputs
in
self
.
bad_build
.
items
():
...
@@ -195,8 +195,11 @@ def make_broadcast_restet(op, expected, checks = {}, **kwargs):
...
@@ -195,8 +195,11 @@ def make_broadcast_restet(op, expected, checks = {}, **kwargs):
if
kwargs
[
'inplace'
]:
if
kwargs
[
'inplace'
]:
_expected
=
expected
_expected
=
expected
expected
=
lambda
*
inputs
:
numpy
.
array
(
_expected
(
*
inputs
),
dtype
=
inputs
[
0
]
.
dtype
)
expected
=
lambda
*
inputs
:
numpy
.
array
(
_expected
(
*
inputs
),
dtype
=
inputs
[
0
]
.
dtype
)
checks
=
dict
(
checks
,
def
inplace_check
(
inputs
,
outputs
):
inplace_check
=
lambda
inputs
,
outputs
:
inputs
[
0
]
is
outputs
[
0
])
# this used to be inputs[0] is output[0]
# I changed it so that it was easier to satisfy by the DebugMode
return
numpy
.
all
(
inputs
[
0
]
==
outputs
[
0
])
checks
=
dict
(
checks
,
inplace_check
=
inplace_check
)
#lambda inputs, outputs: numpy.all(inputs[0] == outputs[0]))
del
kwargs
[
'inplace'
]
del
kwargs
[
'inplace'
]
return
make_restet
(
name
,
op
,
expected
,
checks
,
**
kwargs
)
return
make_restet
(
name
,
op
,
expected
,
checks
,
**
kwargs
)
...
...
theano/tensor/tests/test_blas.py
浏览文件 @
b031dfe0
...
@@ -141,9 +141,10 @@ class t_gemm(TestCase):
...
@@ -141,9 +141,10 @@ class t_gemm(TestCase):
"""test that dot args can be aliased"""
"""test that dot args can be aliased"""
Z
=
value
(
self
.
rand
(
2
,
2
))
Z
=
value
(
self
.
rand
(
2
,
2
))
A
=
value
(
self
.
rand
(
2
,
2
))
A
=
value
(
self
.
rand
(
2
,
2
))
eval_outputs
([
gemm
(
Z
,
1.0
,
A
,
A
,
1.0
)])
f
=
inplace_func
([
A
,
Z
],
gemm
(
Z
,
1.0
,
A
,
A
,
1.0
))
eval_outputs
([
gemm
(
Z
,
1.0
,
A
,
A
.
T
,
1.0
)])
f
(
A
.
data
,
Z
.
data
)
f
=
inplace_func
([
A
,
Z
],
gemm
(
Z
,
1.0
,
A
,
A
.
T
,
1.0
))
f
(
A
.
data
,
Z
.
data
)
def
test_transposes
(
self
):
def
test_transposes
(
self
):
# three square matrices which are not contiguous
# three square matrices which are not contiguous
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论