Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
2e855e05
提交
2e855e05
authored
7月 06, 2011
作者:
James Bergstra
浏览文件
操作
浏览文件
下载
差异文件
merge no conflict
上级
c9b42fdf
2f95561e
全部展开
隐藏空白字符变更
内嵌
并排
正在显示
14 个修改的文件
包含
268 行增加
和
112 行删除
+268
-112
.hgignore
.hgignore
+2
-0
debugmode.py
theano/compile/debugmode.py
+8
-3
function.py
theano/compile/function.py
+11
-2
pfunc.py
theano/compile/pfunc.py
+36
-8
profiling.py
theano/compile/profiling.py
+0
-0
test_modes.py
theano/compile/tests/test_modes.py
+39
-0
__init__.py
theano/gof/__init__.py
+1
-1
env.py
theano/gof/env.py
+24
-5
nvcc_compiler.py
theano/sandbox/cuda/nvcc_compiler.py
+2
-4
basic.py
theano/sparse/basic.py
+76
-59
blas.py
theano/tensor/blas.py
+3
-0
opt.py
theano/tensor/opt.py
+22
-19
test_raw_random.py
theano/tensor/tests/test_raw_random.py
+27
-8
diverse_tests.py
theano/tests/diverse_tests.py
+17
-3
没有找到文件。
.hgignore
浏览文件 @
2e855e05
...
@@ -35,3 +35,5 @@ theano/version.py
...
@@ -35,3 +35,5 @@ theano/version.py
theano/version.py.out
theano/version.py.out
distribute-*.egg
distribute-*.egg
distribute-*.tar.gz
distribute-*.tar.gz
out1
out2
theano/compile/debugmode.py
浏览文件 @
2e855e05
"""Provides `DebugMode`, an evaluation mode for debugging theano internals."""
"""Provides `DebugMode`, an evaluation mode for debugging theano internals.
:TODO: add support for Cond Op, LazyLinker, PureOp, etc.
"""
__docformat__
=
"restructuredtext en"
__docformat__
=
"restructuredtext en"
import
time
,
copy
,
sys
,
copy_reg
,
gc
,
os
import
time
,
copy
,
sys
,
copy_reg
,
gc
,
os
...
@@ -1552,7 +1556,8 @@ class _Maker(FunctionMaker): #inheritance buys a few helper functions
...
@@ -1552,7 +1556,8 @@ class _Maker(FunctionMaker): #inheritance buys a few helper functions
def
__init__
(
self
,
inputs
,
outputs
,
optimizer
,
mode
,
def
__init__
(
self
,
inputs
,
outputs
,
optimizer
,
mode
,
accept_inplace
=
False
,
accept_inplace
=
False
,
function_builder
=
Function
):
function_builder
=
Function
,
profile
=
None
):
"""
"""
:type inputs: a list of SymbolicInput instances
:type inputs: a list of SymbolicInput instances
...
@@ -1567,7 +1572,7 @@ class _Maker(FunctionMaker): #inheritance buys a few helper functions
...
@@ -1567,7 +1572,7 @@ class _Maker(FunctionMaker): #inheritance buys a few helper functions
:note: this function sets TensorType.filter_checks_isfinite when `mode.check_isfinite` is True
:note: this function sets TensorType.filter_checks_isfinite when `mode.check_isfinite` is True
"""
"""
self
.
profile
=
profile
# Handle the case where inputs and/or outputs is a single Variable (not in a list)
# Handle the case where inputs and/or outputs is a single Variable (not in a list)
unpack_single
=
False
unpack_single
=
False
return_none
=
False
return_none
=
False
...
...
theano/compile/function.py
浏览文件 @
2e855e05
...
@@ -7,12 +7,13 @@ _logger = logging.getLogger('theano.compile.function')
...
@@ -7,12 +7,13 @@ _logger = logging.getLogger('theano.compile.function')
from
io
import
In
from
io
import
In
from
function_module
import
orig_function
from
function_module
import
orig_function
from
profiling
import
ProfileStats
from
pfunc
import
pfunc
from
pfunc
import
pfunc
from
numpy
import
any
#for to work in python 2.4
from
numpy
import
any
#for to work in python 2.4
def
function
(
inputs
,
outputs
=
None
,
mode
=
None
,
updates
=
[],
givens
=
[],
def
function
(
inputs
,
outputs
=
None
,
mode
=
None
,
updates
=
[],
givens
=
[],
no_default_updates
=
False
,
accept_inplace
=
False
,
name
=
None
,
no_default_updates
=
False
,
accept_inplace
=
False
,
name
=
None
,
rebuild_strict
=
True
,
allow_input_downcast
=
None
):
rebuild_strict
=
True
,
allow_input_downcast
=
None
,
profile
=
None
):
"""
"""
Return a callable object that will calculate `outputs` from `inputs`.
Return a callable object that will calculate `outputs` from `inputs`.
...
@@ -62,6 +63,11 @@ def function(inputs, outputs=None, mode=None, updates=[], givens=[],
...
@@ -62,6 +63,11 @@ def function(inputs, outputs=None, mode=None, updates=[], givens=[],
precise, type. None (default) is almost like False, but allows
precise, type. None (default) is almost like False, but allows
downcasting of Python float scalars to floatX.
downcasting of Python float scalars to floatX.
:type profile: None, True, or ProfileStats instance
:param profile: accumulate profiling information into a given ProfileStats
instance. If argument is `True` then a new ProfileStats instance will be
used. This profiling object will be available via self.profile.
:note: Regarding givens: Be careful to make sure that these substitutions are
:note: Regarding givens: Be careful to make sure that these substitutions are
independent--behaviour when Var1 of one pair appears in the graph leading to Var2 in
independent--behaviour when Var1 of one pair appears in the graph leading to Var2 in
another expression is undefined. Replacements specified with givens are different from
another expression is undefined. Replacements specified with givens are different from
...
@@ -88,6 +94,8 @@ def function(inputs, outputs=None, mode=None, updates=[], givens=[],
...
@@ -88,6 +94,8 @@ def function(inputs, outputs=None, mode=None, updates=[], givens=[],
if
uses_In
or
uses_tuple
:
if
uses_In
or
uses_tuple
:
# we must use old semantics in this case.
# we must use old semantics in this case.
if
profile
:
raise
NotImplementedError
(
'profiling not supported in old-style function'
)
if
uses_updates
or
uses_givens
:
if
uses_updates
or
uses_givens
:
raise
NotImplementedError
(
"In() instances and tuple inputs triggers the old semantics, which disallow using updates and givens"
)
raise
NotImplementedError
(
"In() instances and tuple inputs triggers the old semantics, which disallow using updates and givens"
)
fn
=
orig_function
(
inputs
,
outputs
,
fn
=
orig_function
(
inputs
,
outputs
,
...
@@ -102,7 +110,8 @@ def function(inputs, outputs=None, mode=None, updates=[], givens=[],
...
@@ -102,7 +110,8 @@ def function(inputs, outputs=None, mode=None, updates=[], givens=[],
no_default_updates
=
no_default_updates
,
no_default_updates
=
no_default_updates
,
accept_inplace
=
accept_inplace
,
name
=
name
,
accept_inplace
=
accept_inplace
,
name
=
name
,
rebuild_strict
=
rebuild_strict
,
rebuild_strict
=
rebuild_strict
,
allow_input_downcast
=
allow_input_downcast
)
allow_input_downcast
=
allow_input_downcast
,
profile
=
profile
)
# We need to add the flag check_aliased inputs if we have any mutable or
# We need to add the flag check_aliased inputs if we have any mutable or
# borrowed used defined inputs
# borrowed used defined inputs
fn
.
_check_for_aliased_inputs
=
check_for_aliased_inputs
fn
.
_check_for_aliased_inputs
=
check_for_aliased_inputs
...
...
theano/compile/pfunc.py
浏览文件 @
2e855e05
"""Provide a simple user friendly API """
"""Provide a simple user friendly API """
__docformat__
=
'restructuredtext en'
__docformat__
=
'restructuredtext en'
import
numpy
# for backport to 2.4, to get any().
from
profiling
import
ProfileStats
from
theano.gof
import
Container
,
Variable
,
generic
,
graph
,
Constant
,
Value
from
theano.gof
import
Container
,
Variable
,
generic
,
graph
,
Constant
,
Value
from
theano.compile
import
orig_function
,
In
,
Out
from
theano.compile
import
orig_function
,
In
,
Out
from
theano.compile.sharedvalue
import
SharedVariable
,
shared
from
theano.compile.sharedvalue
import
SharedVariable
,
shared
import
numpy
# for backport to 2.4, to get any().
from
theano
import
config
def
rebuild_collect_shared
(
outputs
def
rebuild_collect_shared
(
outputs
,
inputs
=
None
,
inputs
=
None
...
@@ -292,7 +295,8 @@ class Param(object):
...
@@ -292,7 +295,8 @@ class Param(object):
def
pfunc
(
params
,
outputs
=
None
,
mode
=
None
,
updates
=
[],
givens
=
[],
def
pfunc
(
params
,
outputs
=
None
,
mode
=
None
,
updates
=
[],
givens
=
[],
no_default_updates
=
False
,
accept_inplace
=
False
,
name
=
None
,
no_default_updates
=
False
,
accept_inplace
=
False
,
name
=
None
,
rebuild_strict
=
True
,
allow_input_downcast
=
None
):
rebuild_strict
=
True
,
allow_input_downcast
=
None
,
profile
=
None
):
"""Function-constructor for graphs with shared variables.
"""Function-constructor for graphs with shared variables.
:type params: list of either Variable or Param instances.
:type params: list of either Variable or Param instances.
...
@@ -319,11 +323,9 @@ def pfunc(params, outputs=None, mode=None, updates=[], givens=[],
...
@@ -319,11 +323,9 @@ def pfunc(params, outputs=None, mode=None, updates=[], givens=[],
If False (default), perform them all. Else, perform automatic updates on all Variables
If False (default), perform them all. Else, perform automatic updates on all Variables
that are neither in "updates" nor in "no_default_updates".
that are neither in "updates" nor in "no_default_updates".
:param name: an optional name for this fct. If used, the profile mode will print the time spent in this fct.
:type name: None or string
:param name: attaches a name to the Profiling result of this function when
:rtype: theano.compile.Function
using ProfileMode (will be deprecated).
:returns: a callable object that will compute the outputs (given the inputs)
and update the implicit function arguments according to the `updates`.
:type allow_input_downcast: Boolean
:type allow_input_downcast: Boolean
:param allow_input_downcast: True means that the values passed as
:param allow_input_downcast: True means that the values passed as
...
@@ -333,6 +335,21 @@ def pfunc(params, outputs=None, mode=None, updates=[], givens=[],
...
@@ -333,6 +335,21 @@ def pfunc(params, outputs=None, mode=None, updates=[], givens=[],
precise, type. None (default) is almost like False, but allows
precise, type. None (default) is almost like False, but allows
downcasting of Python float scalars to floatX.
downcasting of Python float scalars to floatX.
:type profile: None, True, str, or ProfileStats instance
:param profile: accumulate profiling information into a given ProfileStats
instance. None is the default, and means to use the value of
config.profile.
If argument is `True` then a new ProfileStats instance will be
used. If argument is a string, a new ProfileStats instance will be created
with that string as its `message` attribute. This profiling object will be
available via self.profile.
:rtype: theano.compile.Function
:returns: a callable object that will compute the outputs (given the inputs)
and update the implicit function arguments according to the `updates`.
:note: Regarding givens: Be careful to make sure that these substitutions are
:note: Regarding givens: Be careful to make sure that these substitutions are
independent--behaviour when Var1 of one pair appears in the graph leading to Var2 in
independent--behaviour when Var1 of one pair appears in the graph leading to Var2 in
another expression is undefined. Replacements specified with givens are different from
another expression is undefined. Replacements specified with givens are different from
...
@@ -354,6 +371,17 @@ def pfunc(params, outputs=None, mode=None, updates=[], givens=[],
...
@@ -354,6 +371,17 @@ def pfunc(params, outputs=None, mode=None, updates=[], givens=[],
# Then it clones the outputs and the update expressions. This rebuilds a computation graph
# Then it clones the outputs and the update expressions. This rebuilds a computation graph
# from the inputs and the givens.
# from the inputs and the givens.
#
#
if
profile
is
None
:
profile
=
config
.
profile
# profile -> True or False
if
profile
==
True
:
profile
=
ProfileStats
(
message
=
name
)
# profile -> object
if
type
(
profile
)
==
str
:
profile
=
ProfileStats
(
message
=
profile
)
# profile is typically either False or an object at this point.
# No need to block other objects being passed through though. It might be
# useful.
if
not
isinstance
(
params
,(
list
,
tuple
)):
if
not
isinstance
(
params
,(
list
,
tuple
)):
raise
Exception
(
"in pfunc() the first argument must be a list or a tuple"
)
raise
Exception
(
"in pfunc() the first argument must be a list or a tuple"
)
...
@@ -393,7 +421,7 @@ def pfunc(params, outputs=None, mode=None, updates=[], givens=[],
...
@@ -393,7 +421,7 @@ def pfunc(params, outputs=None, mode=None, updates=[], givens=[],
inputs
.
append
(
si
)
inputs
.
append
(
si
)
return
orig_function
(
inputs
,
cloned_outputs
,
mode
,
return
orig_function
(
inputs
,
cloned_outputs
,
mode
,
accept_inplace
=
accept_inplace
,
name
=
name
)
accept_inplace
=
accept_inplace
,
name
=
name
,
profile
=
profile
)
def
_pfunc_param_to_in
(
param
,
strict
=
False
,
allow_downcast
=
None
):
def
_pfunc_param_to_in
(
param
,
strict
=
False
,
allow_downcast
=
None
):
...
...
theano/compile/profiling.py
0 → 100644
浏览文件 @
2e855e05
差异被折叠。
点击展开。
theano/compile/tests/test_modes.py
0 → 100644
浏览文件 @
2e855e05
"""
Test compilation modes
"""
from
nose.plugins.skip
import
SkipTest
import
unittest
import
theano
import
numpy
import
random
import
numpy.random
from
theano.tests
import
unittest_tools
as
utt
class
T_bunch_of_modes
(
unittest
.
TestCase
):
def
test1
(
self
):
# this is a quick test after the LazyLinker branch merge
# to check that all the current modes can still be used.
linker_classes_involved
=
[]
for
modename
in
theano
.
config
.
__class__
.
__dict__
[
'mode'
]
.
all
:
x
=
T
.
matrix
()
y
=
T
.
vector
()
f
=
theano
.
function
([
x
,
y
],
x
+
y
,
mode
=
modename
)
# test that it runs something
f
([[
1
,
2
],[
3
,
4
]],
[
5
,
6
])
linker_classes_involved
.
append
(
f
.
maker
.
mode
.
linker
.
__class__
)
print
'MODE:'
,
modename
,
f
.
maker
.
mode
.
linker
,
'stop'
# regression check:
# there should be
# - VM_Linker
# - OpWiseCLinker (FAST_RUN)
# - WrapLinker (PROFILE_MODE)
# - PerformLinker (FAST_COMPILE)
# - DebugMode's Linker (DEBUG_MODE)
assert
5
==
len
(
set
(
linker_classes_involved
))
if
__name__
==
'__main__'
:
unittest
.
main
()
theano/gof/__init__.py
浏览文件 @
2e855e05
...
@@ -146,7 +146,7 @@ from link import \
...
@@ -146,7 +146,7 @@ from link import \
Container
,
Linker
,
LocalLinker
,
PerformLinker
,
WrapLinker
,
WrapLinkerMany
Container
,
Linker
,
LocalLinker
,
PerformLinker
,
WrapLinker
,
WrapLinkerMany
from
op
import
\
from
op
import
\
Op
Op
,
PureOp
from
opt
import
(
Optimizer
,
optimizer
,
SeqOptimizer
,
from
opt
import
(
Optimizer
,
optimizer
,
SeqOptimizer
,
MergeOptimizer
,
MergeOptMerge
,
MergeOptimizer
,
MergeOptMerge
,
...
...
theano/gof/env.py
浏览文件 @
2e855e05
...
@@ -311,6 +311,9 @@ class Env(utils.object2):
...
@@ -311,6 +311,9 @@ class Env(utils.object2):
self
.
__import_r__
([
new_r
])
self
.
__import_r__
([
new_r
])
self
.
__add_clients__
(
new_r
,
[(
node
,
i
)])
self
.
__add_clients__
(
new_r
,
[(
node
,
i
)])
prune
=
self
.
__remove_clients__
(
r
,
[(
node
,
i
)],
False
)
prune
=
self
.
__remove_clients__
(
r
,
[(
node
,
i
)],
False
)
# Precondition: the substitution is semantically valid
# However it may introduce cycles to the graph, in which case the
# transaction will be reverted later.
self
.
execute_callbacks
(
'on_change_input'
,
node
,
i
,
r
,
new_r
,
reason
=
reason
)
self
.
execute_callbacks
(
'on_change_input'
,
node
,
i
,
r
,
new_r
,
reason
=
reason
)
if
prune
:
if
prune
:
...
@@ -438,16 +441,32 @@ class Env(utils.object2):
...
@@ -438,16 +441,32 @@ class Env(utils.object2):
if
len
(
self
.
nodes
)
<
2
:
if
len
(
self
.
nodes
)
<
2
:
# optimization
# optimization
# when there are 0 or 1 nodes, no sorting is necessary
# when there are 0 or 1 nodes, no sorting is necessary
# This special case happens a lot because the OpWiseCLinker produces
# 1-element graphs.
return
list
(
self
.
nodes
)
return
list
(
self
.
nodes
)
env
=
self
env
=
self
ords
=
{}
ords
=
self
.
orderings
()
for
feature
in
env
.
_features
:
if
hasattr
(
feature
,
'orderings'
):
for
op
,
prereqs
in
feature
.
orderings
(
env
)
.
items
():
ords
.
setdefault
(
op
,
[])
.
extend
(
prereqs
)
order
=
graph
.
io_toposort
(
env
.
inputs
,
env
.
outputs
,
ords
)
order
=
graph
.
io_toposort
(
env
.
inputs
,
env
.
outputs
,
ords
)
return
order
return
order
def
orderings
(
self
):
"""
Return dict d s.t. d[node] is a list of nodes that must be evaluated
before node itself can be evaluated.
This is used primarily by the destroy_handler feature to ensure that all
clients of any destroyed inputs have already computed their outputs.
"""
ords
=
{}
for
feature
in
self
.
_features
:
if
hasattr
(
feature
,
'orderings'
):
for
node
,
prereqs
in
feature
.
orderings
(
self
)
.
items
():
ords
.
setdefault
(
node
,
[])
.
extend
(
prereqs
)
# eliminate duplicate prereqs
for
(
node
,
prereqs
)
in
ords
.
items
():
ords
[
node
]
=
list
(
set
(
prereqs
))
return
ords
def
nclients
(
self
,
r
):
def
nclients
(
self
,
r
):
"""WRITEME Same as len(self.clients(r))."""
"""WRITEME Same as len(self.clients(r))."""
return
len
(
self
.
clients
(
r
))
return
len
(
self
.
clients
(
r
))
...
...
theano/sandbox/cuda/nvcc_compiler.py
浏览文件 @
2e855e05
...
@@ -13,7 +13,7 @@ AddConfigVar('nvcc.compiler_bindir',
...
@@ -13,7 +13,7 @@ AddConfigVar('nvcc.compiler_bindir',
"If defined, nvcc compiler driver will seek g++ and gcc in this directory"
,
"If defined, nvcc compiler driver will seek g++ and gcc in this directory"
,
StrParam
(
""
))
StrParam
(
""
))
AddConfigVar
(
'
cuda.nvcc
flags'
,
AddConfigVar
(
'
nvcc.
flags'
,
"Extra compiler flags for nvcc"
,
"Extra compiler flags for nvcc"
,
StrParam
(
""
))
StrParam
(
""
))
...
@@ -183,11 +183,9 @@ def nvcc_module_compile_str(
...
@@ -183,11 +183,9 @@ def nvcc_module_compile_str(
if
sys
.
platform
!=
'darwin'
:
if
sys
.
platform
!=
'darwin'
:
# the 64bit CUDA libs are in the same files as are named by the function above
# the 64bit CUDA libs are in the same files as are named by the function above
rpaths
.
append
(
os
.
path
.
join
(
config
.
cuda
.
root
,
'lib64'
))
rpaths
.
append
(
os
.
path
.
join
(
config
.
cuda
.
root
,
'lib64'
))
for
rpath
in
rpaths
:
for
rpath
in
rpaths
:
cmd
.
extend
([
'-Xlinker'
,
','
.
join
([
'-rpath'
,
rpath
])])
cmd
.
extend
([
'-Xlinker'
,
','
.
join
([
'-rpath'
,
rpath
])])
nvccflags
=
[
flag
for
flag
in
config
.
cuda
.
nvccflags
.
split
(
' '
)
if
flag
]
cmd
.
extend
([
flag
for
flag
in
config
.
nvcc
.
flags
.
split
(
' '
)
if
flag
])
cmd
.
extend
(
nvccflags
)
cmd
.
extend
(
'-I
%
s'
%
idir
for
idir
in
include_dirs
)
cmd
.
extend
(
'-I
%
s'
%
idir
for
idir
in
include_dirs
)
cmd
.
extend
([
'-o'
,
lib_filename
])
cmd
.
extend
([
'-o'
,
lib_filename
])
cmd
.
append
(
os
.
path
.
split
(
cppfilename
)[
-
1
])
cmd
.
append
(
os
.
path
.
split
(
cppfilename
)[
-
1
])
...
...
theano/sparse/basic.py
浏览文件 @
2e855e05
...
@@ -133,6 +133,79 @@ def sp_ones_like(x):
...
@@ -133,6 +133,79 @@ def sp_ones_like(x):
data
,
indices
,
indptr
,
shape
=
csm_properties
(
x
)
#TODO: don't restrict to CSM formats
data
,
indices
,
indptr
,
shape
=
csm_properties
(
x
)
#TODO: don't restrict to CSM formats
return
CSM
(
format
=
x
.
format
)(
tensor
.
ones_like
(
data
),
indices
,
indptr
,
shape
)
return
CSM
(
format
=
x
.
format
)(
tensor
.
ones_like
(
data
),
indices
,
indptr
,
shape
)
class
_sparse_py_operators
:
T
=
property
(
lambda
self
:
transpose
(
self
),
doc
=
"Return aliased transpose of self (read-only)"
)
def
__neg__
(
self
):
return
neg
(
self
)
def
__add__
(
left
,
right
):
return
add
(
left
,
right
)
def
__radd__
(
right
,
left
):
return
add
(
left
,
right
)
def
__sub__
(
left
,
right
):
return
sub
(
left
,
right
)
def
__rsub__
(
right
,
left
):
return
sub
(
left
,
right
)
def
__mul__
(
left
,
right
):
return
mul
(
left
,
right
)
def
__rmul__
(
left
,
right
):
return
mul
(
left
,
right
)
#extra pseudo-operator symbols
def
__dot__
(
left
,
right
):
return
structured_dot
(
left
,
right
)
def
__rdot__
(
right
,
left
):
return
structured_dot
(
left
,
right
)
#N.B. THIS IS COMMENTED OUT ON PURPOSE!!!
# Discussion with Fred & James (at least, and maybe others before)
# we decided that casting from a sparse to dense should be explicit
# because it's usually something you want to be pretty careful about,
# and not to do by accident.
#def _as_TensorVariable(self):
# return dense_from_sparse(self)
shape
=
property
(
lambda
self
:
tensor
.
shape
(
dense_from_sparse
(
self
)))
# don't worry!
# ... the plan is that the ShapeFeature in tensor.opt will do shape propagation
# ... and remove the dense_from_sparse from the graph. This will *NOT* actually expand
# ... your sparse matrix just to get the shape.
ndim
=
property
(
lambda
self
:
self
.
type
.
ndim
)
dtype
=
property
(
lambda
self
:
self
.
type
.
dtype
)
class
SparseVariable
(
gof
.
Variable
,
_sparse_py_operators
):
dtype
=
property
(
lambda
self
:
self
.
type
.
dtype
)
format
=
property
(
lambda
self
:
self
.
type
.
format
)
def
__str__
(
self
):
return
'
%
s{
%
s,
%
s}'
%
(
self
.
__class__
.
__name__
,
self
.
format
,
self
.
dtype
)
def
__repr__
(
self
):
return
str
(
self
)
class
SparseConstantSignature
(
tuple
):
def
__eq__
(
self
,
other
):
(
a
,
b
),
(
x
,
y
)
=
self
,
other
return
a
==
x
\
and
(
b
.
dtype
==
y
.
dtype
)
\
and
(
type
(
b
)
==
type
(
y
))
\
and
(
b
.
shape
==
y
.
shape
)
\
and
(
abs
(
b
-
y
)
.
sum
()
<
1e-6
*
b
.
nnz
)
def
__hash__
(
self
):
(
a
,
b
)
=
self
return
hash
(
type
(
self
))
^
hash
(
a
)
^
hash
(
type
(
b
))
class
SparseConstant
(
gof
.
Constant
,
_sparse_py_operators
):
dtype
=
property
(
lambda
self
:
self
.
type
.
dtype
)
format
=
property
(
lambda
self
:
self
.
type
.
format
)
def
signature
(
self
):
assert
self
.
data
is
not
None
return
SparseConstantSignature
((
self
.
type
,
self
.
data
))
def
__str__
(
self
):
return
'
%
s{
%
s,
%
s,shape=
%
s,nnz=
%
s}'
%
(
self
.
__class__
.
__name__
,
self
.
format
,
self
.
dtype
,
self
.
data
.
shape
,
self
.
data
.
nnz
)
def
__repr__
(
self
):
return
str
(
self
)
class
SparseValue
(
gof
.
Value
,
_sparse_py_operators
):
dtype
=
property
(
lambda
self
:
self
.
type
.
dtype
)
format
=
property
(
lambda
self
:
self
.
type
.
format
)
class
SparseType
(
gof
.
Type
):
class
SparseType
(
gof
.
Type
):
"""
"""
...
@@ -149,6 +222,9 @@ class SparseType(gof.Type):
...
@@ -149,6 +222,9 @@ class SparseType(gof.Type):
dtype_set
=
set
([
'int'
,
'int8'
,
'int16'
,
'int32'
,
'int64'
,
'float32'
,
'float64'
,
'complex64'
,
'complex128'
])
dtype_set
=
set
([
'int'
,
'int8'
,
'int16'
,
'int32'
,
'int64'
,
'float32'
,
'float64'
,
'complex64'
,
'complex128'
])
ndim
=
2
ndim
=
2
Variable
=
SparseVariable
Constant
=
SparseConstant
def
__init__
(
self
,
format
,
dtype
):
def
__init__
(
self
,
format
,
dtype
):
"""
"""
Fundamental way to create a sparse node.
Fundamental way to create a sparse node.
...
@@ -248,65 +324,6 @@ csr_dmatrix = SparseType(format='csr', dtype='float64')
...
@@ -248,65 +324,6 @@ csr_dmatrix = SparseType(format='csr', dtype='float64')
csc_fmatrix
=
SparseType
(
format
=
'csc'
,
dtype
=
'float32'
)
csc_fmatrix
=
SparseType
(
format
=
'csc'
,
dtype
=
'float32'
)
csr_fmatrix
=
SparseType
(
format
=
'csr'
,
dtype
=
'float32'
)
csr_fmatrix
=
SparseType
(
format
=
'csr'
,
dtype
=
'float32'
)
class
_sparse_py_operators
:
T
=
property
(
lambda
self
:
transpose
(
self
),
doc
=
"Return aliased transpose of self (read-only)"
)
def
__neg__
(
self
):
return
neg
(
self
)
def
__add__
(
left
,
right
):
return
add
(
left
,
right
)
def
__radd__
(
right
,
left
):
return
add
(
left
,
right
)
def
__sub__
(
left
,
right
):
return
sub
(
left
,
right
)
def
__rsub__
(
right
,
left
):
return
sub
(
left
,
right
)
def
__mul__
(
left
,
right
):
return
mul
(
left
,
right
)
def
__rmul__
(
left
,
right
):
return
mul
(
left
,
right
)
#extra pseudo-operator symbols
def
__dot__
(
left
,
right
):
return
structured_dot
(
left
,
right
)
def
__rdot__
(
right
,
left
):
return
structured_dot
(
left
,
right
)
#N.B. THIS IS COMMENTED OUT ON PURPOSE!!!
# Discussion with Fred & James (at least, and maybe others before)
# we decided that casting from a sparse to dense should be explicit
# because it's usually something you want to be pretty careful about,
# and not to do by accident.
#def _as_TensorVariable(self):
# return dense_from_sparse(self)
shape
=
property
(
lambda
self
:
tensor
.
shape
(
dense_from_sparse
(
self
)))
# don't worry!
# ... the plan is that the ShapeFeature in tensor.opt will do shape propagation
# ... and remove the dense_from_sparse from the graph. This will *NOT* actually expand
# ... your sparse matrix just to get the shape.
ndim
=
property
(
lambda
self
:
self
.
type
.
ndim
)
dtype
=
property
(
lambda
self
:
self
.
type
.
dtype
)
class
SparseVariable
(
gof
.
Variable
,
_sparse_py_operators
):
dtype
=
property
(
lambda
self
:
self
.
type
.
dtype
)
format
=
property
(
lambda
self
:
self
.
type
.
format
)
class
SparseConstantSignature
(
tuple
):
def
__eq__
(
self
,
other
):
(
a
,
b
),
(
x
,
y
)
=
self
,
other
return
a
==
x
\
and
(
b
.
dtype
==
y
.
dtype
)
\
and
(
type
(
b
)
==
type
(
y
))
\
and
(
b
.
shape
==
y
.
shape
)
\
and
(
abs
(
b
-
y
)
.
sum
()
<
1e-6
*
b
.
nnz
)
def
__hash__
(
self
):
(
a
,
b
)
=
self
return
hash
(
type
(
self
))
^
hash
(
a
)
^
hash
(
type
(
b
))
class
SparseConstant
(
gof
.
Constant
,
_sparse_py_operators
):
dtype
=
property
(
lambda
self
:
self
.
type
.
dtype
)
format
=
property
(
lambda
self
:
self
.
type
.
format
)
def
signature
(
self
):
assert
self
.
data
is
not
None
return
SparseConstantSignature
((
self
.
type
,
self
.
data
))
class
SparseValue
(
gof
.
Value
,
_sparse_py_operators
):
dtype
=
property
(
lambda
self
:
self
.
type
.
dtype
)
format
=
property
(
lambda
self
:
self
.
type
.
format
)
# CONSTRUCTION
# CONSTRUCTION
class
CSMProperties
(
gof
.
Op
):
class
CSMProperties
(
gof
.
Op
):
"""Extract all of .data .indices and .indptr"""
"""Extract all of .data .indices and .indptr"""
...
...
theano/tensor/blas.py
浏览文件 @
2e855e05
...
@@ -937,6 +937,9 @@ def _gemm_from_node2(node):
...
@@ -937,6 +937,9 @@ def _gemm_from_node2(node):
lst
=
_factor_canonicalized
(
lst
)
lst
=
_factor_canonicalized
(
lst
)
rval
=
_gemm_from_factored_list
(
lst
)
rval
=
_gemm_from_factored_list
(
lst
)
#print "RVAL", rval
#print "RVAL", rval
# THIS GOT COMMENTED OUT AT SOME POINT - ASK P.Lamblin maybe why?
#if rval:
# assert rval[0].type == node.outputs[0].type, (rval[0].type, node.outputs[0].type)
if
rval
and
(
rval
[
0
]
.
type
==
node
.
outputs
[
0
]
.
type
):
if
rval
and
(
rval
[
0
]
.
type
==
node
.
outputs
[
0
]
.
type
):
return
rval
return
rval
...
...
theano/tensor/opt.py
浏览文件 @
2e855e05
...
@@ -3057,30 +3057,33 @@ def constant_folding(node):
...
@@ -3057,30 +3057,33 @@ def constant_folding(node):
for
input
in
node
.
inputs
:
for
input
in
node
.
inputs
:
if
not
isinstance
(
input
,
Constant
):
if
not
isinstance
(
input
,
Constant
):
return
False
return
False
try
:
#condition: all inputs are constant
storage
=
[[
None
]
for
output
in
node
.
outputs
]
node
.
op
.
perform
(
node
,
[
x
.
data
for
x
in
node
.
inputs
],
storage
)
storage_map
=
dict
([(
i
,[
i
.
data
])
for
i
in
node
.
inputs
])
except
MethodNotDefined
:
compute_map
=
dict
([(
i
,[
True
])
for
i
in
node
.
inputs
])
tmp_inputs
=
[
x
.
type
()
for
x
in
node
.
inputs
]
for
o
in
node
.
outputs
:
f
=
compile
.
function
(
storage_map
[
o
]
=
[
None
]
inputs
=
tmp_inputs
,
compute_map
[
o
]
=
[
False
]
outputs
=
node
.
op
.
make_node
(
*
tmp_inputs
)
.
outputs
,
mode
=
compile
.
Mode
(
linker
=
'c|py'
,
optimizer
=
None
))
thunk
=
node
.
op
.
make_thunk
(
node
,
storage_map
,
compute_map
,
xvals
=
f
(
*
[
x
.
data
for
x
in
node
.
inputs
])
no_recycling
=
[])
storage
=
[[
xv
]
for
xv
in
xvals
]
required
=
thunk
()
msg
=
[]
assert
not
required
# a node whose inputs are all provided should always
assert
len
(
storage
)
==
len
(
node
.
outputs
)
# return successfully
for
s
,
output
in
zip
(
storage
,
node
.
outputs
):
rval
=
[]
for
output
in
node
.
outputs
:
assert
compute_map
[
output
][
0
],
(
output
,
storage_map
[
output
][
0
])
try
:
try
:
constant
=
output
.
type
.
Constant
constant
=
output
.
type
.
Constant
except
:
except
AttributeError
:
constant
=
Constant
constant
=
Constant
msg
+=
[
constant
(
output
.
type
,
s
[
0
])]
rval
.
append
(
constant
(
output
.
type
,
storage_map
[
output
][
0
]))
return
msg
return
rval
register_canonicalize
(
constant_folding
,
'fast_compile'
)
register_canonicalize
(
constant_folding
,
'fast_compile'
)
register_stabilize
(
constant_folding
)
# because
register_stabilize
(
constant_folding
)
register_specialize
(
constant_folding
)
register_specialize
(
constant_folding
)
def
_is_1
(
expr
):
def
_is_1
(
expr
):
...
...
theano/tensor/tests/test_raw_random.py
浏览文件 @
2e855e05
...
@@ -49,11 +49,14 @@ class T_random_function(unittest.TestCase):
...
@@ -49,11 +49,14 @@ class T_random_function(unittest.TestCase):
rng_R
=
random_state_type
()
rng_R
=
random_state_type
()
# use make_node to override some of the self.args
# use make_node to override some of the self.args
post_r2
,
out2
=
rf2
(
rng_R
,
(
4
,),
-
2
,
2
)
post_r2
,
out2
=
rf2
(
rng_R
,
(
4
,),
-
2
,
2
)
# NOT INPLACE
post_r2_4
,
out2_4
=
rf2
(
rng_R
,
(
4
,),
-
4.0
,
2
)
post_r4
,
out4
=
rf4
(
rng_R
,
(
4
,),
-
4
,
4
)
# INPLACE
post_r2_4_4
,
out2_4_4
=
rf2
(
rng_R
,
(
4
,),
-
4.0
,
4.0
)
post_r2_4
,
out2_4
=
rf2
(
rng_R
,
(
4
,),
-
4.0
,
2
)
# NOT INPLACE
post_r4
,
out4
=
rf4
(
rng_R
,
(
4
,),
-
4
,
4
)
post_r2_4_4
,
out2_4_4
=
rf2
(
rng_R
,
(
4
,),
-
4.0
,
4.0
)
# NOT INPLACE
# configure out4 to be computed inplace
# The update expression means that the random state rng_R will
# be maintained by post_r4
f
=
compile
.
function
(
f
=
compile
.
function
(
[
compile
.
In
(
rng_R
,
[
compile
.
In
(
rng_R
,
value
=
numpy
.
random
.
RandomState
(
utt
.
fetch_seed
()),
value
=
numpy
.
random
.
RandomState
(
utt
.
fetch_seed
()),
...
@@ -65,9 +68,25 @@ class T_random_function(unittest.TestCase):
...
@@ -65,9 +68,25 @@ class T_random_function(unittest.TestCase):
f2
,
f4
,
f2_4
,
f2_4_4
=
f
()
f2
,
f4
,
f2_4
,
f2_4_4
=
f
()
f2b
,
f4b
,
f2_4b
,
f2_4_4b
=
f
()
f2b
,
f4b
,
f2_4b
,
f2_4_4b
=
f
()
assert
numpy
.
allclose
(
f2
*
2
,
f4
)
print
f2
assert
numpy
.
allclose
(
f2_4_4
,
f4
)
print
f4
assert
not
numpy
.
allclose
(
f4
,
f4b
)
print
f2_4
print
f2_4_4
#print f2b
#print f4b
#print f2_4b
#print f2_4_4b
# setting bounds is same as multiplying by 2
assert
numpy
.
allclose
(
f2
*
2
,
f4
),
(
f2
,
f4
)
# retrieving from non-inplace generator
# is same as inplace one for first call
assert
numpy
.
allclose
(
f2_4_4
,
f4
),
(
f2_4_4
,
f4
)
# f4 changes from call to call, that the update has worked
assert
not
numpy
.
allclose
(
f4
,
f4b
),
(
f4
,
f4b
)
def
test_inplace_optimization
(
self
):
def
test_inplace_optimization
(
self
):
"""Test that FAST_RUN includes the random_make_inplace optimization"""
"""Test that FAST_RUN includes the random_make_inplace optimization"""
...
...
theano/tests/diverse_tests.py
浏览文件 @
2e855e05
...
@@ -13,19 +13,32 @@ from theano.tests import unittest_tools as utt
...
@@ -13,19 +13,32 @@ from theano.tests import unittest_tools as utt
should ensure that it will remain operational
should ensure that it will remain operational
'''
'''
class
T_
diverse
(
unittest
.
TestCase
):
class
T_
scipy
(
unittest
.
TestCase
):
def
setUp
(
self
):
def
setUp
(
self
):
utt
.
seed_rng
()
utt
.
seed_rng
()
self
.
orig_floatX
=
theano
.
config
.
floatX
def
tearDown
(
self
):
theano
.
config
.
floatX
=
self
.
orig_floatX
def
scipy_paper_example1
(
self
):
def
test_
scipy_paper_example1
(
self
):
a
=
theano
.
tensor
.
vector
(
'a'
)
# declare variable
a
=
theano
.
tensor
.
vector
(
'a'
)
# declare variable
b
=
a
+
a
**
10
# build expression
b
=
a
+
a
**
10
# build expression
f
=
theano
.
function
([
a
],
b
)
# compile function
f
=
theano
.
function
([
a
],
b
)
# compile function
assert
numpy
.
all
(
f
([
0
,
1
,
2
])
==
numpy
.
array
([
0
,
2
,
1026
]))
assert
numpy
.
all
(
f
([
0
,
1
,
2
])
==
numpy
.
array
([
0
,
2
,
1026
]))
def
scipy_papa
er_example2
(
self
):
def
test_scipy_pap
er_example2
(
self
):
''' This just sees if things compile well and if they run '''
''' This just sees if things compile well and if they run '''
# PREAMPBLE
T
=
theano
.
tensor
shared
=
theano
.
shared
function
=
theano
.
function
rng
=
numpy
.
random
theano
.
config
.
floatX
=
'float64'
#
# ACTUAL SCRIPT FROM PAPER
x
=
T
.
matrix
()
x
=
T
.
matrix
()
y
=
T
.
vector
()
y
=
T
.
vector
()
w
=
shared
(
rng
.
randn
(
100
))
w
=
shared
(
rng
.
randn
(
100
))
...
@@ -52,6 +65,7 @@ class T_diverse(unittest.TestCase):
...
@@ -52,6 +65,7 @@ class T_diverse(unittest.TestCase):
for
i
in
range
(
training_steps
):
for
i
in
range
(
training_steps
):
pred
,
err
=
train
(
D
[
0
],
D
[
1
])
pred
,
err
=
train
(
D
[
0
],
D
[
1
])
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
unittest
.
main
()
unittest
.
main
()
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论