Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
fd7875ad
提交
fd7875ad
authored
2月 01, 2017
作者:
bscellier
提交者:
GitHub
2月 01, 2017
浏览文件
操作
浏览文件
下载
差异文件
Merge branch 'master' into import_numpy_gpuarray
上级
045cda93
8b9f7336
全部展开
隐藏空白字符变更
内嵌
并排
正在显示
62 个修改的文件
包含
588 行增加
和
374 行删除
+588
-374
jenkins_test1.sh
.jenkins/jenkins_test1.sh
+1
-1
jenkins_test2.sh
.jenkins/jenkins_test2.sh
+1
-1
theano_cache.py
bin/theano_cache.py
+2
-2
how_to_release.txt
doc/internal/how_to_release.txt
+20
-6
shared.txt
doc/library/compile/shared.txt
+28
-5
bn.txt
doc/library/tensor/nnet/bn.txt
+4
-1
builders.py
theano/compile/builders.py
+2
-2
debugmode.py
theano/compile/debugmode.py
+31
-31
function.py
theano/compile/function.py
+2
-2
function_module.py
theano/compile/function_module.py
+25
-14
monitormode.py
theano/compile/monitormode.py
+3
-3
ops.py
theano/compile/ops.py
+6
-6
pfunc.py
theano/compile/pfunc.py
+1
-1
profiling.py
theano/compile/profiling.py
+94
-57
sharedvalue.py
theano/compile/sharedvalue.py
+27
-2
test_builders.py
theano/compile/tests/test_builders.py
+34
-34
test_debugmode.py
theano/compile/tests/test_debugmode.py
+19
-19
test_function.py
theano/compile/tests/test_function.py
+20
-20
test_function_module.py
theano/compile/tests/test_function_module.py
+37
-25
test_misc.py
theano/compile/tests/test_misc.py
+5
-5
test_monitormode.py
theano/compile/tests/test_monitormode.py
+4
-4
test_nanguardmode.py
theano/compile/tests/test_nanguardmode.py
+16
-16
test_ops.py
theano/compile/tests/test_ops.py
+2
-3
test_pfunc.py
theano/compile/tests/test_pfunc.py
+0
-0
test_profiling.py
theano/compile/tests/test_profiling.py
+2
-2
test_shared.py
theano/compile/tests/test_shared.py
+0
-0
configdefaults.py
theano/configdefaults.py
+7
-1
opt.py
theano/gof/opt.py
+8
-4
test_opt.py
theano/gof/tests/test_opt.py
+1
-0
utils.py
theano/gof/utils.py
+2
-0
vm.py
theano/gof/vm.py
+5
-5
dnn.py
theano/gpuarray/dnn.py
+0
-0
dnn_batchnorm.c
theano/gpuarray/dnn_batchnorm.c
+57
-3
dnn_batchnorm_grad.c
theano/gpuarray/dnn_batchnorm_grad.c
+3
-1
dnn_batchnorm_inf.c
theano/gpuarray/dnn_batchnorm_inf.c
+9
-1
test_abstractconv.py
theano/gpuarray/tests/test_abstractconv.py
+4
-0
test_dnn.py
theano/gpuarray/tests/test_dnn.py
+0
-0
check_blas.py
theano/misc/check_blas.py
+11
-11
check_multi_gpu.py
theano/misc/check_multi_gpu.py
+2
-2
latence_gpu_transfert.py
theano/misc/latence_gpu_transfert.py
+3
-3
may_share_memory.py
theano/misc/may_share_memory.py
+3
-3
pkl_utils.py
theano/misc/pkl_utils.py
+8
-8
pycuda_example.py
theano/misc/pycuda_example.py
+5
-5
safe_asarray.py
theano/misc/safe_asarray.py
+3
-3
test_cudamat_utils.py
theano/misc/tests/test_cudamat_utils.py
+5
-5
test_gnumpy_utils.py
theano/misc/tests/test_gnumpy_utils.py
+4
-5
test_may_share_memory.py
theano/misc/tests/test_may_share_memory.py
+3
-3
test_pkl_utils.py
theano/misc/tests/test_pkl_utils.py
+7
-8
test_pycuda_example.py
theano/misc/tests/test_pycuda_example.py
+12
-12
test_pycuda_theano_simple.py
theano/misc/tests/test_pycuda_theano_simple.py
+15
-15
test_pycuda_utils.py
theano/misc/tests/test_pycuda_utils.py
+13
-13
__init__.py
theano/sandbox/cuda/__init__.py
+12
-1
dnn.py
theano/sandbox/cuda/dnn.py
+0
-0
opt.py
theano/sandbox/cuda/opt.py
+0
-0
test_dnn.py
theano/sandbox/cuda/tests/test_dnn.py
+0
-0
var.py
theano/sandbox/cuda/var.py
+0
-0
__init__.py
theano/tensor/nnet/__init__.py
+0
-0
bn.py
theano/tensor/nnet/bn.py
+0
-0
test_abstract_conv.py
theano/tensor/nnet/tests/test_abstract_conv.py
+0
-0
test_bn.py
theano/tensor/nnet/tests/test_bn.py
+0
-0
test_var.py
theano/tensor/tests/test_var.py
+0
-0
var.py
theano/tensor/var.py
+0
-0
没有找到文件。
.jenkins/jenkins_test1.sh
浏览文件 @
fd7875ad
...
...
@@ -13,5 +13,5 @@ echo "===== Testing theano core"
# Test theano core
PARTS
=
"theano -e cuda -e gpuarray"
THEANO_PARAM
=
"
${
PARTS
}
--with-timer --timer-top-n 10 --with-xunit --xunit-file=theanocore_tests.xml"
FLAGS
=
"mode=FAST_RUN,floatX=float32"
FLAGS
=
"mode=FAST_RUN,floatX=float32
,on_opt_error=raise,on_shape_error=raise
"
THEANO_FLAGS
=
${
FLAGS
}
bin/theano-nose
${
THEANO_PARAM
}
.jenkins/jenkins_test2.sh
浏览文件 @
fd7875ad
...
...
@@ -76,5 +76,5 @@ THEANO_GPUARRAY_TESTS="theano/gpuarray/tests \
theano/sandbox/tests/test_rng_mrg.py:test_consistency_GPUA_parallel
\
theano/sandbox/tests/test_rng_mrg.py:test_GPUA_full_fill
\
theano/scan_module/tests/test_scan.py:T_Scan_Gpuarray"
FLAGS
=
"init_gpu_device=
$DEVICE
,gpuarray.preallocate=1000,mode=FAST_RUN"
FLAGS
=
"init_gpu_device=
$DEVICE
,gpuarray.preallocate=1000,mode=FAST_RUN
,on_opt_error=raise,on_shape_error=raise
"
THEANO_FLAGS
=
${
FLAGS
}
time
nosetests
-v
--with-xunit
--xunit-file
=
theanogpuarray_tests.xml
${
THEANO_GPUARRAY_TESTS
}
bin/theano_cache.py
浏览文件 @
fd7875ad
...
...
@@ -5,11 +5,11 @@ import os
import
sys
if
sys
.
platform
==
'win32'
:
config_
cxx
=
'cxx=
'
config_
for_theano_cache_script
=
'cxx=,device=cpu
'
theano_flags
=
os
.
environ
[
'THEANO_FLAGS'
]
if
'THEANO_FLAGS'
in
os
.
environ
else
''
if
theano_flags
:
theano_flags
+=
','
theano_flags
+=
config_
cxx
theano_flags
+=
config_
for_theano_cache_script
os
.
environ
[
'THEANO_FLAGS'
]
=
theano_flags
import
theano
...
...
doc/internal/how_to_release.txt
浏览文件 @
fd7875ad
...
...
@@ -64,11 +64,18 @@ The documentation will be automatically regenerated in the next few hours.
Generate and upload the package
===============================
For release candidates, only upload on PyPI.
On PyPI
-------
Set your umask to ``0022`` to ensure that the package file will be readable from other people.
To check your umask::
umask
To set your umask::
umask 0022
Now change ``ISRELEASED`` in ``setup.py`` to ``True``.
Finally, use setuptools to register and upload the release::
...
...
@@ -84,8 +91,8 @@ UnicodeDecodeError if there are non-ASCII characters in NEWS.txt. You
would need to change NEWS.txt so it contains only ASCII characters (the
problem usually comes from diacritics in people's names).
On mloss.org
------------
On mloss.org
(for final releases only)
------------
--------------------------
Project page is at http://mloss.org/software/view/241/.
Account jaberg is listed as submitter.
...
...
@@ -138,8 +145,10 @@ then run the script.
Announce the release
====================
Generate an e-mail from the template in in ``EMAIL.txt``, including content
from ``NEWS.txt``, and send it to the following mailing lists:
Generate an e-mail from the template in ``EMAIL.txt``, including content
from ``NEWS.txt``.
For final releases, send the e-mail to the following mailing lists:
* theano-users
* theano-announce
...
...
@@ -152,3 +161,8 @@ For release candidates, only e-mail:
* theano-announce
* theano-dev
* theano-users
For alpha and beta releases, only e-mail:
* theano-dev
* theano-users
doc/library/compile/shared.txt
浏览文件 @
fd7875ad
...
...
@@ -19,11 +19,34 @@
The user-friendly constructor is :func:`shared`
.. attribute:: value
Read/write access to the [non-symbolic] value/data associated with this SharedVariable.
Changes to this value will be visible to all functions using this SharedVariable.
.. method:: get_value(self, borrow=False, return_internal_type=False)
:param borrow: True to permit returning of an object aliased to internal memory.
:type borrow: bool
:param return_internal_type: True to permit the returning of an arbitrary type object used
internally to store the shared variable.
:type return_internal_type: bool
By default, return a copy of the data. If ``borrow=True`` (and
``return_internal_type=False``), maybe it will return a copy.
For tensor, it will always return a ndarray by default, so if
the data is on the GPU, it will return a copy, but if the data
is on the CPU, it will return the original data. If you do
``borrow=True`` and ``return_internal_type=True``, it will
always return the original data, not a copy, but this can be a
GPU object.
.. method:: set_value(self, new_value, borrow=False)
:param new_value: The new value.
:type new_value: A compatible type for this shared variable.
:param borrow: True to use the new_value directly, potentially creating problems
related to aliased memory.
:type borrow: bool
The new value will be seen by all functions using this SharedVariable.
.. method:: __init__(self, name, type, value, strict, container=None)
...
...
doc/library/tensor/nnet/bn.txt
浏览文件 @
fd7875ad
...
...
@@ -10,6 +10,9 @@
.. moduleauthor:: LISA
.. seealso:: cuDNN batch normalization: :class:`theano.gpuarray.dnn.dnn_batch_normalization_train`, :class:`theano.gpuarray.dnn.dnn_batch_normalization_test>`. They must be added manually as they do not have the same user interface.
.. autofunction:: theano.tensor.nnet.bn.batch_normalization_train
.. autofunction:: theano.tensor.nnet.bn.batch_normalization_test
.. seealso:: cuDNN batch normalization: :class:`theano.gpuarray.dnn.dnn_batch_normalization_train`, :class:`theano.gpuarray.dnn.dnn_batch_normalization_test>`.
.. autofunction:: theano.tensor.nnet.bn.batch_normalization
theano/compile/builders.py
浏览文件 @
fd7875ad
...
...
@@ -59,11 +59,11 @@ class OpFromGraph(gof.Op):
.. code-block:: python
import numpy
import numpy
as np
import theano
from theano import config, function, OpFromGraph, tensor
x, y, z = tensor.scalars('xyz')
s = theano.shared(n
umpy
.random.rand(2, 2).astype(config.floatX))
s = theano.shared(n
p
.random.rand(2, 2).astype(config.floatX))
e = x + y * z + s
op = OpFromGraph([x, y, z], [e])
# op behaves like a normal theano op
...
...
theano/compile/debugmode.py
浏览文件 @
fd7875ad
...
...
@@ -14,7 +14,7 @@ import six.moves.copyreg as copyreg
from
itertools
import
chain
,
product
as
itertools_product
from
theano.compat
import
izip
import
numpy
import
numpy
as
np
import
theano
from
theano
import
gof
,
config
...
...
@@ -270,15 +270,15 @@ class BadOptimization(DebugModeError):
print
(
" New Value: "
,
str
(
self
.
new_r_val
),
file
=
sio
)
try
:
ov
=
n
umpy
.
asarray
(
self
.
old_r_val
)
nv
=
n
umpy
.
asarray
(
self
.
new_r_val
)
ov
=
n
p
.
asarray
(
self
.
old_r_val
)
nv
=
n
p
.
asarray
(
self
.
new_r_val
)
ssio
=
StringIO
()
abs_diff
=
n
umpy
.
absolute
(
nv
-
ov
)
print
(
" Max Abs Diff: "
,
n
umpy
.
max
(
abs_diff
),
file
=
ssio
)
print
(
" Mean Abs Diff: "
,
n
umpy
.
mean
(
abs_diff
),
file
=
ssio
)
print
(
" Median Abs Diff: "
,
n
umpy
.
median
(
abs_diff
),
file
=
ssio
)
print
(
" Std Abs Diff: "
,
n
umpy
.
std
(
abs_diff
),
file
=
ssio
)
arg_max_val
=
n
umpy
.
argmax
(
abs_diff
)
abs_diff
=
n
p
.
absolute
(
nv
-
ov
)
print
(
" Max Abs Diff: "
,
n
p
.
max
(
abs_diff
),
file
=
ssio
)
print
(
" Mean Abs Diff: "
,
n
p
.
mean
(
abs_diff
),
file
=
ssio
)
print
(
" Median Abs Diff: "
,
n
p
.
median
(
abs_diff
),
file
=
ssio
)
print
(
" Std Abs Diff: "
,
n
p
.
std
(
abs_diff
),
file
=
ssio
)
arg_max_val
=
n
p
.
argmax
(
abs_diff
)
values_at_max
=
(
nv
.
flatten
()[
arg_max_val
],
ov
.
flatten
()[
arg_max_val
])
print
(
" Value at Max Diff: "
,
values_at_max
,
file
=
ssio
)
...
...
@@ -286,13 +286,13 @@ class BadOptimization(DebugModeError):
# N.B. the maximum(..., 1e-8) protects against div by 0 when
# nv == ov == 0
reldiff
=
(
abs_diff
/
n
umpy
.
maaximum
(
numpy
.
absolute
(
nv
)
+
numpy
.
absolute
(
ov
),
1e-8
))
print
(
" Max Rel Diff: "
,
n
umpy
.
max
(
reldiff
),
file
=
ssio
)
print
(
" Mean Rel Diff: "
,
n
umpy
.
mean
(
reldiff
),
file
=
ssio
)
print
(
" Median Rel Diff: "
,
n
umpy
.
median
(
reldiff
),
file
=
ssio
)
print
(
" Std Rel Diff: "
,
n
umpy
.
std
(
reldiff
),
file
=
ssio
)
arg_max_val
=
n
umpy
.
argmax
(
reldiff
)
n
p
.
maximum
(
np
.
absolute
(
nv
)
+
np
.
absolute
(
ov
),
1e-8
))
print
(
" Max Rel Diff: "
,
n
p
.
max
(
reldiff
),
file
=
ssio
)
print
(
" Mean Rel Diff: "
,
n
p
.
mean
(
reldiff
),
file
=
ssio
)
print
(
" Median Rel Diff: "
,
n
p
.
median
(
reldiff
),
file
=
ssio
)
print
(
" Std Rel Diff: "
,
n
p
.
std
(
reldiff
),
file
=
ssio
)
arg_max_val
=
n
p
.
argmax
(
reldiff
)
values_at_max
=
(
nv
.
flatten
()[
arg_max_val
],
ov
.
flatten
()[
arg_max_val
])
print
(
" Value at Max Diff: "
,
values_at_max
,
file
=
ssio
)
...
...
@@ -342,8 +342,8 @@ class BadDestroyMap(DebugModeError):
print
(
" repr (old val):"
,
repr
(
self
.
old_val
),
file
=
sio
)
print
(
" repr (new val):"
,
repr
(
self
.
new_val
),
file
=
sio
)
try
:
npy_old_val
=
n
umpy
.
asarray
(
self
.
old_val
)
npy_new_val
=
n
umpy
.
asarray
(
self
.
new_val
)
npy_old_val
=
n
p
.
asarray
(
self
.
old_val
)
npy_new_val
=
n
p
.
asarray
(
self
.
new_val
)
print
(
" value dtype (new <space> old):"
,
npy_new_val
.
dtype
,
npy_old_val
.
dtype
,
file
=
sio
)
print
(
" value shape (new <space> old):"
,
npy_new_val
.
shape
,
...
...
@@ -356,13 +356,13 @@ class BadDestroyMap(DebugModeError):
print
(
" value min (new-old):"
,
delta
.
min
(),
file
=
sio
)
print
(
" value max (new-old):"
,
delta
.
max
(),
file
=
sio
)
print
(
" value argmin (new-old):"
,
n
umpy
.
unravel_index
(
delta
.
argmin
(),
npy_new_val
.
shape
),
n
p
.
unravel_index
(
delta
.
argmin
(),
npy_new_val
.
shape
),
file
=
sio
)
print
(
" value argmax (new-old):"
,
n
umpy
.
unravel_index
(
delta
.
argmax
(),
npy_new_val
.
shape
),
n
p
.
unravel_index
(
delta
.
argmax
(),
npy_new_val
.
shape
),
file
=
sio
)
print
(
" location of first 10 mismatches:"
,
n
umpy
.
transpose
(
numpy
.
nonzero
(
delta
))[:
10
],
file
=
sio
)
n
p
.
transpose
(
np
.
nonzero
(
delta
))[:
10
],
file
=
sio
)
print
(
""
,
file
=
sio
)
except
Exception
as
e
:
print
(
"(Numpy-hints failed with:
%
s)"
%
str
(
e
),
file
=
sio
)
...
...
@@ -453,7 +453,7 @@ class InvalidValueError(DebugModeError):
v_dtype
=
v
.
dtype
v_min
=
v
.
min
()
v_max
=
v
.
max
()
v_isfinite
=
n
umpy
.
all
(
numpy
.
isfinite
(
v
))
v_isfinite
=
n
p
.
all
(
np
.
isfinite
(
v
))
except
Exception
:
pass
client_node
=
self
.
client_node
...
...
@@ -1025,7 +1025,7 @@ def _lessbroken_deepcopy(a):
# this exists because copy.deepcopy on numpy arrays is broken
# This logic is also in link.py
from
theano.gof.type
import
_cdata_type
if
type
(
a
)
in
(
n
umpy
.
ndarray
,
numpy
.
memmap
):
if
type
(
a
)
in
(
n
p
.
ndarray
,
np
.
memmap
):
rval
=
a
.
copy
()
elif
type
(
a
)
is
_cdata_type
:
# This is not copyable (and should be used for constant data).
...
...
@@ -1034,7 +1034,7 @@ def _lessbroken_deepcopy(a):
rval
=
copy
.
deepcopy
(
a
)
assert
type
(
rval
)
==
type
(
a
),
(
type
(
rval
),
type
(
a
))
if
isinstance
(
rval
,
n
umpy
.
ndarray
):
if
isinstance
(
rval
,
n
p
.
ndarray
):
assert
rval
.
dtype
==
a
.
dtype
return
rval
...
...
@@ -1241,7 +1241,7 @@ def _get_preallocated_maps(node, thunk, prealloc_modes, def_val,
# There is no risk to overwrite inputs, since r does not work
# inplace.
if
isinstance
(
r
.
type
,
(
TensorType
,
CudaNdarrayType
)):
reuse_outputs
[
r
][
...
]
=
n
umpy
.
asarray
(
reuse_outputs
[
r
][
...
]
=
n
p
.
asarray
(
def_val
)
.
astype
(
r
.
type
.
dtype
)
if
reuse_outputs
:
...
...
@@ -1259,7 +1259,7 @@ def _get_preallocated_maps(node, thunk, prealloc_modes, def_val,
new_buf
=
r
.
type
.
value_zeros
(
r_vals
[
r
]
.
shape
)
# CudaNdarray don't have flags field
# assert new_buf.flags["C_CONTIGUOUS"]
new_buf
[
...
]
=
n
umpy
.
asarray
(
def_val
)
.
astype
(
r
.
type
.
dtype
)
new_buf
[
...
]
=
n
p
.
asarray
(
def_val
)
.
astype
(
r
.
type
.
dtype
)
c_cont_outputs
[
r
]
=
new_buf
...
...
@@ -1273,7 +1273,7 @@ def _get_preallocated_maps(node, thunk, prealloc_modes, def_val,
f_cont_outputs
=
{}
for
r
in
considered_outputs
:
if
isinstance
(
r
.
type
,
(
TensorType
,
CudaNdarrayType
)):
new_buf
=
n
umpy
.
zeros
(
new_buf
=
n
p
.
zeros
(
shape
=
r_vals
[
r
]
.
shape
,
dtype
=
r_vals
[
r
]
.
dtype
,
order
=
'F'
)
...
...
@@ -1331,7 +1331,7 @@ def _get_preallocated_maps(node, thunk, prealloc_modes, def_val,
else
:
buf_shape
.
append
(
s
*
2
)
new_buf
=
r
.
type
.
value_zeros
(
buf_shape
)
new_buf
[
...
]
=
n
umpy
.
asarray
(
def_val
)
.
astype
(
r
.
type
.
dtype
)
new_buf
[
...
]
=
n
p
.
asarray
(
def_val
)
.
astype
(
r
.
type
.
dtype
)
init_strided
[
r
]
=
new_buf
# The number of combinations is exponential in the number of
...
...
@@ -1377,7 +1377,7 @@ def _get_preallocated_maps(node, thunk, prealloc_modes, def_val,
r_buf
=
r_buf
[
tuple
(
strides
)][
tuple
(
shapes
)]
assert
r_buf
.
shape
==
r_vals
[
r
]
.
shape
r_buf
[
...
]
=
n
umpy
.
asarray
(
def_val
)
.
astype
(
r_buf
.
dtype
)
r_buf
[
...
]
=
n
p
.
asarray
(
def_val
)
.
astype
(
r_buf
.
dtype
)
strided
[
r
]
=
r_buf
if
strided
:
...
...
@@ -1405,7 +1405,7 @@ def _get_preallocated_maps(node, thunk, prealloc_modes, def_val,
for
s
,
sd
in
zip
(
r_vals
[
r
]
.
shape
,
r_shape_diff
)]
new_buf
=
r
.
type
.
value_zeros
(
out_shape
)
new_buf
[
...
]
=
n
umpy
.
asarray
(
new_buf
[
...
]
=
n
p
.
asarray
(
def_val
)
.
astype
(
r
.
type
.
dtype
)
wrong_size
[
r
]
=
new_buf
...
...
@@ -2261,7 +2261,7 @@ class _Linker(gof.link.LocalLinker):
# HACK TO LOOK LIKE A REAL DESTRUCTIVE ACTION
# TOOK PLACE
if
((
type
(
dr_vals
[
r
][
0
])
in
(
n
umpy
.
ndarray
,
numpy
.
memmap
))
and
(
n
p
.
ndarray
,
np
.
memmap
))
and
(
dr_vals
[
r
][
0
]
.
dtype
==
storage_map
[
r
][
0
]
.
dtype
)
and
(
dr_vals
[
r
][
0
]
.
shape
==
...
...
theano/compile/function.py
浏览文件 @
fd7875ad
...
...
@@ -13,7 +13,7 @@ from six import string_types
from
theano.compile.io
import
In
from
theano.compile.function_module
import
orig_function
from
theano.compile.pfunc
import
pfunc
from
numpy
import
any
import
numpy
as
np
import
warnings
from
theano
import
compat
...
...
@@ -286,7 +286,7 @@ def function(inputs, outputs=None, mode=None, updates=None, givens=None,
"input."
)
# compute some features of the arguments:
uses_tuple
=
any
([
isinstance
(
i
,
(
list
,
tuple
))
for
i
in
inputs
])
uses_tuple
=
np
.
any
([
isinstance
(
i
,
(
list
,
tuple
))
for
i
in
inputs
])
uses_updates
=
bool
(
updates
)
uses_givens
=
bool
(
givens
)
...
...
theano/compile/function_module.py
浏览文件 @
fd7875ad
...
...
@@ -12,13 +12,14 @@ import six.moves.cPickle as pickle
from
itertools
import
chain
import
time
import
warnings
import
numpy
import
numpy
as
np
import
theano
from
theano
import
config
,
gof
from
theano.compat
import
izip
from
theano.gof
import
graph
import
theano.compile.mode
import
theano.compile.profiling
from
theano.compile.io
import
(
In
,
SymbolicInput
,
SymbolicOutput
)
from
theano.compile.ops
import
deep_copy_op
,
view_op
...
...
@@ -663,7 +664,7 @@ class Function(object):
input_storage
=
[
i
.
value
for
i
in
ins
]
# reinitialize new maker and create new function
if
profile
is
None
:
profile
=
config
.
profile
profile
=
config
.
profile
or
config
.
print_global_stats
# profile -> True or False
if
profile
is
True
:
if
name
:
...
...
@@ -749,6 +750,12 @@ class Function(object):
List of outputs on indices/keys from ``output_subset`` or all of them,
if ``output_subset`` is not passed.
"""
def
restore_defaults
():
for
i
,
(
required
,
refeed
,
value
)
in
enumerate
(
self
.
defaults
):
if
refeed
:
if
isinstance
(
value
,
gof
.
Container
):
value
=
value
.
storage
[
0
]
self
[
i
]
=
value
profile
=
self
.
profile
t0
=
time
.
time
()
...
...
@@ -804,6 +811,7 @@ class Function(object):
e
.
args
=
(
"Bad input "
+
argument_name
+
" to "
+
function_name
+
" at index
%
d (0-based).
%
s"
%
(
i
,
where
),)
+
e
.
args
restore_defaults
()
raise
s
.
provided
+=
1
i
+=
1
...
...
@@ -829,9 +837,9 @@ class Function(object):
in
args_share_memory
[
j
]],
[
self
.
input_storage
[
k
]
.
storage
[
0
]
for
k
in
args_share_memory
[
j
]])
if
n
umpy
.
any
([(
var
.
type
is
i_var
.
type
and
var
.
type
.
may_share_memory
(
val
,
i_val
))
for
(
var
,
val
)
in
group_j
]):
if
n
p
.
any
([(
var
.
type
is
i_var
.
type
and
var
.
type
.
may_share_memory
(
val
,
i_val
))
for
(
var
,
val
)
in
group_j
]):
is_aliased
=
True
args_share_memory
[
j
]
.
append
(
i
)
...
...
@@ -853,14 +861,17 @@ class Function(object):
if
not
self
.
trust_input
:
for
c
in
self
.
input_storage
:
if
c
.
required
and
not
c
.
provided
:
restore_defaults
()
raise
TypeError
(
"Missing required input:
%
s"
%
getattr
(
self
.
inv_finder
[
c
],
'variable'
,
self
.
inv_finder
[
c
]))
if
c
.
provided
>
1
:
restore_defaults
()
raise
TypeError
(
"Multiple values for input:
%
s"
%
getattr
(
self
.
inv_finder
[
c
],
'variable'
,
self
.
inv_finder
[
c
]))
if
c
.
implicit
and
c
.
provided
>
0
:
restore_defaults
()
raise
TypeError
(
'Tried to provide value for implicit input:
%
s'
%
getattr
(
self
.
inv_finder
[
c
],
'variable'
,
...
...
@@ -873,6 +884,7 @@ class Function(object):
self
.
fn
()
if
output_subset
is
None
else
\
self
.
fn
(
output_subset
=
output_subset
)
except
Exception
:
restore_defaults
()
if
hasattr
(
self
.
fn
,
'position_of_error'
):
# this is a new vm-provided function or c linker
# they need this because the exception manipulation
...
...
@@ -925,11 +937,7 @@ class Function(object):
outputs
=
outputs
[:
self
.
n_returned_outputs
]
# Put default values back in the storage
for
i
,
(
required
,
refeed
,
value
)
in
enumerate
(
self
.
defaults
):
if
refeed
:
if
isinstance
(
value
,
gof
.
Container
):
value
=
value
.
storage
[
0
]
self
[
i
]
=
value
restore_defaults
()
#
# NOTE: This logic needs to be replicated in
# scan.
...
...
@@ -937,6 +945,7 @@ class Function(object):
#
dt_call
=
time
.
time
()
-
t0
theano
.
compile
.
profiling
.
total_fct_exec_time
+=
dt_call
self
.
maker
.
mode
.
call_time
+=
dt_call
if
profile
:
profile
.
fct_callcount
+=
1
...
...
@@ -1019,9 +1028,9 @@ def _pickle_Function(f):
all_data
=
input_storage
+
inputs_data
for
i
,
d_i
in
enumerate
(
all_data
):
for
j
,
d_j
in
enumerate
(
all_data
):
if
((
i
<
j
)
and
isinstance
(
d_i
,
n
umpy
.
ndarray
)
and
isinstance
(
d_j
,
n
umpy
.
ndarray
)):
if
n
umpy
.
may_share_memory
(
d_i
,
d_j
):
if
((
i
<
j
)
and
isinstance
(
d_i
,
n
p
.
ndarray
)
and
isinstance
(
d_j
,
n
p
.
ndarray
)):
if
n
p
.
may_share_memory
(
d_i
,
d_j
):
if
f
.
pickle_aliased_memory_strategy
==
'warn'
:
_logger
.
warning
(
'aliased relationship between '
'Function arguments
%
s,
%
s '
...
...
@@ -1041,7 +1050,7 @@ def _constructor_Function(maker, input_storage, inputs_data):
assert
len
(
f
.
input_storage
)
==
len
(
inputs_data
)
for
container
,
x
in
zip
(
f
.
input_storage
,
inputs_data
):
assert
(
container
.
data
is
x
)
or
\
(
isinstance
(
x
,
n
umpy
.
ndarray
)
and
(
container
.
data
==
x
)
.
all
())
or
\
(
isinstance
(
x
,
n
p
.
ndarray
)
and
(
container
.
data
==
x
)
.
all
())
or
\
(
container
.
data
==
x
)
return
f
...
...
@@ -1466,6 +1475,7 @@ class FunctionMaker(object):
end_optimizer
=
time
.
time
()
opt_time
=
end_optimizer
-
start_optimizer
theano
.
compile
.
profiling
.
total_graph_opt_time
+=
opt_time
if
profile
:
profile
.
optimizer_time
+=
opt_time
if
theano
.
config
.
profile_optimizer
:
...
...
@@ -1655,6 +1665,7 @@ class FunctionMaker(object):
end_linker
=
time
.
time
()
linker_time
=
end_linker
-
start_linker
theano
.
compile
.
profiling
.
total_time_linker
+=
linker_time
_logger
.
debug
(
'Linker took
%
f seconds'
,
linker_time
)
if
self
.
profile
:
self
.
profile
.
linker_time
+=
linker_time
...
...
theano/compile/monitormode.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
# Note: this code was initially copied from the 'pyutools' package by its
# original author, and re-licensed under Theano's license.
import
numpy
import
numpy
as
np
import
theano
from
theano.compile.mode
import
Mode
...
...
@@ -93,8 +93,8 @@ class MonitorMode(Mode):
def
detect_nan
(
i
,
node
,
fn
):
for
output
in
fn
.
outputs
:
if
(
not
isinstance
(
output
[
0
],
n
umpy
.
random
.
RandomState
)
and
n
umpy
.
isnan
(
output
[
0
])
.
any
()):
if
(
not
isinstance
(
output
[
0
],
n
p
.
random
.
RandomState
)
and
n
p
.
isnan
(
output
[
0
])
.
any
()):
print
(
'*** NaN detected ***'
)
theano
.
printing
.
debugprint
(
node
)
print
(
'Inputs :
%
s'
%
[
input
[
0
]
for
input
in
fn
.
inputs
])
...
...
theano/compile/ops.py
浏览文件 @
fd7875ad
...
...
@@ -17,7 +17,7 @@ from six import iteritems, integer_types
from
six.moves
import
xrange
import
numpy
import
numpy
as
np
def
register_view_op_c_code
(
type
,
code
,
version
=
()):
...
...
@@ -338,7 +338,7 @@ class Shape_i(gof.Op):
def
__init__
(
self
,
i
):
# As i will be used in the hash and that ndarray are not hashable,
# we need to convert it to an int as it is hashable.
if
isinstance
(
i
,
n
umpy
.
ndarray
):
if
isinstance
(
i
,
n
p
.
ndarray
):
assert
i
.
dtype
in
theano
.
tensor
.
integer_dtypes
assert
i
==
int
(
i
)
i
=
int
(
i
)
...
...
@@ -665,11 +665,11 @@ class Rebroadcast(gof.Op):
items
=
sorted
(
axis
)
self
.
axis
=
OrderedDict
(
items
)
for
axis
,
broad
in
iteritems
(
self
.
axis
):
if
not
isinstance
(
axis
,
(
n
umpy
.
integer
,
integer_types
)):
if
not
isinstance
(
axis
,
(
n
p
.
integer
,
integer_types
)):
raise
TypeError
(
"Rebroadcast needs integer axes. "
"Got {}"
.
format
(
axis
))
if
not
isinstance
(
broad
,
(
n
umpy
.
bool_
,
bool
)):
if
not
isinstance
(
broad
,
(
n
p
.
bool_
,
bool
)):
raise
TypeError
(
"Rebroadcast needs bool for new broadcast "
"pattern. Got {}"
.
format
(
broad
))
...
...
@@ -835,8 +835,8 @@ class SpecifyShape(gof.Op):
x
,
shape
=
inp
out
,
=
out_
assert
x
.
ndim
==
shape
.
size
assert
n
umpy
.
all
(
x
.
shape
==
shape
),
(
"got shape"
,
x
.
shape
,
"expected"
,
shape
)
assert
n
p
.
all
(
x
.
shape
==
shape
),
(
"got shape"
,
x
.
shape
,
"expected"
,
shape
)
out
[
0
]
=
x
def
infer_shape
(
self
,
node
,
shapes
):
...
...
theano/compile/pfunc.py
浏览文件 @
fd7875ad
...
...
@@ -364,7 +364,7 @@ def pfunc(params, outputs=None, mode=None, updates=None, givens=None,
if
givens
is
None
:
givens
=
[]
if
profile
is
None
:
profile
=
config
.
profile
profile
=
config
.
profile
or
config
.
print_global_stats
# profile -> True or False
if
profile
is
False
:
profile
=
None
...
...
theano/compile/profiling.py
浏览文件 @
fd7875ad
...
...
@@ -27,7 +27,7 @@ import sys
import
time
from
collections
import
defaultdict
import
numpy
import
numpy
as
np
import
theano
from
six
import
iteritems
...
...
@@ -36,6 +36,9 @@ from theano.gof import graph
logger
=
logging
.
getLogger
(
'theano.compile.profiling'
)
theano_imported_time
=
time
.
time
()
total_fct_exec_time
=
0.
total_graph_opt_time
=
0.
total_time_linker
=
0.
config
=
theano
.
config
_atexit_print_list
=
[]
...
...
@@ -47,7 +50,80 @@ def _atexit_print_fn():
Print ProfileStat objects in _atexit_print_list to _atexit_print_file.
"""
to_sum
=
[]
if
config
.
profile
:
to_sum
=
[]
if
config
.
profiling
.
destination
==
'stderr'
:
destination_file
=
sys
.
stderr
elif
config
.
profiling
.
destination
==
'stdout'
:
destination_file
=
sys
.
stdout
else
:
destination_file
=
open
(
config
.
profiling
.
destination
,
'w'
)
# Reverse sort in the order of compile+exec time
for
ps
in
sorted
(
_atexit_print_list
,
key
=
lambda
a
:
a
.
compile_time
+
a
.
fct_call_time
)[::
-
1
]:
if
ps
.
fct_callcount
>=
1
or
ps
.
compile_time
>
1
:
ps
.
summary
(
file
=
destination_file
,
n_ops_to_print
=
config
.
profiling
.
n_ops
,
n_apply_to_print
=
config
.
profiling
.
n_apply
)
if
not
isinstance
(
ps
,
ScanProfileStats
):
to_sum
.
append
(
ps
)
else
:
# TODO print the name if there is one!
print
(
'Skipping empty Profile'
)
if
len
(
to_sum
)
>
1
:
# Make a global profile
cum
=
copy
.
copy
(
to_sum
[
0
])
msg
=
(
"Sum of all(
%
d) printed profiles at exit excluding Scan op"
" profile."
%
len
(
to_sum
))
cum
.
message
=
msg
for
ps
in
to_sum
[
1
:]:
for
attr
in
[
"compile_time"
,
"fct_call_time"
,
"fct_callcount"
,
"vm_call_time"
,
"optimizer_time"
,
"linker_time"
,
"validate_time"
,
"import_time"
,
"linker_node_make_thunks"
]:
setattr
(
cum
,
attr
,
getattr
(
cum
,
attr
)
+
getattr
(
ps
,
attr
))
# merge dictonary
for
attr
in
[
"apply_time"
,
"apply_callcount"
,
"apply_cimpl"
,
"variable_shape"
,
"variable_strides"
,
"linker_make_thunk_time"
]:
cum_attr
=
getattr
(
cum
,
attr
)
for
key
,
val
in
iteritems
(
getattr
(
ps
,
attr
)):
assert
key
not
in
cum_attr
cum_attr
[
key
]
=
val
if
cum
.
optimizer_profile
and
ps
.
optimizer_profile
:
try
:
merge
=
cum
.
optimizer_profile
[
0
]
.
merge_profile
(
cum
.
optimizer_profile
[
1
],
ps
.
optimizer_profile
[
1
])
assert
len
(
merge
)
==
len
(
cum
.
optimizer_profile
[
1
])
cum
.
optimizer_profile
=
(
cum
.
optimizer_profile
[
0
],
merge
)
except
Exception
as
e
:
print
(
"Got an exception while merging profile"
)
print
(
e
)
cum
.
optimizer_profile
=
None
else
:
cum
.
optimizer_profile
=
None
cum
.
summary
(
file
=
destination_file
,
n_ops_to_print
=
config
.
profiling
.
n_ops
,
n_apply_to_print
=
config
.
profiling
.
n_apply
)
if
config
.
print_global_stats
:
print_global_stats
()
def
print_global_stats
():
"""
Print the following stats:
-- Time elapsed since Theano was imported
-- Time spent inside Theano functions
-- Time spent in compiling Theano functions
-- on graph optimization
-- on linker
"""
if
config
.
profiling
.
destination
==
'stderr'
:
destination_file
=
sys
.
stderr
...
...
@@ -56,57 +132,18 @@ def _atexit_print_fn():
else
:
destination_file
=
open
(
config
.
profiling
.
destination
,
'w'
)
# Reverse sort in the order of compile+exec time
for
ps
in
sorted
(
_atexit_print_list
,
key
=
lambda
a
:
a
.
compile_time
+
a
.
fct_call_time
)[::
-
1
]:
if
ps
.
fct_callcount
>=
1
or
ps
.
compile_time
>
1
:
ps
.
summary
(
file
=
destination_file
,
n_ops_to_print
=
config
.
profiling
.
n_ops
,
n_apply_to_print
=
config
.
profiling
.
n_apply
)
if
not
isinstance
(
ps
,
ScanProfileStats
):
to_sum
.
append
(
ps
)
else
:
# TODO print the name if there is one!
print
(
'Skipping empty Profile'
)
if
len
(
to_sum
)
>
1
:
# Make a global profile
cum
=
copy
.
copy
(
to_sum
[
0
])
msg
=
(
"Sum of all(
%
d) printed profiles at exit excluding Scan op"
" profile."
%
len
(
to_sum
))
cum
.
message
=
msg
for
ps
in
to_sum
[
1
:]:
for
attr
in
[
"compile_time"
,
"fct_call_time"
,
"fct_callcount"
,
"vm_call_time"
,
"optimizer_time"
,
"linker_time"
,
"validate_time"
,
"import_time"
,
"linker_node_make_thunks"
]:
setattr
(
cum
,
attr
,
getattr
(
cum
,
attr
)
+
getattr
(
ps
,
attr
))
# merge dictonary
for
attr
in
[
"apply_time"
,
"apply_callcount"
,
"apply_cimpl"
,
"variable_shape"
,
"variable_strides"
,
"linker_make_thunk_time"
]:
cum_attr
=
getattr
(
cum
,
attr
)
for
key
,
val
in
iteritems
(
getattr
(
ps
,
attr
)):
assert
key
not
in
cum_attr
cum_attr
[
key
]
=
val
if
cum
.
optimizer_profile
and
ps
.
optimizer_profile
:
try
:
merge
=
cum
.
optimizer_profile
[
0
]
.
merge_profile
(
cum
.
optimizer_profile
[
1
],
ps
.
optimizer_profile
[
1
])
assert
len
(
merge
)
==
len
(
cum
.
optimizer_profile
[
1
])
cum
.
optimizer_profile
=
(
cum
.
optimizer_profile
[
0
],
merge
)
except
Exception
as
e
:
print
(
"Got an exception while merging profile"
)
print
(
e
)
cum
.
optimizer_profile
=
None
else
:
cum
.
optimizer_profile
=
None
cum
.
summary
(
file
=
destination_file
,
n_ops_to_print
=
config
.
profiling
.
n_ops
,
n_apply_to_print
=
config
.
profiling
.
n_apply
)
print
(
'='
*
50
,
file
=
destination_file
)
print
(
'Global stats: '
,
'Time elasped since Theano import =
%6.3
fs, '
'Time spent in Theano functions =
%6.3
fs, '
'Time spent compiling Theano functions: '
' optimzation =
%6.3
fs, linker =
%6.3
fs '
%
(
time
.
time
()
-
theano_imported_time
,
total_fct_exec_time
,
total_graph_opt_time
,
total_time_linker
),
file
=
destination_file
)
print
(
'='
*
50
,
file
=
destination_file
)
class
ProfileStats
(
object
):
...
...
@@ -440,7 +477,7 @@ class ProfileStats(object):
hs
+=
[
'<#apply>'
]
es
+=
[
'
%4
d '
]
upto_length
=
n
umpy
.
sum
([
len
(
x
)
for
x
in
hs
])
+
len
(
hs
)
upto_length
=
n
p
.
sum
([
len
(
x
)
for
x
in
hs
])
+
len
(
hs
)
maxlen
=
max
(
self
.
line_width
-
upto_length
,
0
)
hs
+=
[
'<Class name>'
]
es
+=
[
'
%
s'
]
...
...
@@ -522,7 +559,7 @@ class ProfileStats(object):
hs
+=
[
'<#apply>'
]
es
+=
[
'
%4
d '
]
upto_length
=
n
umpy
.
sum
([
len
(
x
)
for
x
in
hs
])
+
len
(
hs
)
upto_length
=
n
p
.
sum
([
len
(
x
)
for
x
in
hs
])
+
len
(
hs
)
maxlen
=
max
(
self
.
line_width
-
upto_length
,
0
)
hs
+=
[
'<Op name>'
]
es
+=
[
'
%
s'
]
...
...
@@ -590,7 +627,7 @@ class ProfileStats(object):
if
self
.
variable_shape
:
hs
+=
[
'<Mflops>'
,
'<Gflops/s>'
]
upto_length
=
n
umpy
.
sum
([
len
(
x
)
for
x
in
hs
])
+
len
(
hs
)
upto_length
=
n
p
.
sum
([
len
(
x
)
for
x
in
hs
])
+
len
(
hs
)
maxlen
=
max
(
self
.
line_width
-
upto_length
,
0
)
hs
+=
[
'<Apply name>'
]
es
+=
[
'
%
s'
]
...
...
@@ -892,7 +929,7 @@ class ProfileStats(object):
node_list
=
list
(
node_list
)
mem_count
=
0
max_mem_count
=
0
mem_bound
=
n
umpy
.
inf
mem_bound
=
n
p
.
inf
# This take only the inputs/outputs dependencies.
dependencies
=
fgraph
.
profile
.
dependencies
done_set
=
set
([])
...
...
theano/compile/sharedvalue.py
浏览文件 @
fd7875ad
...
...
@@ -9,7 +9,7 @@ import copy
import
logging
# Third-party imports
import
numpy
import
numpy
as
np
# Theano imports
from
theano.gof
import
Container
,
Variable
,
generic
,
utils
...
...
@@ -120,6 +120,31 @@ class SharedVariable(Variable):
Changes to this value will be visible to all functions using
this SharedVariable.
Notes
-----
Set_value will work in-place on the GPU, if
the following conditions are met:
* The destination on the GPU must be c_contiguous.
* The source is on the CPU.
* The old value must have the same dtype as the new value
(which is a given for now, since only float32 is
supported).
* The old and new value must have the same shape.
* The old value is being completely replaced by the new
value (not partially modified, e.g. by replacing some
subtensor of it).
* You change the value of the shared variable via
set_value, not via the .value accessors. You should not
use the .value accessors anyway, since they will soon be
deprecated and removed.
It is also worth mentioning that, for efficient transfer to the GPU,
Theano will make the new data ``c_contiguous``. This can require an
extra copy of the data on the host.
The inplace on gpu memory work when borrow is either True or False.
"""
if
borrow
:
self
.
container
.
value
=
new_value
...
...
@@ -162,7 +187,7 @@ class SharedVariable(Variable):
# implemented at all, but with a more explicit error message to help
# Theano users figure out the root of the problem more easily.
value
=
self
.
get_value
(
borrow
=
True
)
if
isinstance
(
value
,
n
umpy
.
ndarray
):
if
isinstance
(
value
,
n
p
.
ndarray
):
# Array probably had an unknown dtype.
msg
=
(
"a Numpy array with dtype: '
%
s'. This data type is not "
"currently recognized by Theano tensors: please cast "
...
...
theano/compile/tests/test_builders.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
from
theano
import
config
,
shared
...
...
@@ -23,14 +23,14 @@ class T_OpFromGraph(unittest_tools.InferShapeTester):
f
=
op
(
x
,
y
,
z
)
-
op
(
y
,
z
,
x
)
fn
=
function
([
x
,
y
,
z
],
f
)
xv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
xv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
# print function, function.__module__
# print fn.maker.fgraph.toposort()
fn
(
xv
,
yv
,
zv
)
assert
n
umpy
.
all
(
8.0
==
fn
(
xv
,
yv
,
zv
))
assert
n
umpy
.
all
(
8.0
==
fn
(
xv
,
yv
,
zv
))
assert
n
p
.
all
(
8.0
==
fn
(
xv
,
yv
,
zv
))
assert
n
p
.
all
(
8.0
==
fn
(
xv
,
yv
,
zv
))
def
test_size_changes
(
self
):
x
,
y
,
z
=
T
.
matrices
(
'xyz'
)
...
...
@@ -38,15 +38,15 @@ class T_OpFromGraph(unittest_tools.InferShapeTester):
op
=
OpFromGraph
([
x
,
y
],
[
e
])
f
=
op
(
x
,
op
(
y
,
z
))
fn
=
function
([
x
,
y
,
z
],
f
)
xv
=
n
umpy
.
ones
((
2
,
3
),
dtype
=
config
.
floatX
)
yv
=
n
umpy
.
ones
((
3
,
4
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
umpy
.
ones
((
4
,
5
),
dtype
=
config
.
floatX
)
*
5
xv
=
n
p
.
ones
((
2
,
3
),
dtype
=
config
.
floatX
)
yv
=
n
p
.
ones
((
3
,
4
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
p
.
ones
((
4
,
5
),
dtype
=
config
.
floatX
)
*
5
res
=
fn
(
xv
,
yv
,
zv
)
assert
res
.
shape
==
(
2
,
5
)
assert
n
umpy
.
all
(
180.0
==
res
)
assert
n
p
.
all
(
180.0
==
res
)
res
=
fn
(
xv
,
yv
,
zv
)
assert
res
.
shape
==
(
2
,
5
)
assert
n
umpy
.
all
(
180.0
==
res
)
assert
n
p
.
all
(
180.0
==
res
)
def
test_grad
(
self
):
x
,
y
,
z
=
T
.
matrices
(
'xyz'
)
...
...
@@ -55,10 +55,10 @@ class T_OpFromGraph(unittest_tools.InferShapeTester):
f
=
op
(
x
,
y
,
z
)
f
=
f
-
T
.
grad
(
T
.
sum
(
f
),
y
)
fn
=
function
([
x
,
y
,
z
],
f
)
xv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
assert
n
umpy
.
all
(
11.0
==
fn
(
xv
,
yv
,
zv
))
xv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
assert
n
p
.
all
(
11.0
==
fn
(
xv
,
yv
,
zv
))
def
test_grad_grad
(
self
):
x
,
y
,
z
=
T
.
matrices
(
'xyz'
)
...
...
@@ -68,47 +68,47 @@ class T_OpFromGraph(unittest_tools.InferShapeTester):
f
=
f
-
T
.
grad
(
T
.
sum
(
f
),
y
)
f
=
f
-
T
.
grad
(
T
.
sum
(
f
),
y
)
fn
=
function
([
x
,
y
,
z
],
f
)
xv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
assert
n
umpy
.
allclose
(
6.0
,
fn
(
xv
,
yv
,
zv
))
xv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
assert
n
p
.
allclose
(
6.0
,
fn
(
xv
,
yv
,
zv
))
def
test_shared
(
self
):
x
,
y
,
z
=
T
.
matrices
(
'xyz'
)
s
=
shared
(
n
umpy
.
random
.
rand
(
2
,
2
)
.
astype
(
config
.
floatX
))
s
=
shared
(
n
p
.
random
.
rand
(
2
,
2
)
.
astype
(
config
.
floatX
))
e
=
x
+
y
*
z
+
s
op
=
OpFromGraph
([
x
,
y
,
z
],
[
e
])
# (1+3*5=array of 16) - (3+1*5=array of 8)
f
=
op
(
x
,
y
,
z
)
-
op
(
y
,
z
,
x
)
fn
=
function
([
x
,
y
,
z
],
f
)
xv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
xv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
# print function, function.__module__
# print fn.maker.fgraph.toposort()
assert
n
umpy
.
allclose
(
8.0
,
fn
(
xv
,
yv
,
zv
))
assert
n
umpy
.
allclose
(
8.0
,
fn
(
xv
,
yv
,
zv
))
assert
n
p
.
allclose
(
8.0
,
fn
(
xv
,
yv
,
zv
))
assert
n
p
.
allclose
(
8.0
,
fn
(
xv
,
yv
,
zv
))
def
test_shared_grad
(
self
):
x
,
y
,
z
=
T
.
matrices
(
'xyz'
)
s
=
shared
(
n
umpy
.
random
.
rand
(
2
,
2
)
.
astype
(
config
.
floatX
))
s
=
shared
(
n
p
.
random
.
rand
(
2
,
2
)
.
astype
(
config
.
floatX
))
e
=
x
+
y
*
z
+
s
op
=
OpFromGraph
([
x
,
y
,
z
],
[
e
])
f
=
op
(
x
,
y
,
z
)
f
=
f
-
T
.
grad
(
T
.
sum
(
f
),
y
)
fn
=
function
([
x
,
y
,
z
],
f
)
xv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
umpy
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
assert
n
umpy
.
allclose
(
11.0
+
s
.
get_value
(),
fn
(
xv
,
yv
,
zv
))
xv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
yv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
3
zv
=
n
p
.
ones
((
2
,
2
),
dtype
=
config
.
floatX
)
*
5
assert
n
p
.
allclose
(
11.0
+
s
.
get_value
(),
fn
(
xv
,
yv
,
zv
))
# grad again the shared variable
f
=
op
(
x
,
y
,
z
)
f
=
f
-
T
.
grad
(
T
.
sum
(
f
),
s
)
fn
=
function
([
x
,
y
,
z
],
f
)
assert
n
umpy
.
allclose
(
15.0
+
s
.
get_value
(),
fn
(
xv
,
yv
,
zv
))
assert
n
p
.
allclose
(
15.0
+
s
.
get_value
(),
fn
(
xv
,
yv
,
zv
))
def
test_connection_pattern
(
self
):
# Basic case
...
...
@@ -163,6 +163,6 @@ class T_OpFromGraph(unittest_tools.InferShapeTester):
p
=
T
.
matrix
(
'p'
)
self
.
_compile_and_check
([
q
,
p
],
op_graph
(
q
,
p
),
[
n
umpy
.
ones
([
3
,
4
],
dtype
=
config
.
floatX
),
n
umpy
.
ones
([
3
,
4
],
dtype
=
config
.
floatX
)],
[
n
p
.
ones
([
3
,
4
],
dtype
=
config
.
floatX
),
n
p
.
ones
([
3
,
4
],
dtype
=
config
.
floatX
)],
OpFromGraph
)
theano/compile/tests/test_debugmode.py
浏览文件 @
fd7875ad
...
...
@@ -2,7 +2,7 @@ from __future__ import absolute_import, print_function, division
from
nose.plugins.skip
import
SkipTest
import
unittest
import
numpy
import
numpy
as
np
from
theano
import
config
from
theano
import
gof
...
...
@@ -316,7 +316,7 @@ def test_just_c_code():
x
=
theano
.
tensor
.
dvector
()
f
=
theano
.
function
([
x
],
wb2
(
x
),
mode
=
debugmode
.
DebugMode
(
check_py_code
=
False
))
assert
n
umpy
.
all
(
f
([
1
,
2
])
==
[
2
,
4
])
assert
n
p
.
all
(
f
([
1
,
2
])
==
[
2
,
4
])
def
test_baddestroymap
():
...
...
@@ -349,7 +349,7 @@ def test_baddestroymap_c():
f
=
theano
.
function
([
x
],
wb2i
(
x
),
mode
=
debugmode
.
DebugMode
(
check_py_code
=
False
))
try
:
assert
n
umpy
.
all
(
f
([
1
,
2
])
==
[
2
,
4
])
assert
n
p
.
all
(
f
([
1
,
2
])
==
[
2
,
4
])
assert
False
# failed to raise error
except
debugmode
.
BadDestroyMap
:
pass
...
...
@@ -445,8 +445,8 @@ class Test_ViewMap(unittest.TestCase):
r0
,
r1
=
f
([
1
,
2
,
3
,
4
],
[
5
,
6
,
7
,
8
])
assert
n
umpy
.
all
(
r0
==
[
1
,
2
,
3
,
4
])
assert
n
umpy
.
all
(
r1
==
[
2
,
3
,
4
])
assert
n
p
.
all
(
r0
==
[
1
,
2
,
3
,
4
])
assert
n
p
.
all
(
r1
==
[
2
,
3
,
4
])
def
test_aliased_outputs_ok_output
(
self
):
# here aliased outputs is ok because they are both outputs of the
...
...
@@ -470,8 +470,8 @@ class Test_ViewMap(unittest.TestCase):
r0
,
r1
=
f
([
1
,
2
,
3
,
4
],
[
5
,
6
,
7
,
8
])
assert
n
umpy
.
all
(
r0
==
[
2
,
4
,
6
,
8
])
assert
n
umpy
.
all
(
r1
==
[
4
,
6
,
8
])
assert
n
p
.
all
(
r0
==
[
2
,
4
,
6
,
8
])
assert
n
p
.
all
(
r1
==
[
4
,
6
,
8
])
def
test_aliased_outputs_ok_shadow
(
self
):
# here the alias between outputs is ok because one of them is not used
...
...
@@ -496,7 +496,7 @@ class Test_ViewMap(unittest.TestCase):
r0
=
f
([
1
,
2
,
3
,
4
],
[
5
,
6
,
7
,
8
])
assert
n
umpy
.
all
(
r0
==
[
2
,
4
,
6
,
8
])
assert
n
p
.
all
(
r0
==
[
2
,
4
,
6
,
8
])
def
test_aliased_outputs_bad
(
self
):
# here the alias between outputs is not ok because destroying one
...
...
@@ -555,31 +555,31 @@ class Test_check_isfinite(unittest.TestCase):
g
=
theano
.
function
([
x
],
theano
.
tensor
.
log
(
x
),
mode
=
'DEBUG_MODE'
)
# this should work
f
(
n
umpy
.
log
([
3
,
4
,
5
])
.
astype
(
config
.
floatX
))
f
(
n
p
.
log
([
3
,
4
,
5
])
.
astype
(
config
.
floatX
))
# if TensorType.filter_checks_isfinite were true, these would raise
# ValueError
# if not, DebugMode will check internally, and raise InvalidValueError
# passing an invalid value as an input should trigger ValueError
self
.
assertRaises
(
debugmode
.
InvalidValueError
,
f
,
n
umpy
.
log
([
3
,
-
4
,
5
])
.
astype
(
config
.
floatX
))
n
p
.
log
([
3
,
-
4
,
5
])
.
astype
(
config
.
floatX
))
self
.
assertRaises
(
debugmode
.
InvalidValueError
,
f
,
(
n
umpy
.
asarray
([
0
,
1.0
,
0
])
/
0
)
.
astype
(
config
.
floatX
))
(
n
p
.
asarray
([
0
,
1.0
,
0
])
/
0
)
.
astype
(
config
.
floatX
))
self
.
assertRaises
(
debugmode
.
InvalidValueError
,
f
,
(
n
umpy
.
asarray
([
1.0
,
1.0
,
1.0
])
/
0
)
.
astype
(
config
.
floatX
))
(
n
p
.
asarray
([
1.0
,
1.0
,
1.0
])
/
0
)
.
astype
(
config
.
floatX
))
# generating an invalid value internally should trigger
# InvalidValueError
self
.
assertRaises
(
debugmode
.
InvalidValueError
,
g
,
n
umpy
.
asarray
([
3
,
-
4
,
5
],
dtype
=
config
.
floatX
))
n
p
.
asarray
([
3
,
-
4
,
5
],
dtype
=
config
.
floatX
))
# this should disable the exception
theano
.
tensor
.
TensorType
.
filter_checks_isfinite
=
False
theano
.
compile
.
mode
.
predefined_modes
[
'DEBUG_MODE'
]
.
check_isfinite
=
False
# insert several Inf
f
(
n
umpy
.
asarray
(
numpy
.
asarray
([
1.0
,
1.0
,
1.0
])
/
0
,
dtype
=
config
.
floatX
))
f
(
n
p
.
asarray
(
np
.
asarray
([
1.0
,
1.0
,
1.0
])
/
0
,
dtype
=
config
.
floatX
))
def
test_check_isfinite_disabled
(
self
):
x
=
theano
.
tensor
.
dvector
()
...
...
@@ -587,10 +587,10 @@ class Test_check_isfinite(unittest.TestCase):
mode
=
debugmode
.
DebugMode
(
check_isfinite
=
False
))
# nan should go through
f
(
n
umpy
.
log
([
3
,
-
4
,
5
]))
f
(
n
p
.
log
([
3
,
-
4
,
5
]))
# inf should go through
infs
=
n
umpy
.
asarray
([
1.0
,
1.
,
1.
])
/
0
infs
=
n
p
.
asarray
([
1.0
,
1.
,
1.
])
/
0
# print infs
f
(
infs
)
return
...
...
@@ -721,14 +721,14 @@ class VecAsRowAndCol(gof.Op):
class
Test_preallocated_output
(
unittest
.
TestCase
):
def
setUp
(
self
):
self
.
rng
=
n
umpy
.
random
.
RandomState
(
seed
=
utt
.
fetch_seed
())
self
.
rng
=
n
p
.
random
.
RandomState
(
seed
=
utt
.
fetch_seed
())
def
test_f_contiguous
(
self
):
a
=
theano
.
tensor
.
fmatrix
(
'a'
)
b
=
theano
.
tensor
.
fmatrix
(
'b'
)
z
=
BrokenCImplementationAdd
()(
a
,
b
)
# In this test, we do not want z to be an output of the graph.
out
=
theano
.
tensor
.
dot
(
z
,
n
umpy
.
eye
(
7
))
out
=
theano
.
tensor
.
dot
(
z
,
n
p
.
eye
(
7
))
a_val
=
self
.
rng
.
randn
(
7
,
7
)
.
astype
(
'float32'
)
b_val
=
self
.
rng
.
randn
(
7
,
7
)
.
astype
(
'float32'
)
...
...
theano/compile/tests/test_function.py
浏览文件 @
fd7875ad
...
...
@@ -5,7 +5,7 @@ import shutil
import
tempfile
import
unittest
import
numpy
import
numpy
as
np
import
theano
from
theano.compile.io
import
In
...
...
@@ -27,7 +27,7 @@ def test_function_dump():
fct2
=
theano
.
function
(
**
l
)
x
=
[
1
,
2
,
3
]
assert
n
umpy
.
allclose
(
fct1
(
x
),
fct2
(
x
))
assert
n
p
.
allclose
(
fct1
(
x
),
fct2
(
x
))
class
TestFunctionIn
(
unittest
.
TestCase
):
...
...
@@ -40,14 +40,14 @@ class TestFunctionIn(unittest.TestCase):
f
=
theano
.
function
([
In
(
a
,
strict
=
False
)],
out
)
# works, rand generates float64 by default
f
(
n
umpy
.
random
.
rand
(
8
))
f
(
n
p
.
random
.
rand
(
8
))
# works, casting is allowed
f
(
n
umpy
.
array
([
1
,
2
,
3
,
4
],
dtype
=
'int32'
))
f
(
n
p
.
array
([
1
,
2
,
3
,
4
],
dtype
=
'int32'
))
f
=
theano
.
function
([
In
(
a
,
strict
=
True
)],
out
)
try
:
# fails, f expects float64
f
(
n
umpy
.
array
([
1
,
2
,
3
,
4
],
dtype
=
'int32'
))
f
(
n
p
.
array
([
1
,
2
,
3
,
4
],
dtype
=
'int32'
))
except
TypeError
:
pass
...
...
@@ -70,17 +70,17 @@ class TestFunctionIn(unittest.TestCase):
# using mutable=True will let f change the value in aval
f
=
theano
.
function
([
In
(
a
,
mutable
=
True
)],
a_out
,
mode
=
'FAST_RUN'
)
aval
=
n
umpy
.
random
.
rand
(
10
)
aval
=
n
p
.
random
.
rand
(
10
)
aval2
=
aval
.
copy
()
assert
n
umpy
.
all
(
f
(
aval
)
==
(
aval2
*
2
))
assert
not
n
umpy
.
all
(
aval
==
aval2
)
assert
n
p
.
all
(
f
(
aval
)
==
(
aval2
*
2
))
assert
not
n
p
.
all
(
aval
==
aval2
)
# using mutable=False should leave the input untouched
f
=
theano
.
function
([
In
(
a
,
mutable
=
False
)],
a_out
,
mode
=
'FAST_RUN'
)
aval
=
n
umpy
.
random
.
rand
(
10
)
aval
=
n
p
.
random
.
rand
(
10
)
aval2
=
aval
.
copy
()
assert
n
umpy
.
all
(
f
(
aval
)
==
(
aval2
*
2
))
assert
n
umpy
.
all
(
aval
==
aval2
)
assert
n
p
.
all
(
f
(
aval
)
==
(
aval2
*
2
))
assert
n
p
.
all
(
aval
==
aval2
)
def
test_in_update
(
self
):
a
=
theano
.
tensor
.
dscalar
(
'a'
)
...
...
@@ -115,7 +115,7 @@ class TestFunctionIn(unittest.TestCase):
# changes occur at the same time and one doesn't overwrite the other.
for
i
in
range
(
5
):
f
()
assert
n
umpy
.
allclose
(
shared_var
.
get_value
(),
i
%
2
)
assert
n
p
.
allclose
(
shared_var
.
get_value
(),
i
%
2
)
def
test_in_allow_downcast_int
(
self
):
a
=
theano
.
tensor
.
wvector
(
'a'
)
# int16
...
...
@@ -128,16 +128,16 @@ class TestFunctionIn(unittest.TestCase):
# Both values are in range. Since they're not ndarrays (but lists),
# they will be converted, and their value checked.
assert
n
umpy
.
all
(
f
([
3
],
[
6
],
1
)
==
10
)
assert
n
p
.
all
(
f
([
3
],
[
6
],
1
)
==
10
)
# Values are in range, but a dtype too large has explicitly been given
# For performance reasons, no check of the data is explicitly performed
# (It might be OK to change this in the future.)
self
.
assertRaises
(
TypeError
,
f
,
[
3
],
n
umpy
.
array
([
6
],
dtype
=
'int16'
),
self
.
assertRaises
(
TypeError
,
f
,
[
3
],
n
p
.
array
([
6
],
dtype
=
'int16'
),
1
)
# Value too big for a, silently ignored
assert
n
umpy
.
all
(
f
([
2
**
20
],
numpy
.
ones
(
1
,
dtype
=
'int8'
),
1
)
==
2
)
assert
n
p
.
all
(
f
([
2
**
20
],
np
.
ones
(
1
,
dtype
=
'int8'
),
1
)
==
2
)
# Value too big for b, raises TypeError
self
.
assertRaises
(
TypeError
,
f
,
[
3
],
[
312
],
1
)
...
...
@@ -156,17 +156,17 @@ class TestFunctionIn(unittest.TestCase):
(
a
+
b
+
c
))
# If the values can be accurately represented, everything is OK
assert
n
umpy
.
all
(
f
(
0
,
0
,
0
)
==
0
)
assert
n
p
.
all
(
f
(
0
,
0
,
0
)
==
0
)
# If allow_downcast is True, idem
assert
n
umpy
.
allclose
(
f
(
0.1
,
0
,
0
),
0.1
)
assert
n
p
.
allclose
(
f
(
0.1
,
0
,
0
),
0.1
)
# If allow_downcast is False, nope
self
.
assertRaises
(
TypeError
,
f
,
0
,
0.1
,
0
)
# If allow_downcast is None, it should work iff floatX=float32
if
theano
.
config
.
floatX
==
'float32'
:
assert
n
umpy
.
allclose
(
f
(
0
,
0
,
0.1
),
0.1
)
assert
n
p
.
allclose
(
f
(
0
,
0
,
0.1
),
0.1
)
else
:
self
.
assertRaises
(
TypeError
,
f
,
0
,
0
,
0.1
)
...
...
@@ -182,10 +182,10 @@ class TestFunctionIn(unittest.TestCase):
# If the values can be accurately represented, everything is OK
z
=
[
0
]
assert
n
umpy
.
all
(
f
(
z
,
z
,
z
)
==
0
)
assert
n
p
.
all
(
f
(
z
,
z
,
z
)
==
0
)
# If allow_downcast is True, idem
assert
n
umpy
.
allclose
(
f
([
0.1
],
z
,
z
),
0.1
)
assert
n
p
.
allclose
(
f
([
0.1
],
z
,
z
),
0.1
)
# If allow_downcast is False, nope
self
.
assertRaises
(
TypeError
,
f
,
z
,
[
0.1
],
z
)
...
...
theano/compile/tests/test_function_module.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
import
copy
import
six.moves.cPickle
as
pickle
import
numpy
import
numpy
as
np
import
unittest
...
...
@@ -18,8 +18,6 @@ from theano import tensor
from
theano
import
tensor
as
T
import
theano
import
numpy
as
N
def
PatternOptimizer
(
p1
,
p2
,
ign
=
True
):
return
gof
.
OpKeyOptimizer
(
gof
.
PatternSub
(
p1
,
p2
),
ignore_newtrees
=
ign
)
...
...
@@ -281,7 +279,7 @@ class T_function(unittest.TestCase):
def
test_swap_SharedVariable
(
self
):
i
=
T
.
iscalar
()
x_list
=
theano
.
shared
(
value
=
n
umpy
.
random
.
rand
(
10
)
.
astype
(
config
.
floatX
))
x_list
=
theano
.
shared
(
value
=
n
p
.
random
.
rand
(
10
)
.
astype
(
config
.
floatX
))
x
=
T
.
scalar
(
'x'
)
# SharedVariable for tests, one of them has update
...
...
@@ -343,11 +341,11 @@ class T_function(unittest.TestCase):
A special testcase for logistic_sgd.py in Deep Learning Tutorial
This test assert that SharedVariable in different function have same storage
"""
train_x
=
theano
.
shared
(
value
=
n
umpy
.
random
.
rand
(
10
,
10
)
.
astype
(
config
.
floatX
))
test_x
=
theano
.
shared
(
value
=
n
umpy
.
random
.
rand
(
10
,
10
)
.
astype
(
config
.
floatX
))
train_x
=
theano
.
shared
(
value
=
n
p
.
random
.
rand
(
10
,
10
)
.
astype
(
config
.
floatX
))
test_x
=
theano
.
shared
(
value
=
n
p
.
random
.
rand
(
10
,
10
)
.
astype
(
config
.
floatX
))
train_y
=
theano
.
shared
(
value
=
n
umpy
.
random
.
rand
(
10
,
1
)
.
astype
(
config
.
floatX
))
test_y
=
theano
.
shared
(
value
=
n
umpy
.
random
.
rand
(
10
,
1
)
.
astype
(
config
.
floatX
))
train_y
=
theano
.
shared
(
value
=
n
p
.
random
.
rand
(
10
,
1
)
.
astype
(
config
.
floatX
))
test_y
=
theano
.
shared
(
value
=
n
p
.
random
.
rand
(
10
,
1
)
.
astype
(
config
.
floatX
))
i
=
T
.
iscalar
(
'index'
)
x
=
T
.
vector
(
'x'
)
...
...
@@ -500,42 +498,42 @@ class T_function(unittest.TestCase):
when borrow=True is implemented.
"""
a
=
T
.
dmatrix
()
aval
=
n
umpy
.
random
.
rand
(
3
,
3
)
aval
=
n
p
.
random
.
rand
(
3
,
3
)
# when borrow=False, test that a destroy map cannot alias output to input
f
=
theano
.
function
([
In
(
a
,
borrow
=
False
)],
Out
(
a
+
1
,
borrow
=
True
))
assert
n
umpy
.
all
(
f
(
aval
)
==
aval
+
1
)
assert
not
n
umpy
.
may_share_memory
(
aval
,
f
(
aval
))
assert
n
p
.
all
(
f
(
aval
)
==
aval
+
1
)
assert
not
n
p
.
may_share_memory
(
aval
,
f
(
aval
))
# when borrow=False, test that a viewmap cannot alias output to input
f
=
theano
.
function
([
In
(
a
,
borrow
=
False
)],
Out
(
a
[
0
,
:],
borrow
=
True
))
assert
n
umpy
.
all
(
f
(
aval
)
==
aval
[
0
,
:])
assert
not
n
umpy
.
may_share_memory
(
aval
,
f
(
aval
))
assert
n
p
.
all
(
f
(
aval
)
==
aval
[
0
,
:])
assert
not
n
p
.
may_share_memory
(
aval
,
f
(
aval
))
def
test_borrow_output
(
self
):
a
=
T
.
dmatrix
()
f
=
function
([
a
],
Out
(
a
,
borrow
=
False
))
o
=
N
.
ones
((
3
,
3
))
o
=
np
.
ones
((
3
,
3
))
assert
o
is
not
f
(
o
)
# function no longer permits aliasing outputs to inputs
f
=
function
([
a
],
Out
(
a
*
4
,
borrow
=
False
))
o
=
N
.
ones
((
3
,
3
))
o
=
np
.
ones
((
3
,
3
))
four
=
f
(
o
)
assert
n
umpy
.
all
(
four
==
4
)
assert
n
p
.
all
(
four
==
4
)
f
(
o
+
.
1
)
# should not clobber the memory used to store four
assert
n
umpy
.
all
(
four
==
4
)
assert
n
p
.
all
(
four
==
4
)
f
=
function
([
a
],
Out
(
a
*
4
,
borrow
=
True
),
mode
=
theano
.
Mode
(
'c|py_nogc'
,
'fast_run'
))
o
=
N
.
ones
((
3
,
3
))
o
=
np
.
ones
((
3
,
3
))
four
=
f
(
o
)
assert
n
umpy
.
all
(
four
==
4
)
assert
n
p
.
all
(
four
==
4
)
f
(
o
+
.
1
)
# should clobber the memory used to store four
if
theano
.
config
.
cxx
:
assert
not
n
umpy
.
all
(
four
==
4
)
assert
not
n
p
.
all
(
four
==
4
)
else
:
# The Elemwise.perform method don't reuse memory
# as some numpy version don't support that correctly.
assert
n
umpy
.
all
(
four
==
4
)
assert
n
p
.
all
(
four
==
4
)
def
test_disconnected_input
(
self
):
a
=
T
.
scalar
(
'a'
)
...
...
@@ -579,6 +577,20 @@ class T_function(unittest.TestCase):
if
not
isinstance
(
key
,
theano
.
gof
.
Constant
):
assert
(
val
[
0
]
is
None
)
def
test_default_values
(
self
):
"""
Check that default values are restored
when an exception occurs in interactive mode.
"""
a
,
b
=
T
.
dscalars
(
'a'
,
'b'
)
c
=
a
+
b
func
=
theano
.
function
([
theano
.
In
(
a
,
name
=
'first'
),
theano
.
In
(
b
,
value
=
1
,
name
=
'second'
)],
c
)
x
=
func
(
first
=
1
)
try
:
func
(
second
=
2
)
except
TypeError
:
assert
(
func
(
first
=
1
)
==
x
)
class
T_picklefunction
(
unittest
.
TestCase
):
...
...
@@ -753,7 +765,7 @@ class T_picklefunction(unittest.TestCase):
assert
f2
.
container
[
s
]
.
storage
is
f1
.
container
[
s
]
.
storage
# now put in a function with non-scalar
v_value
=
n
umpy
.
asarray
([
2
,
3
,
4.
],
dtype
=
config
.
floatX
)
v_value
=
n
p
.
asarray
([
2
,
3
,
4.
],
dtype
=
config
.
floatX
)
f3
=
function
([
x
,
In
(
v
,
value
=
v_value
)],
x
+
v
)
list_of_things
.
append
(
f3
)
...
...
@@ -800,13 +812,13 @@ class T_picklefunction(unittest.TestCase):
assert
nl
[
5
](
3
)
==
ol
[
5
](
3
)
assert
nl
[
4
]
.
value
[
nl
[
0
]]
==
6
assert
n
umpy
.
all
(
nl
[
6
][
nl
[
2
]]
==
numpy
.
asarray
([
2
,
3.
,
4
]))
assert
n
p
.
all
(
nl
[
6
][
nl
[
2
]]
==
np
.
asarray
([
2
,
3.
,
4
]))
def
test_broken_pickle_with_shared
(
self
):
saves
=
[]
def
pers_save
(
obj
):
if
isinstance
(
obj
,
n
umpy
.
ndarray
):
if
isinstance
(
obj
,
n
p
.
ndarray
):
saves
.
append
(
obj
)
return
len
(
saves
)
-
1
else
:
...
...
@@ -815,7 +827,7 @@ class T_picklefunction(unittest.TestCase):
def
pers_load
(
id
):
return
saves
[
id
]
b
=
n
umpy
.
random
.
rand
(
5
,
4
)
b
=
n
p
.
random
.
rand
(
5
,
4
)
x
=
theano
.
tensor
.
matrix
()
y
=
theano
.
shared
(
b
)
...
...
theano/compile/tests/test_misc.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
unittest
from
theano.compile.pfunc
import
pfunc
...
...
@@ -20,8 +20,8 @@ class NNet(object):
self
.
input
=
input
self
.
target
=
target
self
.
lr
=
shared
(
lr
,
'learning_rate'
)
self
.
w1
=
shared
(
n
umpy
.
zeros
((
n_hidden
,
n_input
)),
'w1'
)
self
.
w2
=
shared
(
n
umpy
.
zeros
((
n_output
,
n_hidden
)),
'w2'
)
self
.
w1
=
shared
(
n
p
.
zeros
((
n_hidden
,
n_input
)),
'w1'
)
self
.
w2
=
shared
(
n
p
.
zeros
((
n_output
,
n_hidden
)),
'w2'
)
# print self.lr.type
self
.
hidden
=
sigmoid
(
tensor
.
dot
(
self
.
w1
,
self
.
input
))
...
...
@@ -45,7 +45,7 @@ class NNet(object):
class
TestNnet
(
unittest
.
TestCase
):
def
test_nnet
(
self
):
rng
=
n
umpy
.
random
.
RandomState
(
1827
)
rng
=
n
p
.
random
.
RandomState
(
1827
)
data
=
rng
.
rand
(
10
,
4
)
nnet
=
NNet
(
n_input
=
3
,
n_hidden
=
10
)
for
epoch
in
range
(
3
):
...
...
@@ -60,4 +60,4 @@ class TestNnet(unittest.TestCase):
self
.
assertTrue
(
abs
(
mean_cost
-
0.20588975452
)
<
1e-6
)
# Just call functions to make sure they do not crash.
nnet
.
compute_output
(
input
)
nnet
.
output_from_hidden
(
n
umpy
.
ones
(
10
))
nnet
.
output_from_hidden
(
n
p
.
ones
(
10
))
theano/compile/tests/test_monitormode.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
theano
...
...
@@ -12,7 +12,7 @@ def test_detect_nan():
def
detect_nan
(
i
,
node
,
fn
):
for
output
in
fn
.
outputs
:
if
n
umpy
.
isnan
(
output
[
0
])
.
any
():
if
n
p
.
isnan
(
output
[
0
])
.
any
():
print
(
'*** NaN detected ***'
)
theano
.
printing
.
debugprint
(
node
)
print
(
'Inputs :
%
s'
%
[
input
[
0
]
for
input
in
fn
.
inputs
])
...
...
@@ -36,7 +36,7 @@ def test_optimizer():
def
detect_nan
(
i
,
node
,
fn
):
for
output
in
fn
.
outputs
:
if
n
umpy
.
isnan
(
output
[
0
])
.
any
():
if
n
p
.
isnan
(
output
[
0
])
.
any
():
print
(
'*** NaN detected ***'
)
theano
.
printing
.
debugprint
(
node
)
print
(
'Inputs :
%
s'
%
[
input
[
0
]
for
input
in
fn
.
inputs
])
...
...
@@ -65,7 +65,7 @@ def test_not_inplace():
def
detect_nan
(
i
,
node
,
fn
):
for
output
in
fn
.
outputs
:
if
n
umpy
.
isnan
(
output
[
0
])
.
any
():
if
n
p
.
isnan
(
output
[
0
])
.
any
():
print
(
'*** NaN detected ***'
)
theano
.
printing
.
debugprint
(
node
)
print
(
'Inputs :
%
s'
%
[
input
[
0
]
for
input
in
fn
.
inputs
])
...
...
theano/compile/tests/test_nanguardmode.py
浏览文件 @
fd7875ad
...
...
@@ -6,7 +6,7 @@ from __future__ import absolute_import, print_function, division
import
logging
from
nose.tools
import
assert_raises
import
numpy
import
numpy
as
np
from
theano.compile.nanguardmode
import
NanGuardMode
import
theano
...
...
@@ -18,20 +18,20 @@ def test_NanGuardMode():
# intentionally. A working implementation should be able to capture all
# the abnormalties.
x
=
T
.
matrix
()
w
=
theano
.
shared
(
n
umpy
.
random
.
randn
(
5
,
7
)
.
astype
(
theano
.
config
.
floatX
))
w
=
theano
.
shared
(
n
p
.
random
.
randn
(
5
,
7
)
.
astype
(
theano
.
config
.
floatX
))
y
=
T
.
dot
(
x
,
w
)
fun
=
theano
.
function
(
[
x
],
y
,
mode
=
NanGuardMode
(
nan_is_error
=
True
,
inf_is_error
=
True
)
)
a
=
n
umpy
.
random
.
randn
(
3
,
5
)
.
astype
(
theano
.
config
.
floatX
)
infa
=
n
umpy
.
tile
(
(
n
umpy
.
asarray
(
100.
)
**
1000000
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
5
))
nana
=
n
umpy
.
tile
(
n
umpy
.
asarray
(
numpy
.
nan
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
5
))
biga
=
n
umpy
.
tile
(
n
umpy
.
asarray
(
1e20
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
5
))
a
=
n
p
.
random
.
randn
(
3
,
5
)
.
astype
(
theano
.
config
.
floatX
)
infa
=
n
p
.
tile
(
(
n
p
.
asarray
(
100.
)
**
1000000
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
5
))
nana
=
n
p
.
tile
(
n
p
.
asarray
(
np
.
nan
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
5
))
biga
=
n
p
.
tile
(
n
p
.
asarray
(
1e20
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
5
))
fun
(
a
)
# normal values
...
...
@@ -46,14 +46,14 @@ def test_NanGuardMode():
_logger
.
propagate
=
True
# slices
a
=
n
umpy
.
random
.
randn
(
3
,
4
,
5
)
.
astype
(
theano
.
config
.
floatX
)
infa
=
n
umpy
.
tile
(
(
n
umpy
.
asarray
(
100.
)
**
1000000
)
.
astype
(
theano
.
config
.
floatX
),
a
=
n
p
.
random
.
randn
(
3
,
4
,
5
)
.
astype
(
theano
.
config
.
floatX
)
infa
=
n
p
.
tile
(
(
n
p
.
asarray
(
100.
)
**
1000000
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
4
,
5
))
nana
=
n
umpy
.
tile
(
n
umpy
.
asarray
(
numpy
.
nan
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
4
,
5
))
biga
=
n
umpy
.
tile
(
n
umpy
.
asarray
(
1e20
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
4
,
5
))
nana
=
n
p
.
tile
(
n
p
.
asarray
(
np
.
nan
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
4
,
5
))
biga
=
n
p
.
tile
(
n
p
.
asarray
(
1e20
)
.
astype
(
theano
.
config
.
floatX
),
(
3
,
4
,
5
))
x
=
T
.
tensor3
()
y
=
x
[:,
T
.
arange
(
2
),
T
.
arange
(
2
)]
...
...
theano/compile/tests/test_ops.py
浏览文件 @
fd7875ad
...
...
@@ -9,7 +9,6 @@ from theano.tests import unittest_tools as utt
from
theano
import
function
import
theano
from
theano.tensor
import
dmatrix
,
dvector
from
numpy
import
allclose
from
theano.compile
import
as_op
import
pickle
...
...
@@ -34,7 +33,7 @@ class OpDecoratorTests(utt.InferShapeTester):
r
=
fn
([[
1.5
,
5
],
[
2
,
2
]])
r0
=
np
.
array
([
1.5
,
7.5
,
15.
,
30.
])
assert
allclose
(
r
,
r0
),
(
r
,
r0
)
assert
np
.
allclose
(
r
,
r0
),
(
r
,
r0
)
def
test_2arg
(
self
):
x
=
dmatrix
(
'x'
)
...
...
@@ -50,7 +49,7 @@ class OpDecoratorTests(utt.InferShapeTester):
r
=
fn
([[
1.5
,
5
],
[
2
,
2
]],
[
1
,
100
,
2
,
200
])
r0
=
np
.
array
([
2.5
,
107.5
,
17.
,
230.
])
assert
allclose
(
r
,
r0
),
(
r
,
r0
)
assert
np
.
allclose
(
r
,
r0
),
(
r
,
r0
)
def
test_infer_shape
(
self
):
x
=
dmatrix
(
'x'
)
...
...
theano/compile/tests/test_pfunc.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/compile/tests/test_profiling.py
浏览文件 @
fd7875ad
...
...
@@ -6,7 +6,7 @@ from __future__ import absolute_import, print_function, division
import
unittest
import
numpy
import
numpy
as
np
import
theano
from
six.moves
import
StringIO
...
...
@@ -45,7 +45,7 @@ class Test_profiling(unittest.TestCase):
f
=
theano
.
function
(
x
,
z
,
profile
=
p
,
name
=
"test_profiling"
,
mode
=
m
)
inp
=
[
n
umpy
.
arange
(
1024
,
dtype
=
'float32'
)
+
1
for
i
in
range
(
len
(
x
))]
inp
=
[
n
p
.
arange
(
1024
,
dtype
=
'float32'
)
+
1
for
i
in
range
(
len
(
x
))]
f
(
*
inp
)
buf
=
StringIO
()
...
...
theano/compile/tests/test_shared.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/configdefaults.py
浏览文件 @
fd7875ad
...
...
@@ -126,6 +126,12 @@ AddConfigVar(
BoolParam
(
False
,
allow_override
=
False
),
in_c_key
=
False
)
AddConfigVar
(
'print_global_stats'
,
"Print some global statistics (time spent) at the end"
,
BoolParam
(
False
),
in_c_key
=
False
)
class
ContextsParam
(
ConfigParam
):
def
__init__
(
self
):
...
...
@@ -1111,7 +1117,7 @@ AddConfigVar('optdb.position_cutoff',
AddConfigVar
(
'optdb.max_use_ratio'
,
'A ratio that prevent infinite loop in EquilibriumOptimizer.'
,
FloatParam
(
5
),
FloatParam
(
8
),
in_c_key
=
False
)
AddConfigVar
(
'gcc.cxxflags'
,
...
...
theano/gof/opt.py
浏览文件 @
fd7875ad
...
...
@@ -2510,10 +2510,14 @@ class EquilibriumOptimizer(NavigatorOptimizer):
end_nb_nodes
=
len
(
fgraph
.
apply_nodes
)
if
max_use_abort
:
_logger
.
error
(
"EquilibriumOptimizer max'ed out by '
%
s'"
%
opt_name
+
". You can safely raise the current threshold of "
+
"
%
f with the theano flag 'optdb.max_use_ratio'."
%
config
.
optdb
.
max_use_ratio
)
msg
=
(
"EquilibriumOptimizer max'ed out by '
%
s'"
%
opt_name
+
". You can safely raise the current threshold of "
+
"
%
f with the theano flag 'optdb.max_use_ratio'."
%
config
.
optdb
.
max_use_ratio
)
if
theano
.
config
.
on_opt_error
==
'raise'
:
raise
AssertionError
(
msg
)
else
:
_logger
.
error
(
msg
)
fgraph
.
remove_feature
(
change_tracker
)
assert
len
(
loop_process_count
)
==
len
(
loop_timing
)
assert
len
(
loop_process_count
)
==
len
(
global_opt_timing
)
...
...
theano/gof/tests/test_opt.py
浏览文件 @
fd7875ad
...
...
@@ -571,6 +571,7 @@ class TestEquilibrium(object):
opt
.
optimize
(
g
)
assert
str
(
g
)
==
'[Op2(x, y)]'
@theano.configparser.change_flags
(
on_opt_error
=
'ignore'
)
def
test_low_use_ratio
(
self
):
x
,
y
,
z
=
map
(
MyVariable
,
'xyz'
)
e
=
op3
(
op4
(
x
,
y
))
...
...
theano/gof/utils.py
浏览文件 @
fd7875ad
...
...
@@ -503,6 +503,8 @@ def hist(coll):
return
counts
@deprecated
(
"theano.gof.utils"
,
msg
=
"Use a_theano_variable.auto_name instead"
)
def
give_variables_names
(
variables
):
"""
Gives unique names to an iterable of variables. Modifies input.
...
...
theano/gof/vm.py
浏览文件 @
fd7875ad
...
...
@@ -482,7 +482,7 @@ class Stack(VM):
try
:
_
,
dt
=
self
.
run_thunk_of_node
(
current_apply
)
del
_
if
config
.
profile
:
if
config
.
profile
or
config
.
print_global_stats
:
current_idx
=
self
.
node_idx
[
current_apply
]
self
.
call_counts
[
current_idx
]
+=
1
self
.
call_times
[
current_idx
]
+=
dt
...
...
@@ -596,7 +596,7 @@ class Stack(VM):
if
current_apply
.
inputs
[
r
]
.
owner
:
apply_stack
.
append
(
current_apply
.
inputs
[
r
]
.
owner
)
else
:
if
config
.
profile
:
if
config
.
profile
or
config
.
print_global_stats
:
for
(
idx
,
o
)
in
enumerate
(
thunks
[
self
.
node_idx
[
current_apply
]]
.
outputs
):
var
=
self
.
nodes
[
...
...
@@ -757,7 +757,7 @@ class VM_Linker(link.LocalLinker):
associated to self, else, a new VM_Linker associated to fgraph.
"""
if
(
config
.
profile
and
if
(
(
config
.
profile
or
config
.
print_global_stats
)
and
((
hasattr
(
theano
,
'sandbox'
)
and
hasattr
(
theano
.
sandbox
,
'cuda'
)
and
theano
.
sandbox
.
cuda
.
cuda_enabled
)
or
...
...
@@ -856,7 +856,7 @@ class VM_Linker(link.LocalLinker):
pre_call_clear
=
[
storage_map
[
v
]
for
v
in
self
.
no_recycling
]
if
(
self
.
callback
is
not
None
or
self
.
callback_input
is
not
None
or
(
config
.
profile
and
config
.
profile_memory
)
or
(
(
config
.
profile
or
config
.
print_global_stats
)
and
config
.
profile_memory
)
or
(
self
.
allow_partial_eval
and
not
self
.
use_cloop
)):
if
self
.
use_cloop
and
(
self
.
callback
is
not
None
or
...
...
@@ -1086,7 +1086,7 @@ class VM_Linker(link.LocalLinker):
lazy
=
config
.
vm
.
lazy
if
lazy
is
None
:
lazy
=
not
all
([(
not
th
.
lazy
)
for
th
in
thunks
])
if
not
(
lazy
or
(
config
.
profile
and
config
.
profile_memory
)
or
if
not
(
lazy
or
(
(
config
.
profile
or
config
.
print_global_stats
)
and
config
.
profile_memory
)
or
self
.
use_cloop
or
self
.
callback
or
self
.
callback_input
):
for
pair
in
itervalues
(
reallocated_info
):
storage_map
[
pair
[
1
]]
=
storage_map
[
pair
[
0
]]
...
...
theano/gpuarray/dnn.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/gpuarray/dnn_batchnorm.c
浏览文件 @
fd7875ad
...
...
@@ -2,8 +2,19 @@
int
dnn_batchnorm_op
(
PyGpuArrayObject
*
inp
,
PyGpuArrayObject
*
scale
,
PyGpuArrayObject
*
bias
,
npy_float64
epsilon
,
PyGpuArrayObject
**
outp
,
PyGpuArrayObject
**
x_mean
,
PyGpuArrayObject
**
x_invstd
,
cudnnHandle_t
_handle
)
{
npy_float64
running_average_factor
,
#ifdef RUNNING_AVERAGES
PyGpuArrayObject
*
in_running_mean
,
PyGpuArrayObject
*
in_running_var
,
#endif
PyGpuArrayObject
**
outp
,
PyGpuArrayObject
**
x_mean
,
PyGpuArrayObject
**
x_invstd
,
#ifdef RUNNING_AVERAGES
PyGpuArrayObject
**
out_running_mean
,
PyGpuArrayObject
**
out_running_var
,
#endif
cudnnHandle_t
_handle
)
{
PyGpuContextObject
*
c
=
inp
->
context
;
if
(
c_set_tensorNd
(
inp
,
bn_input
)
!=
0
)
...
...
@@ -11,11 +22,19 @@ int dnn_batchnorm_op(PyGpuArrayObject *inp, PyGpuArrayObject *scale,
if
(
c_set_tensorNd
(
scale
,
bn_params
)
!=
0
)
return
1
;
if
(
epsilon
<
1e-5
)
if
(
epsilon
<
1e-5
)
{
PyErr_Format
(
PyExc_ValueError
,
"epsilon must be at least 1e-5, got %f"
,
epsilon
);
return
1
;
}
#ifdef INPLACE_OUTPUT
Py_XDECREF
(
*
outp
);
*
outp
=
inp
;
Py_INCREF
(
*
outp
);
#else
if
(
theano_prep_output
(
outp
,
inp
->
ga
.
nd
,
inp
->
ga
.
dimensions
,
inp
->
ga
.
typecode
,
GA_C_ORDER
,
c
)
!=
0
)
return
1
;
#endif
if
(
theano_prep_output
(
x_mean
,
scale
->
ga
.
nd
,
scale
->
ga
.
dimensions
,
scale
->
ga
.
typecode
,
GA_C_ORDER
,
c
)
!=
0
)
return
1
;
if
(
theano_prep_output
(
x_invstd
,
scale
->
ga
.
nd
,
scale
->
ga
.
dimensions
,
scale
->
ga
.
typecode
,
GA_C_ORDER
,
c
)
!=
0
)
...
...
@@ -24,6 +43,31 @@ int dnn_batchnorm_op(PyGpuArrayObject *inp, PyGpuArrayObject *scale,
if
(
c_set_tensorNd
(
*
outp
,
bn_output
)
!=
0
)
return
1
;
#ifdef RUNNING_AVERAGES
#ifdef INPLACE_RUNNING_MEAN
Py_XDECREF
(
out_running_mean
);
PyGpuArrayObject
*
running_mean
=
in_running_mean
;
Py_INCREF
(
running_mean
);
#else
PyGpuArrayObject
*
running_mean
=
*
out_running_mean
;
running_mean
=
theano_try_copy
(
running_mean
,
in_running_mean
);
if
(
running_mean
==
NULL
)
{
return
1
;
}
#endif
#ifdef INPLACE_RUNNING_VAR
Py_XDECREF
(
out_running_var
);
PyGpuArrayObject
*
running_var
=
in_running_var
;
Py_INCREF
(
running_var
);
#else
PyGpuArrayObject
*
running_var
=
*
out_running_var
;
running_var
=
theano_try_copy
(
running_var
,
in_running_var
);
if
(
running_var
==
NULL
)
{
return
1
;
}
#endif
#endif
{
const
float
falpha
=
1
.;
const
float
fbeta
=
0
.;
...
...
@@ -50,9 +94,15 @@ int dnn_batchnorm_op(PyGpuArrayObject *inp, PyGpuArrayObject *scale,
bn_params
,
PyGpuArray_DEV_DATA
(
scale
),
PyGpuArray_DEV_DATA
(
bias
),
#ifdef RUNNING_AVERAGES
running_average_factor
,
PyGpuArray_DEV_DATA
(
running_mean
),
PyGpuArray_DEV_DATA
(
running_var
),
#else
0
,
NULL
,
// running mean, deliberately unused
NULL
,
// running var, deliberately unused
#endif
epsilon
,
PyGpuArray_DEV_DATA
(
*
x_mean
),
PyGpuArray_DEV_DATA
(
*
x_invstd
)
...
...
@@ -62,6 +112,10 @@ int dnn_batchnorm_op(PyGpuArrayObject *inp, PyGpuArrayObject *scale,
cudnnGetErrorString
(
err
));
return
1
;
}
#ifdef RUNNING_AVERAGES
*
out_running_mean
=
running_mean
;
*
out_running_var
=
running_var
;
#endif
}
return
0
;
}
theano/gpuarray/dnn_batchnorm_grad.c
浏览文件 @
fd7875ad
...
...
@@ -34,8 +34,10 @@ int dnn_batchnorm_grad(PyGpuArrayObject *inp, PyGpuArrayObject *doutp,
if
(
c_set_tensorNd
(
scale
,
bn_params
)
!=
0
)
return
1
;
if
(
epsilon
<
1e-5
)
if
(
epsilon
<
1e-5
)
{
PyErr_Format
(
PyExc_ValueError
,
"epsilon must be at least 1e-5, got %f"
,
epsilon
);
return
1
;
}
if
(
theano_prep_output
(
dinp
,
inp
->
ga
.
nd
,
inp
->
ga
.
dimensions
,
inp
->
ga
.
typecode
,
GA_C_ORDER
,
c
)
!=
0
)
return
1
;
...
...
theano/gpuarray/dnn_batchnorm_inf.c
浏览文件 @
fd7875ad
...
...
@@ -11,11 +11,19 @@ int dnn_batchnorm_op(PyGpuArrayObject *inp, PyGpuArrayObject *scale,
if
(
c_set_tensorNd
(
scale
,
bn_params
)
!=
0
)
return
1
;
if
(
epsilon
<
1e-5
)
if
(
epsilon
<
1e-5
)
{
PyErr_Format
(
PyExc_ValueError
,
"epsilon must be at least 1e-5, got %f"
,
epsilon
);
return
1
;
}
#ifdef INPLACE_OUTPUT
Py_XDECREF
(
*
outp
);
*
outp
=
inp
;
Py_INCREF
(
*
outp
);
#else
if
(
theano_prep_output
(
outp
,
inp
->
ga
.
nd
,
inp
->
ga
.
dimensions
,
inp
->
ga
.
typecode
,
GA_C_ORDER
,
c
)
!=
0
)
return
1
;
#endif
if
(
c_set_tensorNd
(
*
outp
,
bn_output
)
!=
0
)
return
1
;
...
...
theano/gpuarray/tests/test_abstractconv.py
浏览文件 @
fd7875ad
...
...
@@ -252,3 +252,7 @@ class TestDnnConvTypes(test_abstract_conv.TestConvTypes):
self
.
constant_tensor
=
gpuarray
.
array
(
np
.
zeros
((
3
,
5
,
7
,
11
),
dtype
=
'float32'
),
context
=
get_context
(
test_ctx_name
))
class
TestConv2dTranspose
(
test_abstract_conv
.
TestConv2dTranspose
):
mode
=
mode_with_gpu
theano/gpuarray/tests/test_dnn.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/misc/check_blas.py
浏览文件 @
fd7875ad
...
...
@@ -13,7 +13,7 @@ import time
from
optparse
import
OptionParser
import
subprocess
import
numpy
import
numpy
as
np
import
theano
import
theano.tensor
as
T
...
...
@@ -47,10 +47,10 @@ def execute(execute=True, verbose=True, M=2000, N=2000, K=2000,
print
()
print
(
'Numpy config: (used when the Theano flag'
' "blas.ldflags" is empty)'
)
n
umpy
.
show_config
()
print
(
'Numpy dot module:'
,
n
umpy
.
dot
.
__module__
)
print
(
'Numpy location:'
,
n
umpy
.
__file__
)
print
(
'Numpy version:'
,
n
umpy
.
__version__
)
n
p
.
show_config
()
print
(
'Numpy dot module:'
,
n
p
.
dot
.
__module__
)
print
(
'Numpy location:'
,
n
p
.
__file__
)
print
(
'Numpy version:'
,
n
p
.
__version__
)
if
(
theano
.
config
.
device
.
startswith
(
"gpu"
)
or
theano
.
config
.
init_gpu_device
.
startswith
(
"gpu"
)):
print
(
'nvcc version:'
)
...
...
@@ -58,12 +58,12 @@ def execute(execute=True, verbose=True, M=2000, N=2000, K=2000,
"--version"
))
print
()
a
=
theano
.
shared
(
n
umpy
.
ones
((
M
,
N
),
dtype
=
theano
.
config
.
floatX
,
order
=
order
))
b
=
theano
.
shared
(
n
umpy
.
ones
((
N
,
K
),
dtype
=
theano
.
config
.
floatX
,
order
=
order
))
c
=
theano
.
shared
(
n
umpy
.
ones
((
M
,
K
),
dtype
=
theano
.
config
.
floatX
,
order
=
order
))
a
=
theano
.
shared
(
n
p
.
ones
((
M
,
N
),
dtype
=
theano
.
config
.
floatX
,
order
=
order
))
b
=
theano
.
shared
(
n
p
.
ones
((
N
,
K
),
dtype
=
theano
.
config
.
floatX
,
order
=
order
))
c
=
theano
.
shared
(
n
p
.
ones
((
M
,
K
),
dtype
=
theano
.
config
.
floatX
,
order
=
order
))
f
=
theano
.
function
([],
updates
=
[(
c
,
0.4
*
c
+
.
8
*
T
.
dot
(
a
,
b
))])
if
any
([
x
.
op
.
__class__
.
__name__
==
'Gemm'
for
x
in
...
...
theano/misc/check_multi_gpu.py
浏览文件 @
fd7875ad
...
...
@@ -9,7 +9,7 @@ from __future__ import absolute_import, print_function, division
import
threading
import
time
import
numpy
import
numpy
as
np
import
theano
from
theano.gpuarray
import
init_dev
...
...
@@ -21,7 +21,7 @@ def main(dev1, dev2):
init_dev
(
dev2
,
'ctx2'
)
size
=
1024
*
16
data
=
n
umpy
.
random
.
randn
(
size
,
size
)
.
astype
(
'float32'
)
data
=
n
p
.
random
.
randn
(
size
,
size
)
.
astype
(
'float32'
)
val1a
=
theano
.
shared
(
data
,
target
=
'ctx1'
)
val1b
=
theano
.
shared
(
data
,
target
=
'ctx1'
)
val1c
=
theano
.
shared
(
data
,
target
=
'ctx1'
)
...
...
theano/misc/latence_gpu_transfert.py
浏览文件 @
fd7875ad
...
...
@@ -2,18 +2,18 @@ from __future__ import absolute_import, print_function, division
import
time
import
numpy
import
numpy
as
np
import
theano
y
=
theano
.
tensor
.
fvector
()
x
=
theano
.
shared
(
n
umpy
.
zeros
(
1
,
dtype
=
'float32'
))
x
=
theano
.
shared
(
n
p
.
zeros
(
1
,
dtype
=
'float32'
))
f1
=
theano
.
function
([
y
],
updates
=
{
x
:
y
})
f2
=
theano
.
function
([],
theano
.
sandbox
.
cuda
.
host_from_gpu
(
x
))
print
(
f1
.
maker
.
fgraph
.
toposort
())
print
(
f2
.
maker
.
fgraph
.
toposort
())
for
i
in
[
1
,
10
,
100
,
1000
,
10000
,
100000
,
1000000
,
10000000
]:
o
=
n
umpy
.
zeros
(
i
,
dtype
=
'float32'
)
o
=
n
p
.
zeros
(
i
,
dtype
=
'float32'
)
t0
=
time
.
time
()
f1
(
o
)
t1
=
time
.
time
()
...
...
theano/misc/may_share_memory.py
浏览文件 @
fd7875ad
...
...
@@ -4,7 +4,7 @@ numpy version support only ndarray.
"""
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
from
theano.tensor.basic
import
TensorType
try
:
...
...
@@ -42,8 +42,8 @@ else:
def
may_share_memory
(
a
,
b
,
raise_other_type
=
True
):
a_ndarray
=
isinstance
(
a
,
n
umpy
.
ndarray
)
b_ndarray
=
isinstance
(
b
,
n
umpy
.
ndarray
)
a_ndarray
=
isinstance
(
a
,
n
p
.
ndarray
)
b_ndarray
=
isinstance
(
b
,
n
p
.
ndarray
)
if
a_ndarray
and
b_ndarray
:
return
TensorType
.
may_share_memory
(
a
,
b
)
a_cuda
=
_is_cuda
(
a
)
...
...
theano/misc/pkl_utils.py
浏览文件 @
fd7875ad
...
...
@@ -5,7 +5,7 @@ These pickled graphs can be used, for instance, as cases for
unit tests or regression tests.
"""
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
os
import
pickle
import
sys
...
...
@@ -188,10 +188,10 @@ class PersistentNdarrayID(object):
return
name
def
__call__
(
self
,
obj
):
if
type
(
obj
)
is
n
umpy
.
ndarray
:
if
type
(
obj
)
is
n
p
.
ndarray
:
if
id
(
obj
)
not
in
self
.
seen
:
def
write_array
(
f
):
n
umpy
.
lib
.
format
.
write_array
(
f
,
obj
)
n
p
.
lib
.
format
.
write_array
(
f
,
obj
)
name
=
self
.
_resolve_name
(
obj
)
zipadd
(
write_array
,
self
.
zip_file
,
name
)
self
.
seen
[
id
(
obj
)]
=
'ndarray.{0}'
.
format
(
name
)
...
...
@@ -204,7 +204,7 @@ class PersistentCudaNdarrayID(PersistentNdarrayID):
type
(
obj
)
is
cuda_ndarray
.
cuda_ndarray
.
CudaNdarray
):
if
id
(
obj
)
not
in
self
.
seen
:
def
write_array
(
f
):
n
umpy
.
lib
.
format
.
write_array
(
f
,
numpy
.
asarray
(
obj
))
n
p
.
lib
.
format
.
write_array
(
f
,
np
.
asarray
(
obj
))
name
=
self
.
_resolve_name
(
obj
)
zipadd
(
write_array
,
self
.
zip_file
,
name
)
self
.
seen
[
id
(
obj
)]
=
'cuda_ndarray.{0}'
.
format
(
name
)
...
...
@@ -283,7 +283,7 @@ class PersistentNdarrayLoad(object):
if
name
in
self
.
cache
:
return
self
.
cache
[
name
]
ret
=
None
array
=
n
umpy
.
lib
.
format
.
read_array
(
self
.
zip_file
.
open
(
name
))
array
=
n
p
.
lib
.
format
.
read_array
(
self
.
zip_file
.
open
(
name
))
if
array_type
==
'cuda_ndarray'
:
if
config
.
experimental
.
unpickle_gpu_on_cpu
:
# directly return numpy array
...
...
@@ -335,10 +335,10 @@ def dump(obj, file_handler, protocol=DEFAULT_PROTOCOL,
>>> foo_1 = theano.shared(0, name='foo')
>>> foo_2 = theano.shared(1, name='foo')
>>> with open('model.zip', 'wb') as f:
... dump((foo_1, foo_2, n
umpy
.array(2)), f)
>>> n
umpy
.load('model.zip').keys()
... dump((foo_1, foo_2, n
p
.array(2)), f)
>>> n
p
.load('model.zip').keys()
['foo', 'foo_2', 'array_0', 'pkl']
>>> n
umpy
.load('model.zip')['foo']
>>> n
p
.load('model.zip')['foo']
array(0)
>>> with open('model.zip', 'rb') as f:
... foo_1, foo_2, array = load(f)
...
...
theano/misc/pycuda_example.py
浏览文件 @
fd7875ad
...
...
@@ -22,7 +22,7 @@ TheanoElementwiseKernel.
from
__future__
import
absolute_import
,
print_function
,
division
from
itertools
import
chain
import
numpy
import
numpy
as
np
import
theano
from
six.moves
import
xrange
...
...
@@ -257,13 +257,13 @@ class PycudaElemwiseSourceModuleOp(GpuOp):
" inputs don't have the same shape!"
)
if
inputs
[
0
]
.
size
>
512
:
grid
=
(
int
(
n
umpy
.
ceil
(
inputs
[
0
]
.
size
/
512.
)),
1
)
grid
=
(
int
(
n
p
.
ceil
(
inputs
[
0
]
.
size
/
512.
)),
1
)
block
=
(
512
,
1
,
1
)
else
:
grid
=
(
1
,
1
)
block
=
(
inputs
[
0
]
.
shape
[
0
],
inputs
[
0
]
.
shape
[
1
],
1
)
self
.
pycuda_fct
(
inputs
[
0
],
inputs
[
1
],
z
[
0
],
n
umpy
.
intc
(
inputs
[
1
]
.
size
),
block
=
block
,
grid
=
grid
)
n
p
.
intc
(
inputs
[
1
]
.
size
),
block
=
block
,
grid
=
grid
)
class
PycudaElemwiseSourceModuleMakeThunkOp
(
Op
):
...
...
@@ -349,13 +349,13 @@ class PycudaElemwiseSourceModuleMakeThunkOp(Op):
" inputs don't have the same shape!"
)
if
inputs
[
0
][
0
]
.
size
>
512
:
grid
=
(
int
(
n
umpy
.
ceil
(
inputs
[
0
][
0
]
.
size
/
512.
)),
1
)
grid
=
(
int
(
n
p
.
ceil
(
inputs
[
0
][
0
]
.
size
/
512.
)),
1
)
block
=
(
512
,
1
,
1
)
else
:
grid
=
(
1
,
1
)
block
=
(
inputs
[
0
][
0
]
.
shape
[
0
],
inputs
[
0
][
0
]
.
shape
[
1
],
1
)
pycuda_fct
(
inputs
[
0
][
0
],
inputs
[
1
][
0
],
z
[
0
],
n
umpy
.
intc
(
inputs
[
1
][
0
]
.
size
),
block
=
block
,
n
p
.
intc
(
inputs
[
1
][
0
]
.
size
),
block
=
block
,
grid
=
grid
)
thunk
.
inputs
=
inputs
thunk
.
outputs
=
outputs
...
...
theano/misc/safe_asarray.py
浏览文件 @
fd7875ad
...
...
@@ -3,7 +3,7 @@ Helper function to safely convert an array to a new data type.
"""
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
theano
...
...
@@ -30,8 +30,8 @@ def _asarray(a, dtype, order=None):
"""
if
str
(
dtype
)
==
'floatX'
:
dtype
=
theano
.
config
.
floatX
dtype
=
n
umpy
.
dtype
(
dtype
)
# Convert into dtype object.
rval
=
n
umpy
.
asarray
(
a
,
dtype
=
dtype
,
order
=
order
)
dtype
=
n
p
.
dtype
(
dtype
)
# Convert into dtype object.
rval
=
n
p
.
asarray
(
a
,
dtype
=
dtype
,
order
=
order
)
# Note that dtype comparison must be done by comparing their `num`
# attribute. One cannot assume that two identical data types are pointers
# towards the same object (e.g. under Windows this appears not to be the
...
...
theano/misc/tests/test_cudamat_utils.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
theano
from
theano.misc.cudamat_utils
import
cudamat_available
...
...
@@ -20,7 +20,7 @@ def test(shape=(3, 4)):
U
=
gpu
(
theano
.
tensor
.
fmatrix
(
'U'
))
ii
=
theano
.
function
([
U
],
gpu
(
U
+
1
))
A_cpu
=
n
umpy
.
asarray
(
numpy
.
random
.
rand
(
*
shape
),
dtype
=
"float32"
)
A_cpu
=
n
p
.
asarray
(
np
.
random
.
rand
(
*
shape
),
dtype
=
"float32"
)
A_cnd
=
theano
.
sandbox
.
cuda
.
CudaNdarray
(
A_cpu
)
A_cmat
=
cudandarray_to_cudamat
(
A_cnd
)
...
...
@@ -28,9 +28,9 @@ def test(shape=(3, 4)):
B_cnd
=
ii
(
A_cnd
)
u
=
A_cnd
.
copy
()
u
+=
theano
.
sandbox
.
cuda
.
CudaNdarray
(
n
umpy
.
asarray
([[
1
]],
dtype
=
'float32'
))
u
=
n
umpy
.
asarray
(
u
)
v
=
n
umpy
.
asarray
(
B_cnd
)
u
+=
theano
.
sandbox
.
cuda
.
CudaNdarray
(
n
p
.
asarray
([[
1
]],
dtype
=
'float32'
))
u
=
n
p
.
asarray
(
u
)
v
=
n
p
.
asarray
(
B_cnd
)
w
=
A_cmat
.
add
(
1
)
.
asarray
()
assert
abs
(
u
-
v
)
.
max
()
==
0
...
...
theano/misc/tests/test_gnumpy_utils.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
theano
from
theano.misc.gnumpy_utils
import
gnumpy_available
...
...
@@ -31,11 +31,10 @@ def test(shape=(3, 4, 5)):
B_cnd
=
ii
(
A_cnd
)
B
=
cudandarray_to_garray
(
B_cnd
)
assert
A_cnd
.
shape
==
A
.
shape
from
numpy
import
array
u
=
(
A
+
1
)
.
asarray
()
v
=
B
.
asarray
()
w
=
array
(
B_cnd
)
w
=
np
.
array
(
B_cnd
)
assert
(
u
==
v
)
.
all
()
assert
(
u
==
w
)
.
all
()
...
...
@@ -49,7 +48,7 @@ def test2(shape=(3, 4, 5)):
U
=
gpu
(
theano
.
tensor
.
ftensor3
(
'U'
))
theano
.
function
([
U
],
gpu
(
U
+
1
))
A
=
n
umpy
.
random
.
rand
(
*
shape
)
.
astype
(
'float32'
)
A
=
n
p
.
random
.
rand
(
*
shape
)
.
astype
(
'float32'
)
A_cnd
=
theano
.
sandbox
.
cuda
.
CudaNdarray
(
A
)
A_gar
=
cudandarray_to_garray
(
A_cnd
)
assert
A_cnd
.
shape
==
A_gar
.
shape
...
...
@@ -62,7 +61,7 @@ def test2(shape=(3, 4, 5)):
# dtype always float32
assert
A_cnd
.
_strides
==
B
.
_strides
assert
A_cnd
.
gpudata
==
B
.
gpudata
v
=
n
umpy
.
asarray
(
B
)
v
=
n
p
.
asarray
(
B
)
assert
(
v
==
A
)
.
all
()
...
...
theano/misc/tests/test_may_share_memory.py
浏览文件 @
fd7875ad
...
...
@@ -3,7 +3,7 @@ test the tensor and sparse type. The CudaNdarray type is tested in
sandbox/cuda/tests/test_tensor_op.py.test_may_share_memory_cuda
"""
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
theano
try
:
...
...
@@ -16,8 +16,8 @@ from theano.misc.may_share_memory import may_share_memory
def
test_may_share_memory
():
a
=
n
umpy
.
random
.
rand
(
5
,
4
)
b
=
n
umpy
.
random
.
rand
(
5
,
4
)
a
=
n
p
.
random
.
rand
(
5
,
4
)
b
=
n
p
.
random
.
rand
(
5
,
4
)
va
=
a
.
view
()
vb
=
b
.
view
()
ra
=
a
.
reshape
((
4
,
5
))
...
...
theano/misc/tests/test_pkl_utils.py
浏览文件 @
fd7875ad
...
...
@@ -4,8 +4,7 @@ import shutil
import
unittest
from
tempfile
import
mkdtemp
import
numpy
from
numpy.testing
import
assert_allclose
import
numpy
as
np
from
nose.plugins.skip
import
SkipTest
import
theano
...
...
@@ -44,7 +43,7 @@ class T_dump_load(unittest.TestCase):
x
=
load
(
f
)
assert
x
.
name
==
'x'
assert_allclose
(
x
.
get_value
(),
[[
1
]])
np
.
testing
.
assert_allclose
(
x
.
get_value
(),
[[
1
]])
def
test_dump_load_mrg
(
self
):
rng
=
MRG_RandomStreams
(
use_cuda
=
cuda_ndarray
.
cuda_enabled
)
...
...
@@ -62,14 +61,14 @@ class T_dump_load(unittest.TestCase):
foo_2
=
theano
.
shared
(
1
,
name
=
'foo'
)
foo_3
=
theano
.
shared
(
2
,
name
=
'foo'
)
with
open
(
'model.zip'
,
'wb'
)
as
f
:
dump
((
foo_1
,
foo_2
,
foo_3
,
n
umpy
.
array
(
3
)),
f
)
keys
=
list
(
n
umpy
.
load
(
'model.zip'
)
.
keys
())
dump
((
foo_1
,
foo_2
,
foo_3
,
n
p
.
array
(
3
)),
f
)
keys
=
list
(
n
p
.
load
(
'model.zip'
)
.
keys
())
assert
keys
==
[
'foo'
,
'foo_2'
,
'foo_3'
,
'array_0'
,
'pkl'
]
foo_3
=
n
umpy
.
load
(
'model.zip'
)[
'foo_3'
]
assert
foo_3
==
n
umpy
.
array
(
2
)
foo_3
=
n
p
.
load
(
'model.zip'
)[
'foo_3'
]
assert
foo_3
==
n
p
.
array
(
2
)
with
open
(
'model.zip'
,
'rb'
)
as
f
:
foo_1
,
foo_2
,
foo_3
,
array
=
load
(
f
)
assert
array
==
n
umpy
.
array
(
3
)
assert
array
==
n
p
.
array
(
3
)
class
TestStripPickler
(
unittest
.
TestCase
):
...
...
theano/misc/tests/test_pycuda_example.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
theano
import
theano.misc.pycuda_init
...
...
@@ -58,11 +58,11 @@ def test_pycuda_elemwise_source_module():
PycudaElemwiseSourceModuleMakeThunkOp
)
for
node
in
f4
.
maker
.
fgraph
.
toposort
()])
val1
=
n
umpy
.
asarray
(
numpy
.
random
.
rand
(
*
shape
),
dtype
=
'float32'
)
val2
=
n
umpy
.
asarray
(
numpy
.
random
.
rand
(
*
shape
),
dtype
=
'float32'
)
assert
n
umpy
.
allclose
(
f
(
val1
,
val2
),
f2
(
val1
,
val2
))
assert
n
umpy
.
allclose
(
f
(
val1
,
val2
),
f3
(
val1
,
val2
))
assert
n
umpy
.
allclose
(
f
(
val1
,
val2
),
f4
(
val1
,
val2
))
val1
=
n
p
.
asarray
(
np
.
random
.
rand
(
*
shape
),
dtype
=
'float32'
)
val2
=
n
p
.
asarray
(
np
.
random
.
rand
(
*
shape
),
dtype
=
'float32'
)
assert
n
p
.
allclose
(
f
(
val1
,
val2
),
f2
(
val1
,
val2
))
assert
n
p
.
allclose
(
f
(
val1
,
val2
),
f3
(
val1
,
val2
))
assert
n
p
.
allclose
(
f
(
val1
,
val2
),
f4
(
val1
,
val2
))
# print f(val1,val2)
# print f2(val1,val2)
...
...
@@ -82,10 +82,10 @@ def test_pycuda_elemwise_kernel():
assert any([isinstance(node.op, PycudaElemwiseKernelOp)
for node in f2.maker.fgraph.toposort()])
val1 = n
umpy.asarray(numpy
.random.rand(5, 5), dtype='float32')
val2 = n
umpy.asarray(numpy
.random.rand(5, 5), dtype='float32')
#val1 = n
umpy
.ones((5,5))
#val2 = n
umpy
.arange(25).reshape(5,5)
val1 = n
p.asarray(np
.random.rand(5, 5), dtype='float32')
val2 = n
p.asarray(np
.random.rand(5, 5), dtype='float32')
#val1 = n
p
.ones((5,5))
#val2 = n
p
.arange(25).reshape(5,5)
assert (f(val1, val2) == f2(val1, val2)).all()
print(f(val1, val2))
print(f2(val1, val2))
...
...
@@ -99,8 +99,8 @@ def test_pycuda_elemwise_kernel():
assert any([isinstance(node.op, PycudaElemwiseKernelOp)
for node in f4.maker.fgraph.toposort()])
val1 = n
umpy
.random.rand(2, 2, 2)
val1 = n
p
.random.rand(2, 2, 2)
print(val1)
print(f4(val1, val1, val1))
assert n
umpy
.allclose(f4(val1, val1, val1), val1 * val1 + val1)
assert n
p
.allclose(f4(val1, val1, val1), val1 * val1 + val1)
"""
theano/misc/tests/test_pycuda_theano_simple.py
浏览文件 @
fd7875ad
...
...
@@ -8,7 +8,7 @@ from __future__ import absolute_import, print_function, division
import
sys
import
numpy
import
numpy
as
np
import
theano
import
theano.sandbox.cuda
as
cuda_ndarray
...
...
@@ -42,9 +42,9 @@ __global__ void multiply_them(float *dest, float *a, float *b)
multiply_them
=
mod
.
get_function
(
"multiply_them"
)
# Test with pycuda in/out of numpy.ndarray
a
=
n
umpy
.
random
.
randn
(
100
)
.
astype
(
numpy
.
float32
)
b
=
n
umpy
.
random
.
randn
(
100
)
.
astype
(
numpy
.
float32
)
dest
=
n
umpy
.
zeros_like
(
a
)
a
=
n
p
.
random
.
randn
(
100
)
.
astype
(
np
.
float32
)
b
=
n
p
.
random
.
randn
(
100
)
.
astype
(
np
.
float32
)
dest
=
n
p
.
zeros_like
(
a
)
multiply_them
(
drv
.
Out
(
dest
),
drv
.
In
(
a
),
drv
.
In
(
b
),
block
=
(
400
,
1
,
1
),
grid
=
(
1
,
1
))
...
...
@@ -64,8 +64,8 @@ __global__ void multiply_them(float *dest, float *a, float *b)
multiply_them
=
mod
.
get_function
(
"multiply_them"
)
a
=
n
umpy
.
random
.
randn
(
100
)
.
astype
(
numpy
.
float32
)
b
=
n
umpy
.
random
.
randn
(
100
)
.
astype
(
numpy
.
float32
)
a
=
n
p
.
random
.
randn
(
100
)
.
astype
(
np
.
float32
)
b
=
n
p
.
random
.
randn
(
100
)
.
astype
(
np
.
float32
)
# Test with Theano object
ga
=
cuda_ndarray
.
CudaNdarray
(
a
)
...
...
@@ -73,7 +73,7 @@ __global__ void multiply_them(float *dest, float *a, float *b)
dest
=
cuda_ndarray
.
CudaNdarray
.
zeros
(
a
.
shape
)
multiply_them
(
dest
,
ga
,
gb
,
block
=
(
400
,
1
,
1
),
grid
=
(
1
,
1
))
assert
(
n
umpy
.
asarray
(
dest
)
==
a
*
b
)
.
all
()
assert
(
n
p
.
asarray
(
dest
)
==
a
*
b
)
.
all
()
def
test_pycuda_memory_to_theano
():
...
...
@@ -87,7 +87,7 @@ def test_pycuda_memory_to_theano():
print
(
"gpuarray ref count before creating a CudaNdarray"
,
end
=
' '
)
print
(
sys
.
getrefcount
(
y
))
assert
sys
.
getrefcount
(
y
)
==
initial_refcount
rand
=
n
umpy
.
random
.
randn
(
*
y
.
shape
)
.
astype
(
numpy
.
float32
)
rand
=
n
p
.
random
.
randn
(
*
y
.
shape
)
.
astype
(
np
.
float32
)
cuda_rand
=
cuda_ndarray
.
CudaNdarray
(
rand
)
strides
=
[
1
]
...
...
@@ -102,7 +102,7 @@ def test_pycuda_memory_to_theano():
z
=
cuda_ndarray
.
from_gpu_pointer
(
y_ptr
,
y
.
shape
,
strides
,
y
)
print
(
"gpuarray ref count after creating a CudaNdarray"
,
sys
.
getrefcount
(
y
))
assert
sys
.
getrefcount
(
y
)
==
initial_refcount
+
1
assert
(
n
umpy
.
asarray
(
z
)
==
0
)
.
all
()
assert
(
n
p
.
asarray
(
z
)
==
0
)
.
all
()
assert
z
.
base
is
y
# Test that we can take a view from this cuda view on pycuda memory
...
...
@@ -112,17 +112,17 @@ def test_pycuda_memory_to_theano():
del
zz
assert
sys
.
getrefcount
(
y
)
==
initial_refcount
+
1
cuda_ones
=
cuda_ndarray
.
CudaNdarray
(
n
umpy
.
asarray
([[[
1
]]],
dtype
=
'float32'
))
cuda_ones
=
cuda_ndarray
.
CudaNdarray
(
n
p
.
asarray
([[[
1
]]],
dtype
=
'float32'
))
z
+=
cuda_ones
assert
(
n
umpy
.
asarray
(
z
)
==
numpy
.
ones
(
y
.
shape
))
.
all
()
assert
(
n
umpy
.
asarray
(
z
)
==
1
)
.
all
()
assert
(
n
p
.
asarray
(
z
)
==
np
.
ones
(
y
.
shape
))
.
all
()
assert
(
n
p
.
asarray
(
z
)
==
1
)
.
all
()
assert
cuda_rand
.
shape
==
z
.
shape
assert
cuda_rand
.
_strides
==
z
.
_strides
,
(
cuda_rand
.
_strides
,
z
.
_strides
)
assert
(
n
umpy
.
asarray
(
cuda_rand
)
==
rand
)
.
all
()
assert
(
n
p
.
asarray
(
cuda_rand
)
==
rand
)
.
all
()
z
+=
cuda_rand
assert
(
n
umpy
.
asarray
(
z
)
==
(
rand
+
1
))
.
all
()
assert
(
n
p
.
asarray
(
z
)
==
(
rand
+
1
))
.
all
()
# Check that the ref count to the gpuarray is right.
del
z
...
...
theano/misc/tests/test_pycuda_utils.py
浏览文件 @
fd7875ad
from
__future__
import
absolute_import
,
print_function
,
division
import
numpy
import
numpy
as
np
import
theano.sandbox.cuda
as
cuda
import
theano.misc.pycuda_init
...
...
@@ -22,30 +22,30 @@ def test_to_gpuarray():
px
=
to_gpuarray
(
cx
)
assert
isinstance
(
px
,
pycuda
.
gpuarray
.
GPUArray
)
cx
[
0
,
0
]
=
n
umpy
.
asarray
(
1
,
dtype
=
"float32"
)
cx
[
0
,
0
]
=
n
p
.
asarray
(
1
,
dtype
=
"float32"
)
# Check that they share the same memory space
assert
px
.
gpudata
==
cx
.
gpudata
assert
n
umpy
.
asarray
(
cx
[
0
,
0
])
==
1
assert
n
p
.
asarray
(
cx
[
0
,
0
])
==
1
assert
n
umpy
.
allclose
(
numpy
.
asarray
(
cx
),
px
.
get
())
assert
n
p
.
allclose
(
np
.
asarray
(
cx
),
px
.
get
())
assert
px
.
dtype
==
cx
.
dtype
assert
px
.
shape
==
cx
.
shape
assert
all
(
n
umpy
.
asarray
(
cx
.
_strides
)
*
4
==
px
.
strides
)
assert
all
(
n
p
.
asarray
(
cx
.
_strides
)
*
4
==
px
.
strides
)
# Test when the CudaNdarray is strided
cx
=
cx
[::
2
,
::]
px
=
to_gpuarray
(
cx
,
copyif
=
True
)
assert
isinstance
(
px
,
pycuda
.
gpuarray
.
GPUArray
)
cx
[
0
,
0
]
=
n
umpy
.
asarray
(
2
,
dtype
=
"float32"
)
cx
[
0
,
0
]
=
n
p
.
asarray
(
2
,
dtype
=
"float32"
)
# Check that they do not share the same memory space
assert
px
.
gpudata
!=
cx
.
gpudata
assert
n
umpy
.
asarray
(
cx
[
0
,
0
])
==
2
assert
not
n
umpy
.
allclose
(
numpy
.
asarray
(
cx
),
px
.
get
())
assert
n
p
.
asarray
(
cx
[
0
,
0
])
==
2
assert
not
n
p
.
allclose
(
np
.
asarray
(
cx
),
px
.
get
())
assert
px
.
dtype
==
cx
.
dtype
assert
px
.
shape
==
cx
.
shape
assert
not
all
(
n
umpy
.
asarray
(
cx
.
_strides
)
*
4
==
px
.
strides
)
assert
not
all
(
n
p
.
asarray
(
cx
.
_strides
)
*
4
==
px
.
strides
)
# Test that we return an error
try
:
...
...
@@ -59,11 +59,11 @@ def test_to_cudandarray():
px
=
pycuda
.
gpuarray
.
zeros
((
3
,
4
,
5
),
'float32'
)
cx
=
to_cudandarray
(
px
)
assert
isinstance
(
cx
,
cuda
.
CudaNdarray
)
assert
n
umpy
.
allclose
(
px
.
get
(),
numpy
.
asarray
(
cx
))
assert
n
p
.
allclose
(
px
.
get
(),
np
.
asarray
(
cx
))
assert
px
.
dtype
==
cx
.
dtype
assert
px
.
shape
==
cx
.
shape
assert
all
(
n
umpy
.
asarray
(
cx
.
_strides
)
*
4
==
px
.
strides
)
assert
all
(
n
p
.
asarray
(
cx
.
_strides
)
*
4
==
px
.
strides
)
try
:
px
=
pycuda
.
gpuarray
.
zeros
((
3
,
4
,
5
),
'float64'
)
...
...
@@ -73,7 +73,7 @@ def test_to_cudandarray():
pass
try
:
to_cudandarray
(
n
umpy
.
zeros
(
4
))
to_cudandarray
(
n
p
.
zeros
(
4
))
assert
False
except
ValueError
:
pass
theano/sandbox/cuda/__init__.py
浏览文件 @
fd7875ad
...
...
@@ -12,7 +12,7 @@ import warnings
import
theano
from
theano.compat
import
get_unbound_function
from
theano.compile
import
optdb
from
theano.gof
import
EquilibriumDB
,
SequenceDB
from
theano.gof
import
EquilibriumDB
,
SequenceDB
,
TopoOptimizer
from
theano.gof.cmodule
import
get_lib_extension
from
theano.gof.compilelock
import
get_lock
,
release_lock
from
theano
import
config
...
...
@@ -40,6 +40,17 @@ def register_opt(*tags, **kwargs):
return
f
def
register_inplace
(
*
tags
,
**
kwargs
):
def
f
(
local_opt
):
name
=
(
kwargs
and
kwargs
.
pop
(
'name'
))
or
local_opt
.
__name__
optdb
.
register
(
name
,
TopoOptimizer
(
local_opt
,
failure_callback
=
TopoOptimizer
.
warn_inplace
),
60
,
'fast_run'
,
'inplace'
,
'gpu'
,
*
tags
)
return
local_opt
return
f
_logger_name
=
'theano.sandbox.cuda'
_logger
=
logging
.
getLogger
(
_logger_name
)
...
...
theano/sandbox/cuda/dnn.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/sandbox/cuda/opt.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/sandbox/cuda/tests/test_dnn.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/sandbox/cuda/var.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/tensor/nnet/__init__.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/tensor/nnet/bn.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/tensor/nnet/tests/test_abstract_conv.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/tensor/nnet/tests/test_bn.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/tensor/tests/test_var.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
theano/tensor/var.py
浏览文件 @
fd7875ad
差异被折叠。
点击展开。
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论