Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
1a18f3db
提交
1a18f3db
authored
12月 17, 2012
作者:
nouiz
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #1136 from delallea/minor
Minor fixes
上级
9555a94f
92ec12fd
隐藏空白字符变更
内嵌
并排
正在显示
4 个修改的文件
包含
41 行增加
和
15 行删除
+41
-15
sched.py
theano/gof/sched.py
+20
-0
__init__.py
theano/sandbox/cuda/__init__.py
+4
-4
basic.py
theano/tensor/basic.py
+13
-7
nnet.py
theano/tensor/nnet/nnet.py
+4
-4
没有找到文件。
theano/gof/sched.py
浏览文件 @
1a18f3db
...
@@ -3,6 +3,25 @@ from theano.gof.python25 import any, defaultdict
...
@@ -3,6 +3,25 @@ from theano.gof.python25 import any, defaultdict
## {{{ http://code.activestate.com/recipes/578231/ (r1)
## {{{ http://code.activestate.com/recipes/578231/ (r1)
# Copyright (c) Oren Tirosh 2012
#
# Permission is hereby granted, free of charge, to any person obtaining a copy of
# this software and associated documentation files (the "Software"), to deal in
# the Software without restriction, including without limitation the rights to
# use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies
# of the Software, and to permit persons to whom the Software is furnished to do
# so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
def
memodict
(
f
):
def
memodict
(
f
):
""" Memoization decorator for a function taking a single argument """
""" Memoization decorator for a function taking a single argument """
class
memodict
(
defaultdict
):
class
memodict
(
defaultdict
):
...
@@ -12,6 +31,7 @@ def memodict(f):
...
@@ -12,6 +31,7 @@ def memodict(f):
return
memodict
()
.
__getitem__
return
memodict
()
.
__getitem__
## end of http://code.activestate.com/recipes/578231/ }}}
## end of http://code.activestate.com/recipes/578231/ }}}
def
make_depends
():
def
make_depends
():
@memodict
@memodict
def
depends
((
a
,
b
)):
def
depends
((
a
,
b
)):
...
...
theano/sandbox/cuda/__init__.py
浏览文件 @
1a18f3db
...
@@ -375,11 +375,11 @@ def use(device,
...
@@ -375,11 +375,11 @@ def use(device,
import
theano.sandbox.cuda.tests.test_driver
import
theano.sandbox.cuda.tests.test_driver
theano
.
sandbox
.
cuda
.
tests
.
test_driver
.
test_nvidia_driver1
()
theano
.
sandbox
.
cuda
.
tests
.
test_driver
.
test_nvidia_driver1
()
if
device_properties
(
use
.
device_number
)[
"warpSize"
]
!=
32
:
if
device_properties
(
use
.
device_number
)[
"warpSize"
]
!=
32
:
raise
ValueError
(
"Your GPU ha
ve a warpSize of
32. Currently"
raise
ValueError
(
"Your GPU ha
s a warpSize !=
32. Currently"
" we have code that depend on this. Email"
" we have code that depend
s
on this. Email"
" Theano mailing list to tell us about"
"
the
Theano mailing list to tell us about"
" this new GPU as we don't know any with"
" this new GPU as we don't know any with"
" this propert
ies
"
)
" this propert
y
"
)
if
move_shared_float32_to_gpu
:
if
move_shared_float32_to_gpu
:
handle_shared_float32
(
True
)
handle_shared_float32
(
True
)
...
...
theano/tensor/basic.py
浏览文件 @
1a18f3db
...
@@ -1443,7 +1443,7 @@ class _tensor_py_operators:
...
@@ -1443,7 +1443,7 @@ class _tensor_py_operators:
def
__sub__
(
self
,
other
):
def
__sub__
(
self
,
other
):
# See explanation in __add__ for the error catched
# See explanation in __add__ for the error catched
# a
dn
the return value in that case
# a
nd
the return value in that case
try
:
try
:
return
sub
(
self
,
other
)
return
sub
(
self
,
other
)
except
(
NotImplementedError
,
TypeError
):
except
(
NotImplementedError
,
TypeError
):
...
@@ -1451,7 +1451,7 @@ class _tensor_py_operators:
...
@@ -1451,7 +1451,7 @@ class _tensor_py_operators:
def
__mul__
(
self
,
other
):
def
__mul__
(
self
,
other
):
# See explanation in __add__ for the error catched
# See explanation in __add__ for the error catched
# a
dn
the return value in that case
# a
nd
the return value in that case
try
:
try
:
return
mul
(
self
,
other
)
return
mul
(
self
,
other
)
except
(
NotImplementedError
,
TypeError
):
except
(
NotImplementedError
,
TypeError
):
...
@@ -1459,7 +1459,7 @@ class _tensor_py_operators:
...
@@ -1459,7 +1459,7 @@ class _tensor_py_operators:
def
__div__
(
self
,
other
):
def
__div__
(
self
,
other
):
# See explanation in __add__ for the error catched
# See explanation in __add__ for the error catched
# a
dn
the return value in that case
# a
nd
the return value in that case
try
:
try
:
return
div_proxy
(
self
,
other
)
return
div_proxy
(
self
,
other
)
except
IntegerDivisionError
:
except
IntegerDivisionError
:
...
@@ -6833,7 +6833,7 @@ def take(a, indices, axis=None, mode='raise'):
...
@@ -6833,7 +6833,7 @@ def take(a, indices, axis=None, mode='raise'):
# Reuse advanced_subtensor1 if indices is a vector
# Reuse advanced_subtensor1 if indices is a vector
if
indices
.
ndim
==
1
:
if
indices
.
ndim
==
1
:
if
mode
==
'clip'
:
if
mode
==
'clip'
:
indices
=
clip
(
indices
,
0
,
a
.
shape
[
axis
]
-
1
)
indices
=
clip
(
indices
,
0
,
a
.
shape
[
axis
]
-
1
)
elif
mode
==
'wrap'
:
elif
mode
==
'wrap'
:
indices
=
indices
%
a
.
shape
[
axis
]
indices
=
indices
%
a
.
shape
[
axis
]
if
axis
is
None
:
if
axis
is
None
:
...
@@ -6853,10 +6853,12 @@ def take(a, indices, axis=None, mode='raise'):
...
@@ -6853,10 +6853,12 @@ def take(a, indices, axis=None, mode='raise'):
shape
=
indices
.
shape
shape
=
indices
.
shape
ndim
=
indices
.
ndim
ndim
=
indices
.
ndim
else
:
else
:
shape
=
concatenate
([
a
.
shape
[:
axis
],
indices
.
shape
,
a
.
shape
[
axis
+
1
:]])
shape
=
concatenate
(
[
a
.
shape
[:
axis
],
indices
.
shape
,
a
.
shape
[
axis
+
1
:]])
ndim
=
a
.
ndim
+
indices
.
ndim
-
1
ndim
=
a
.
ndim
+
indices
.
ndim
-
1
return
take
(
a
,
indices
.
flatten
(),
axis
,
mode
)
.
reshape
(
shape
,
ndim
)
return
take
(
a
,
indices
.
flatten
(),
axis
,
mode
)
.
reshape
(
shape
,
ndim
)
#########################
#########################
# Linalg : Dot
# Linalg : Dot
#########################
#########################
...
@@ -7283,6 +7285,7 @@ def all(x, axis=None, keepdims=False):
...
@@ -7283,6 +7285,7 @@ def all(x, axis=None, keepdims=False):
out
=
makeKeepDims
(
x
,
out
,
axis
)
out
=
makeKeepDims
(
x
,
out
,
axis
)
return
out
return
out
class
Diagonal
(
Op
):
class
Diagonal
(
Op
):
"""Return specified diagonals.
"""Return specified diagonals.
...
@@ -7310,7 +7313,7 @@ class Diagonal(Op):
...
@@ -7310,7 +7313,7 @@ class Diagonal(Op):
x
=
as_tensor_variable
(
x
)
x
=
as_tensor_variable
(
x
)
assert
x
.
ndim
>=
2
assert
x
.
ndim
>=
2
return
Apply
(
self
,
[
x
],
[
tensor
(
dtype
=
x
.
dtype
,
return
Apply
(
self
,
[
x
],
[
tensor
(
dtype
=
x
.
dtype
,
broadcastable
=
[
False
]
*
(
x
.
ndim
-
1
))])
broadcastable
=
[
False
]
*
(
x
.
ndim
-
1
))])
def
perform
(
self
,
node
,
(
x
,),
(
z
,)):
def
perform
(
self
,
node
,
(
x
,),
(
z
,)):
z
[
0
]
=
x
.
diagonal
(
self
.
offset
,
self
.
axis1
,
self
.
axis2
)
z
[
0
]
=
x
.
diagonal
(
self
.
offset
,
self
.
axis1
,
self
.
axis2
)
...
@@ -7322,7 +7325,7 @@ class Diagonal(Op):
...
@@ -7322,7 +7325,7 @@ class Diagonal(Op):
in_shape
,
=
shapes
in_shape
,
=
shapes
dim1
=
in_shape
[
self
.
axis1
]
dim1
=
in_shape
[
self
.
axis1
]
dim2
=
in_shape
[
self
.
axis2
]
dim2
=
in_shape
[
self
.
axis2
]
out_shape
=
[
d
for
i
,
d
in
enumerate
(
in_shape
)
out_shape
=
[
d
for
i
,
d
in
enumerate
(
in_shape
)
if
i
not
in
(
self
.
axis1
,
self
.
axis2
)]
if
i
not
in
(
self
.
axis1
,
self
.
axis2
)]
# The following logic is inspired by C code of PyArray_Diagonal().
# The following logic is inspired by C code of PyArray_Diagonal().
offset
=
self
.
offset
offset
=
self
.
offset
...
@@ -7338,12 +7341,14 @@ class Diagonal(Op):
...
@@ -7338,12 +7341,14 @@ class Diagonal(Op):
def
__str__
(
self
):
def
__str__
(
self
):
return
self
.
__class__
.
__name__
return
self
.
__class__
.
__name__
def
diagonal
(
a
,
offset
=
0
,
axis1
=
0
,
axis2
=
1
):
def
diagonal
(
a
,
offset
=
0
,
axis1
=
0
,
axis2
=
1
):
if
(
offset
,
axis1
,
axis2
)
==
(
0
,
0
,
1
):
if
(
offset
,
axis1
,
axis2
)
==
(
0
,
0
,
1
):
from
theano.sandbox.linalg
import
extract_diag
from
theano.sandbox.linalg
import
extract_diag
return
extract_diag
(
a
)
return
extract_diag
(
a
)
return
Diagonal
(
offset
,
axis1
,
axis2
)(
a
)
return
Diagonal
(
offset
,
axis1
,
axis2
)(
a
)
class
Diag
(
Op
):
class
Diag
(
Op
):
def
__eq__
(
self
,
other
):
def
__eq__
(
self
,
other
):
...
@@ -7371,6 +7376,7 @@ class Diag(Op):
...
@@ -7371,6 +7376,7 @@ class Diag(Op):
def
__str__
(
self
):
def
__str__
(
self
):
return
self
.
__class__
.
__name__
return
self
.
__class__
.
__name__
def
diag
(
v
,
k
=
0
):
def
diag
(
v
,
k
=
0
):
if
v
.
ndim
==
1
:
if
v
.
ndim
==
1
:
assert
k
==
0
,
"diagonals other than main are not implemented"
assert
k
==
0
,
"diagonals other than main are not implemented"
...
...
theano/tensor/nnet/nnet.py
浏览文件 @
1a18f3db
...
@@ -109,7 +109,7 @@ class SoftmaxWithBias(gof.Op):
...
@@ -109,7 +109,7 @@ class SoftmaxWithBias(gof.Op):
if (PyArray_NDIM(
%(x)
s) != 2)
if (PyArray_NDIM(
%(x)
s) != 2)
{
{
PyErr_SetString(PyExc_ValueError, "
a not
2d tensor");
PyErr_SetString(PyExc_ValueError, "
not a
2d tensor");
%(fail)
s;
%(fail)
s;
}
}
if (PyArray_NDIM(
%(b)
s) != 1)
if (PyArray_NDIM(
%(b)
s) != 1)
...
@@ -120,7 +120,7 @@ class SoftmaxWithBias(gof.Op):
...
@@ -120,7 +120,7 @@ class SoftmaxWithBias(gof.Op):
if ((PyArray_DESCR(
%(x)
s)->type_num != NPY_DOUBLE) &&
if ((PyArray_DESCR(
%(x)
s)->type_num != NPY_DOUBLE) &&
(PyArray_DESCR(
%(x)
s)->type_num != NPY_FLOAT))
(PyArray_DESCR(
%(x)
s)->type_num != NPY_FLOAT))
{
{
PyErr_SetString(PyExc_TypeError, "
a not
float");
PyErr_SetString(PyExc_TypeError, "
not a
float");
%(fail)
s;
%(fail)
s;
}
}
if ((PyArray_DESCR(
%(b)
s)->type_num != NPY_DOUBLE) &&
if ((PyArray_DESCR(
%(b)
s)->type_num != NPY_DOUBLE) &&
...
@@ -401,13 +401,13 @@ class Softmax(gof.Op):
...
@@ -401,13 +401,13 @@ class Softmax(gof.Op):
if (
%(x)
s->nd != 2)
if (
%(x)
s->nd != 2)
{
{
PyErr_SetString(PyExc_ValueError, "
a not
2d tensor");
PyErr_SetString(PyExc_ValueError, "
not a
2d tensor");
%(fail)
s;
%(fail)
s;
}
}
if ((
%(x)
s->descr->type_num != PyArray_DOUBLE) &&
if ((
%(x)
s->descr->type_num != PyArray_DOUBLE) &&
(
%(x)
s->descr->type_num != PyArray_FLOAT))
(
%(x)
s->descr->type_num != PyArray_FLOAT))
{
{
PyErr_SetString(PyExc_TypeError, "
a not
float");
PyErr_SetString(PyExc_TypeError, "
not a
float");
%(fail)
s;
%(fail)
s;
}
}
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论