Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
842da1f2
提交
842da1f2
authored
2月 26, 2016
作者:
Taesup (TS) Kim
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fix PEP8 in sandbox/*.py
上级
c51b2833
显示空白字符变更
内嵌
并排
正在显示
9 个修改的文件
包含
144 行增加
和
77 行删除
+144
-77
conv.py
theano/sandbox/conv.py
+2
-1
debug.py
theano/sandbox/debug.py
+28
-13
fourier.py
theano/sandbox/fourier.py
+11
-5
minimal.py
theano/sandbox/minimal.py
+10
-8
multinomial.py
theano/sandbox/multinomial.py
+10
-5
neighbourhoods.py
theano/sandbox/neighbourhoods.py
+10
-2
rng_mrg.py
theano/sandbox/rng_mrg.py
+70
-42
softsign.py
theano/sandbox/softsign.py
+1
-0
solve.py
theano/sandbox/solve.py
+2
-1
没有找到文件。
theano/sandbox/conv.py
浏览文件 @
842da1f2
from
__future__
import
print_function
from
__future__
import
print_function
import
sys
import
sys
print
(
"DEPRECATION: theano.sandbox.conv no longer provides conv. They have been moved to theano.tensor.nnet.conv"
,
file
=
sys
.
stderr
)
print
(
"DEPRECATION: theano.sandbox.conv no longer provides conv. "
"They have been moved to theano.tensor.nnet.conv"
,
file
=
sys
.
stderr
)
from
theano.tensor.nnet.conv
import
*
from
theano.tensor.nnet.conv
import
*
theano/sandbox/debug.py
浏览文件 @
842da1f2
...
@@ -71,10 +71,15 @@ class DebugLinker(gof.WrapLinker):
...
@@ -71,10 +71,15 @@ class DebugLinker(gof.WrapLinker):
r
.
type
.
filter
(
r
.
value
,
strict
=
True
)
r
.
type
.
filter
(
r
.
value
,
strict
=
True
)
except
TypeError
as
e
:
except
TypeError
as
e
:
exc_type
,
exc_value
,
exc_trace
=
sys
.
exc_info
()
exc_type
,
exc_value
,
exc_trace
=
sys
.
exc_info
()
exc
=
DebugException
(
e
,
"The output
%
s was filled with data with the wrong type using linker "
\
exc
=
DebugException
(
(
"
%
s. This happened at step
%
i of the program."
%
(
r
,
linker
,
i
))
+
\
e
,
"For more info, inspect this exception's 'original_exception', 'debugger', "
\
"The output
%
s was filled with data with the wrong "
"'output_at_fault', 'step', 'node', 'thunk' and 'linker' fields."
)
"type using linker "
+
(
"
%
s. This happened at step
%
i of the program."
%
(
r
,
linker
,
i
))
+
"For more info, inspect this exception's "
"'original_exception', 'debugger', 'output_at_fault', "
"'step', 'node', 'thunk' and 'linker' fields."
)
exc
.
debugger
=
self
exc
.
debugger
=
self
exc
.
original_exception
=
e
exc
.
original_exception
=
e
exc
.
output_at_fault
=
r
exc
.
output_at_fault
=
r
...
@@ -88,11 +93,18 @@ class DebugLinker(gof.WrapLinker):
...
@@ -88,11 +93,18 @@ class DebugLinker(gof.WrapLinker):
thunk0
=
thunks
[
0
]
thunk0
=
thunks
[
0
]
linker0
=
self
.
linkers
[
0
]
linker0
=
self
.
linkers
[
0
]
for
thunk
,
linker
in
zip
(
thunks
[
1
:],
self
.
linkers
[
1
:]):
for
thunk
,
linker
in
zip
(
thunks
[
1
:],
self
.
linkers
[
1
:]):
for
o
,
output0
,
output
in
zip
(
node
.
outputs
,
thunk0
.
outputs
,
thunk
.
outputs
):
for
o
,
output0
,
output
in
zip
(
node
.
outputs
,
thunk0
.
outputs
,
thunk
.
outputs
):
if
not
self
.
compare_fn
(
output0
[
0
],
output
[
0
]):
if
not
self
.
compare_fn
(
output0
[
0
],
output
[
0
]):
exc
=
DebugException
((
"The variables from
%
s and
%
s for output
%
s are not the same. This happened at step
%
i."
%
(
linker0
,
linker
,
o
,
step
))
+
\
exc
=
DebugException
(
"For more info, inspect this exception's 'debugger', 'output', 'output_value1', 'output_value2', "
\
(
"The variables from
%
s and
%
s for output
%
s are not "
"'step', 'node', 'thunk1', 'thunk2', 'linker1' and 'linker2' fields."
)
"the same. This happened at step
%
i."
%
(
linker0
,
linker
,
o
,
step
))
+
"For more info, inspect this exception's 'debugger', "
"'output', 'output_value1', 'output_value2', 'step', "
"'node', 'thunk1', 'thunk2', 'linker1' "
"and 'linker2' fields."
)
exc
.
debugger
=
self
exc
.
debugger
=
self
exc
.
output
=
o
exc
.
output
=
o
exc
.
output_value1
=
output0
exc
.
output_value1
=
output0
...
@@ -140,8 +152,12 @@ class DebugLinker(gof.WrapLinker):
...
@@ -140,8 +152,12 @@ class DebugLinker(gof.WrapLinker):
exc_type
,
exc_value
,
exc_trace
=
sys
.
exc_info
()
exc_type
,
exc_value
,
exc_trace
=
sys
.
exc_info
()
if
isinstance
(
e
,
DebugException
):
if
isinstance
(
e
,
DebugException
):
raise
raise
exc
=
DebugException
(
e
,
(
"An exception occurred while processing node
%
s at step
%
i of the program."
%
(
node
,
i
))
+
\
exc
=
DebugException
(
"For more info, inspect this exception's 'original_exception', 'debugger', 'step', 'node' and 'thunks' fields."
)
e
,
(
"An exception occurred while processing node
%
s at step
%
i "
"of the program."
%
(
node
,
i
))
+
"For more info, inspect this exception's 'original_exception',"
"'debugger', 'step', 'node' and 'thunks' fields."
)
exc
.
debugger
=
self
exc
.
debugger
=
self
exc
.
original_exception
=
e
exc
.
original_exception
=
e
exc
.
step
=
i
exc
.
step
=
i
...
@@ -169,7 +185,8 @@ def print_input_shapes(i, node, *thunks):
...
@@ -169,7 +185,8 @@ def print_input_shapes(i, node, *thunks):
def
print_input_types
(
i
,
node
,
*
thunks
):
def
print_input_types
(
i
,
node
,
*
thunks
):
print
(
"input types:"
,
", "
.
join
(
str
(
type
(
input
.
value
))
for
input
in
node
.
inputs
))
print
(
"input types:"
,
", "
.
join
(
str
(
type
(
input
.
value
))
for
input
in
node
.
inputs
))
def
print_sep
(
i
,
node
,
*
thunks
):
def
print_sep
(
i
,
node
,
*
thunks
):
...
@@ -192,5 +209,3 @@ def numpy_debug_linker(pre, post=None):
...
@@ -192,5 +209,3 @@ def numpy_debug_linker(pre, post=None):
pre
,
pre
,
post
,
post
,
compare_fn
=
numpy_compare
)
compare_fn
=
numpy_compare
)
theano/sandbox/fourier.py
浏览文件 @
842da1f2
...
@@ -12,7 +12,7 @@ from theano.gof import Op, Apply, generic
...
@@ -12,7 +12,7 @@ from theano.gof import Op, Apply, generic
class
GradTodo
(
Op
):
class
GradTodo
(
Op
):
# TODO : need description for class
__props__
=
()
__props__
=
()
def
make_node
(
self
,
x
):
def
make_node
(
self
,
x
):
...
@@ -24,6 +24,7 @@ grad_todo = GradTodo()
...
@@ -24,6 +24,7 @@ grad_todo = GradTodo()
class
FFT
(
Op
):
class
FFT
(
Op
):
# TODO : need description for parameters
"""
"""
Fast Fourier Transform.
Fast Fourier Transform.
...
@@ -44,7 +45,8 @@ class FFT(Op):
...
@@ -44,7 +45,8 @@ class FFT(Op):
# don't return the plan object in the 'buf' output
# don't return the plan object in the 'buf' output
half
=
False
half
=
False
"""Only return the first half (positive-valued) of the frequency components."""
"""Only return the first half (positive-valued) of the frequency
components."""
__props__
=
(
"half"
,
"inverse"
)
__props__
=
(
"half"
,
"inverse"
)
def
__init__
(
self
,
half
=
False
,
inverse
=
False
):
def
__init__
(
self
,
half
=
False
,
inverse
=
False
):
...
@@ -82,11 +84,13 @@ class FFT(Op):
...
@@ -82,11 +84,13 @@ class FFT(Op):
M
,
N
=
fft
.
shape
M
,
N
=
fft
.
shape
if
axis
==
0
:
if
axis
==
0
:
if
(
M
%
2
):
if
(
M
%
2
):
raise
ValueError
(
'halfFFT on odd-length vectors is undefined'
)
raise
ValueError
(
'halfFFT on odd-length vectors is undefined'
)
spectrogram
[
0
]
=
fft
[
0
:
M
/
2
,
:]
spectrogram
[
0
]
=
fft
[
0
:
M
/
2
,
:]
elif
axis
==
1
:
elif
axis
==
1
:
if
(
N
%
2
):
if
(
N
%
2
):
raise
ValueError
(
'halfFFT on odd-length vectors is undefined'
)
raise
ValueError
(
'halfFFT on odd-length vectors is undefined'
)
spectrogram
[
0
]
=
fft
[:,
0
:
N
/
2
]
spectrogram
[
0
]
=
fft
[:,
0
:
N
/
2
]
else
:
else
:
raise
NotImplementedError
()
raise
NotImplementedError
()
...
@@ -105,6 +109,7 @@ half_ifft = FFT(half=True, inverse=True)
...
@@ -105,6 +109,7 @@ half_ifft = FFT(half=True, inverse=True)
def
dct_matrix
(
rows
,
cols
,
unitary
=
True
):
def
dct_matrix
(
rows
,
cols
,
unitary
=
True
):
# TODO : need description for parameters
"""
"""
Return a (rows x cols) matrix implementing a discrete cosine transform.
Return a (rows x cols) matrix implementing a discrete cosine transform.
...
@@ -115,7 +120,8 @@ def dct_matrix(rows, cols, unitary=True):
...
@@ -115,7 +120,8 @@ def dct_matrix(rows, cols, unitary=True):
col_range
=
numpy
.
arange
(
cols
)
col_range
=
numpy
.
arange
(
cols
)
scale
=
numpy
.
sqrt
(
2.0
/
cols
)
scale
=
numpy
.
sqrt
(
2.0
/
cols
)
for
i
in
xrange
(
rows
):
for
i
in
xrange
(
rows
):
rval
[
i
]
=
numpy
.
cos
(
i
*
(
col_range
*
2
+
1
)
/
(
2.0
*
cols
)
*
numpy
.
pi
)
*
scale
rval
[
i
]
=
numpy
.
cos
(
i
*
(
col_range
*
2
+
1
)
/
(
2.0
*
cols
)
*
numpy
.
pi
)
*
scale
if
unitary
:
if
unitary
:
rval
[
0
]
*=
numpy
.
sqrt
(
0.5
)
rval
[
0
]
*=
numpy
.
sqrt
(
0.5
)
...
...
theano/sandbox/minimal.py
浏览文件 @
842da1f2
...
@@ -9,17 +9,19 @@ from theano.tests import unittest_tools as utt
...
@@ -9,17 +9,19 @@ from theano.tests import unittest_tools as utt
class
Minimal
(
gof
.
Op
):
class
Minimal
(
gof
.
Op
):
# TODO : need description for class
# if the Op has any attributes,
# if the Op has any attributes, consider using them in the eq function.
# consider using them in the eq function. If two Apply nodes have the same inputs and the
# If two Apply nodes have the same inputs and the ops compare equal...
# ops compare equal... then they will be MERGED so they had better have computed the same
# then they will be MERGED so they had better have computed the same thing!
# thing!
def
__init__
(
self
):
def
__init__
(
self
):
# If you put things here, think about whether they change the outputs computed by
# If you put things here, think about whether they change the outputs
# self.perform()
# computed by # self.perform()
# - If they do, then you should take them into consideration in __eq__ and __hash__
# - If they do, then you should take them into consideration in
# - If they do not, then you should not use them in __eq__ and __hash__
# __eq__ and __hash__
# - If they do not, then you should not use them in
# __eq__ and __hash__
super
(
Minimal
,
self
)
.
__init__
()
super
(
Minimal
,
self
)
.
__init__
()
...
...
theano/sandbox/multinomial.py
浏览文件 @
842da1f2
...
@@ -16,6 +16,7 @@ if cuda_available:
...
@@ -16,6 +16,7 @@ if cuda_available:
class
MultinomialFromUniform
(
Op
):
class
MultinomialFromUniform
(
Op
):
# TODO : need description for parameter 'odtype'
"""
"""
Converts samples from a uniform into sample from a multinomial.
Converts samples from a uniform into sample from a multinomial.
...
@@ -197,7 +198,8 @@ class MultinomialFromUniform(Op):
...
@@ -197,7 +198,8 @@ class MultinomialFromUniform(Op):
class
MultinomialWOReplacementFromUniform
(
MultinomialFromUniform
):
class
MultinomialWOReplacementFromUniform
(
MultinomialFromUniform
):
"""
"""
Converts samples from a uniform into sample (without replacement) from a multinomial.
Converts samples from a uniform into sample (without replacement) from a
multinomial.
"""
"""
...
@@ -222,8 +224,8 @@ class MultinomialWOReplacementFromUniform(MultinomialFromUniform):
...
@@ -222,8 +224,8 @@ class MultinomialWOReplacementFromUniform(MultinomialFromUniform):
(
z
,)
=
outs
(
z
,)
=
outs
if
n_samples
>
pvals
.
shape
[
1
]:
if
n_samples
>
pvals
.
shape
[
1
]:
raise
ValueError
(
"Cannot sample without replacement n samples
bigger
"
raise
ValueError
(
"Cannot sample without replacement n samples "
"than the size of the distribution."
)
"
bigger
than the size of the distribution."
)
if
unis
.
shape
[
0
]
!=
pvals
.
shape
[
0
]
*
n_samples
:
if
unis
.
shape
[
0
]
!=
pvals
.
shape
[
0
]
*
n_samples
:
raise
ValueError
(
"unis.shape[0] != pvals.shape[0] * n_samples"
,
raise
ValueError
(
"unis.shape[0] != pvals.shape[0] * n_samples"
,
...
@@ -233,7 +235,8 @@ class MultinomialWOReplacementFromUniform(MultinomialFromUniform):
...
@@ -233,7 +235,8 @@ class MultinomialWOReplacementFromUniform(MultinomialFromUniform):
odtype
=
'int64'
odtype
=
'int64'
else
:
else
:
odtype
=
self
.
odtype
odtype
=
self
.
odtype
if
z
[
0
]
is
None
or
not
numpy
.
all
(
z
[
0
]
.
shape
==
[
pvals
.
shape
[
0
],
n_samples
]):
if
(
z
[
0
]
is
None
or
not
numpy
.
all
(
z
[
0
]
.
shape
==
[
pvals
.
shape
[
0
],
n_samples
])):
z
[
0
]
=
-
1
*
numpy
.
ones
((
pvals
.
shape
[
0
],
n_samples
),
dtype
=
odtype
)
z
[
0
]
=
-
1
*
numpy
.
ones
((
pvals
.
shape
[
0
],
n_samples
),
dtype
=
odtype
)
nb_multi
=
pvals
.
shape
[
0
]
nb_multi
=
pvals
.
shape
[
0
]
...
@@ -249,7 +252,8 @@ class MultinomialWOReplacementFromUniform(MultinomialFromUniform):
...
@@ -249,7 +252,8 @@ class MultinomialWOReplacementFromUniform(MultinomialFromUniform):
cummul
+=
pvals
[
n
,
m
]
cummul
+=
pvals
[
n
,
m
]
if
(
cummul
>
unis_n
):
if
(
cummul
>
unis_n
):
z
[
0
][
n
,
c
]
=
m
z
[
0
][
n
,
c
]
=
m
# set to zero and re-normalize so that it's not selected again
# set to zero and re-normalize so that it's not
# selected again
pvals
[
n
,
m
]
=
0.
pvals
[
n
,
m
]
=
0.
pvals
[
n
]
/=
pvals
[
n
]
.
sum
()
pvals
[
n
]
/=
pvals
[
n
]
.
sum
()
break
break
...
@@ -443,6 +447,7 @@ class GpuMultinomialFromUniform(MultinomialFromUniform, GpuOp):
...
@@ -443,6 +447,7 @@ class GpuMultinomialFromUniform(MultinomialFromUniform, GpuOp):
@local_optimizer
([
MultinomialFromUniform
])
@local_optimizer
([
MultinomialFromUniform
])
def
local_gpu_multinomial
(
node
):
def
local_gpu_multinomial
(
node
):
# TODO : need description for function
if
type
(
node
.
op
)
is
MultinomialFromUniform
:
if
type
(
node
.
op
)
is
MultinomialFromUniform
:
if
len
(
node
.
inputs
)
==
2
:
if
len
(
node
.
inputs
)
==
2
:
p
,
u
=
node
.
inputs
p
,
u
=
node
.
inputs
...
...
theano/sandbox/neighbourhoods.py
浏览文件 @
842da1f2
...
@@ -116,7 +116,8 @@ class NeighbourhoodsFromImages(Op):
...
@@ -116,7 +116,8 @@ class NeighbourhoodsFromImages(Op):
return
dims
,
num_strides
return
dims
,
num_strides
# for inverse mode
# for inverse mode
# "output" here actually referes to the Op's input shape (but it's inverse mode)
# "output" here actually referes to the Op's input shape (but it's inverse
# mode)
def
in_shape
(
self
,
output_shape
):
def
in_shape
(
self
,
output_shape
):
out_dims
=
list
(
output_shape
[:
self
.
n_dims_before
])
out_dims
=
list
(
output_shape
[:
self
.
n_dims_before
])
num_strides
=
[]
num_strides
=
[]
...
@@ -168,7 +169,8 @@ class NeighbourhoodsFromImages(Op):
...
@@ -168,7 +169,8 @@ class NeighbourhoodsFromImages(Op):
for
dim
in
self
.
dims_neighbourhoods
:
for
dim
in
self
.
dims_neighbourhoods
:
prod
*=
dim
prod
*=
dim
if
x
.
shape
[
-
1
]
!=
prod
:
if
x
.
shape
[
-
1
]
!=
prod
:
raise
ValueError
(
"Last dimension of neighbourhoods (
%
s) is not"
raise
ValueError
(
"Last dimension of neighbourhoods (
%
s) is not"
" the product of the neighbourhoods dimensions"
" the product of the neighbourhoods dimensions"
" (
%
s)"
%
(
str
(
x
.
shape
[
-
1
]),
str
(
prod
)))
" (
%
s)"
%
(
str
(
x
.
shape
[
-
1
]),
str
(
prod
)))
else
:
else
:
...
@@ -195,6 +197,7 @@ class NeighbourhoodsFromImages(Op):
...
@@ -195,6 +197,7 @@ class NeighbourhoodsFromImages(Op):
exec
(
self
.
code
)
exec
(
self
.
code
)
def
make_py_code
(
self
):
def
make_py_code
(
self
):
# TODO : need description for method and return
code
=
self
.
_py_outerloops
()
code
=
self
.
_py_outerloops
()
for
i
in
xrange
(
len
(
self
.
strides
)):
for
i
in
xrange
(
len
(
self
.
strides
)):
code
+=
self
.
_py_innerloop
(
i
)
code
+=
self
.
_py_innerloop
(
i
)
...
@@ -202,6 +205,7 @@ class NeighbourhoodsFromImages(Op):
...
@@ -202,6 +205,7 @@ class NeighbourhoodsFromImages(Op):
return
code
,
builtins
.
compile
(
code
,
'<string>'
,
'exec'
)
return
code
,
builtins
.
compile
(
code
,
'<string>'
,
'exec'
)
def
_py_outerloops
(
self
):
def
_py_outerloops
(
self
):
# TODO : need description for method, parameter and return
code_before
=
""
code_before
=
""
for
dim_idx
in
xrange
(
self
.
n_dims_before
):
for
dim_idx
in
xrange
(
self
.
n_dims_before
):
code_before
+=
(
'
\t
'
*
(
dim_idx
))
+
\
code_before
+=
(
'
\t
'
*
(
dim_idx
))
+
\
...
@@ -210,6 +214,7 @@ class NeighbourhoodsFromImages(Op):
...
@@ -210,6 +214,7 @@ class NeighbourhoodsFromImages(Op):
return
code_before
return
code_before
def
_py_innerloop
(
self
,
inner_dim_no
):
def
_py_innerloop
(
self
,
inner_dim_no
):
# TODO : need description for method, parameter and return
base_indent
=
(
'
\t
'
*
(
self
.
n_dims_before
+
inner_dim_no
*
2
))
base_indent
=
(
'
\t
'
*
(
self
.
n_dims_before
+
inner_dim_no
*
2
))
code_before
=
base_indent
+
\
code_before
=
base_indent
+
\
"for stride_idx_
%
d in xrange(num_strides[
%
d]):
\n
"
%
\
"for stride_idx_
%
d in xrange(num_strides[
%
d]):
\n
"
%
\
...
@@ -229,10 +234,12 @@ class NeighbourhoodsFromImages(Op):
...
@@ -229,10 +234,12 @@ class NeighbourhoodsFromImages(Op):
return
code_before
return
code_before
def
_py_flattened_idx
(
self
):
def
_py_flattened_idx
(
self
):
# TODO : need description for method and return
return
"+"
.
join
([
"neigh_strides[
%
d]*neigh_idx_
%
d"
%
(
i
,
i
)
return
"+"
.
join
([
"neigh_strides[
%
d]*neigh_idx_
%
d"
%
(
i
,
i
)
for
i
in
xrange
(
len
(
self
.
strides
))])
for
i
in
xrange
(
len
(
self
.
strides
))])
def
_py_assignment
(
self
):
def
_py_assignment
(
self
):
# TODO : need description for method and return
input_idx
=
""
.
join
([
"outer_idx_
%
d,"
%
(
i
,)
input_idx
=
""
.
join
([
"outer_idx_
%
d,"
%
(
i
,)
for
i
in
xrange
(
self
.
n_dims_before
)])
for
i
in
xrange
(
self
.
n_dims_before
)])
input_idx
+=
""
.
join
([
"dim_
%
d_offset+neigh_idx_
%
d,"
%
input_idx
+=
""
.
join
([
"dim_
%
d_offset+neigh_idx_
%
d,"
%
...
@@ -259,6 +266,7 @@ class NeighbourhoodsFromImages(Op):
...
@@ -259,6 +266,7 @@ class NeighbourhoodsFromImages(Op):
class
ImagesFromNeighbourhoods
(
NeighbourhoodsFromImages
):
class
ImagesFromNeighbourhoods
(
NeighbourhoodsFromImages
):
# TODO : need description for class, parameters
def
__init__
(
self
,
n_dims_before
,
dims_neighbourhoods
,
def
__init__
(
self
,
n_dims_before
,
dims_neighbourhoods
,
strides
=
None
,
ignore_border
=
False
):
strides
=
None
,
ignore_border
=
False
):
NeighbourhoodsFromImages
.
__init__
(
self
,
n_dims_before
,
NeighbourhoodsFromImages
.
__init__
(
self
,
n_dims_before
,
...
...
theano/sandbox/rng_mrg.py
浏览文件 @
842da1f2
...
@@ -33,11 +33,13 @@ from theano.sandbox.gpuarray.fp16_help import write_w
...
@@ -33,11 +33,13 @@ from theano.sandbox.gpuarray.fp16_help import write_w
def
matVecModM
(
A
,
s
,
m
):
def
matVecModM
(
A
,
s
,
m
):
# TODO : need description for method, parameter and return
assert
A
.
dtype
==
'int64'
assert
A
.
dtype
==
'int64'
return
numpy
.
int32
(
numpy
.
sum
((
A
*
s
)
%
m
,
1
)
%
m
)
return
numpy
.
int32
(
numpy
.
sum
((
A
*
s
)
%
m
,
1
)
%
m
)
def
multMatVect
(
v
,
A
,
m1
,
B
,
m2
):
def
multMatVect
(
v
,
A
,
m1
,
B
,
m2
):
# TODO : need description for parameter and return
"""
"""
Multiply the first half of v by A with a modulo of m1 and the second half
Multiply the first half of v by A with a modulo of m1 and the second half
by B with a modulo of m2.
by B with a modulo of m2.
...
@@ -193,13 +195,13 @@ class DotModulo(Op):
...
@@ -193,13 +195,13 @@ class DotModulo(Op):
# MRG31k3p
# MRG31k3p
# generator constants :
# generator constants :
M1
=
numpy
.
asarray
(
numpy
.
int32
(
2147483647
))
#2^31 - 1
M1
=
numpy
.
asarray
(
numpy
.
int32
(
2147483647
))
#
2^31 - 1
M2
=
numpy
.
asarray
(
numpy
.
int32
(
2147462579
))
#2^31 - 21069
M2
=
numpy
.
asarray
(
numpy
.
int32
(
2147462579
))
#
2^31 - 21069
MASK12
=
numpy
.
int32
(
511
)
#2^9 - 1
MASK12
=
numpy
.
int32
(
511
)
#
2^9 - 1
MASK13
=
numpy
.
int32
(
16777215
)
#2^24 - 1
MASK13
=
numpy
.
int32
(
16777215
)
#
2^24 - 1
MASK2
=
numpy
.
int32
(
65535
)
#2^16 - 1
MASK2
=
numpy
.
int32
(
65535
)
#
2^16 - 1
MULT2
=
numpy
.
int32
(
21069
)
MULT2
=
numpy
.
int32
(
21069
)
NORM
=
4.656612873077392578125e-10
#1./2^31
NORM
=
4.656612873077392578125e-10
#
1./2^31
# A1p0 = numpy.asarray([[0, 4194304, 129], [1, 0, 0], [0, 1, 0]],
# A1p0 = numpy.asarray([[0, 4194304, 129], [1, 0, 0], [0, 1, 0]],
# dtype='int64')
# dtype='int64')
...
@@ -229,14 +231,17 @@ np_int32_vals = [numpy.int32(i) for i in (0, 7, 9, 15, 16, 22, 24)]
...
@@ -229,14 +231,17 @@ np_int32_vals = [numpy.int32(i) for i in (0, 7, 9, 15, 16, 22, 24)]
def
ff_2p134
(
rstate
):
def
ff_2p134
(
rstate
):
# TODO : need description for method, parameter and return
return
multMatVect
(
rstate
,
A1p134
,
M1
,
A2p134
,
M2
)
return
multMatVect
(
rstate
,
A1p134
,
M1
,
A2p134
,
M2
)
def
ff_2p72
(
rstate
):
def
ff_2p72
(
rstate
):
# TODO : need description for method, parameter and return
return
multMatVect
(
rstate
,
A1p72
,
M1
,
A2p72
,
M2
)
return
multMatVect
(
rstate
,
A1p72
,
M1
,
A2p72
,
M2
)
def
mrg_next_value
(
rstate
,
new_rstate
):
def
mrg_next_value
(
rstate
,
new_rstate
):
# TODO : need description for method, parameter and return
x11
,
x12
,
x13
,
x21
,
x22
,
x23
=
rstate
x11
,
x12
,
x13
,
x21
,
x22
,
x23
=
rstate
assert
type
(
x11
)
==
numpy
.
int32
assert
type
(
x11
)
==
numpy
.
int32
...
@@ -286,7 +291,7 @@ def mrg_next_value(rstate, new_rstate):
...
@@ -286,7 +291,7 @@ def mrg_next_value(rstate, new_rstate):
class
mrg_uniform_base
(
Op
):
class
mrg_uniform_base
(
Op
):
# TODO : need description for class, parameter
__props__
=
(
"output_type"
,
"inplace"
)
__props__
=
(
"output_type"
,
"inplace"
)
def
__init__
(
self
,
output_type
,
inplace
=
False
):
def
__init__
(
self
,
output_type
,
inplace
=
False
):
...
@@ -314,9 +319,9 @@ class mrg_uniform_base(Op):
...
@@ -314,9 +319,9 @@ class mrg_uniform_base(Op):
[
rstate
.
type
(),
self
.
output_type
()])
[
rstate
.
type
(),
self
.
output_type
()])
def
grad
(
self
,
inputs
,
ograd
):
def
grad
(
self
,
inputs
,
ograd
):
return
[
gradient
.
grad_undefined
(
return
[
gradient
.
grad_undefined
(
self
,
k
,
inp
,
self
,
k
,
inp
,
'No gradient defined through '
'No gradient defined through
random sampling op'
)
'
random sampling op'
)
for
k
,
inp
in
enumerate
(
inputs
)]
for
k
,
inp
in
enumerate
(
inputs
)]
def
R_op
(
self
,
inputs
,
eval_points
):
def
R_op
(
self
,
inputs
,
eval_points
):
...
@@ -371,9 +376,12 @@ class mrg_uniform(mrg_uniform_base):
...
@@ -371,9 +376,12 @@ class mrg_uniform(mrg_uniform_base):
assert
isinstance
(
node
.
inputs
[
0
]
.
type
,
TensorType
)
assert
isinstance
(
node
.
inputs
[
0
]
.
type
,
TensorType
)
o_rstate
,
o_sample
=
out
o_rstate
,
o_sample
=
out
if
self
.
inplace
:
if
self
.
inplace
:
o_rstate_requirement
=
'NPY_ARRAY_C_CONTIGUOUS|NPY_ARRAY_ALIGNED'
o_rstate_requirement
=
(
'NPY_ARRAY_C_CONTIGUOUS|NPY_ARRAY_ALIGNED'
)
else
:
else
:
o_rstate_requirement
=
'NPY_ARRAY_ENSURECOPY|NPY_ARRAY_C_CONTIGUOUS|NPY_ARRAY_ALIGNED'
o_rstate_requirement
=
(
'NPY_ARRAY_ENSURECOPY|NPY_ARRAY_C_CONTIGUOUS|'
'NPY_ARRAY_ALIGNED'
)
ndim
=
self
.
output_type
.
ndim
ndim
=
self
.
output_type
.
ndim
o_type_num
=
numpy
.
asarray
(
0
,
dtype
=
self
.
output_type
.
dtype
)
.
dtype
.
num
o_type_num
=
numpy
.
asarray
(
0
,
dtype
=
self
.
output_type
.
dtype
)
.
dtype
.
num
fail
=
sub
[
'fail'
]
fail
=
sub
[
'fail'
]
...
@@ -1021,6 +1029,7 @@ class GPUA_mrg_uniform(GpuKernelBase, mrg_uniform_base):
...
@@ -1021,6 +1029,7 @@ class GPUA_mrg_uniform(GpuKernelBase, mrg_uniform_base):
def
guess_n_streams
(
size
,
warn
=
False
):
def
guess_n_streams
(
size
,
warn
=
False
):
# TODO : need description for parameter 'size'
"""
"""
Return a guess at a good number of streams.
Return a guess at a good number of streams.
...
@@ -1044,8 +1053,9 @@ def guess_n_streams(size, warn=False):
...
@@ -1044,8 +1053,9 @@ def guess_n_streams(size, warn=False):
r
=
r
//
6
# chosen as fastest for rbm_benchmark
r
=
r
//
6
# chosen as fastest for rbm_benchmark
# The purpose of sampling from many streams is to be able to use
# The purpose of sampling from many streams is to be able to use
# the GPU to its full capacity. It just wastes RAM and stream-initialization time to
# the GPU to its full capacity. It just wastes RAM and
# allocate more streams than necessary for the GPU.
# stream-initialization time to allocate more streams than necessary
# for the GPU.
# XXX: This number is chosen to be good for 280 and 480 architectures,
# XXX: This number is chosen to be good for 280 and 480 architectures,
# Better would be to use pycuda to query the number of
# Better would be to use pycuda to query the number of
# processors on the GPU device,
# processors on the GPU device,
...
@@ -1053,14 +1063,15 @@ def guess_n_streams(size, warn=False):
...
@@ -1053,14 +1063,15 @@ def guess_n_streams(size, warn=False):
return
min
(
r
,
60
*
256
)
return
min
(
r
,
60
*
256
)
else
:
else
:
if
warn
:
if
warn
:
warnings
.
warn
(
(
warnings
.
warn
(
"MRG_RandomStreams Can't determine #streams from
"
(
"MRG_RandomStreams Can't determine #streams
"
"
size (
%
s), guessing 60*256"
)
%
str
(
size
),
"from
size (
%
s), guessing 60*256"
)
%
str
(
size
),
stacklevel
=
3
)
stacklevel
=
3
)
return
60
*
256
return
60
*
256
class
MRG_RandomStreams
(
object
):
class
MRG_RandomStreams
(
object
):
# TODO : need description for parameter 'use_cuda'
"""
"""
Module component with similar interface to numpy.random
Module component with similar interface to numpy.random
(numpy.random.RandomState).
(numpy.random.RandomState).
...
@@ -1077,11 +1088,13 @@ class MRG_RandomStreams(object):
...
@@ -1077,11 +1088,13 @@ class MRG_RandomStreams(object):
"""
"""
def
updates
(
self
):
def
updates
(
self
):
# TODO : need description for method and return
return
list
(
self
.
state_updates
)
return
list
(
self
.
state_updates
)
def
__init__
(
self
,
seed
=
12345
,
use_cuda
=
None
):
def
__init__
(
self
,
seed
=
12345
,
use_cuda
=
None
):
# A list of pairs of the form (input_r, output_r), representing the
# A list of pairs of the form (input_r, output_r), representing the
# update rules of all the random states generated by this RandomStreams.
# update rules of all the random states generated
# by this RandomStreams.
self
.
state_updates
=
[]
self
.
state_updates
=
[]
super
(
MRG_RandomStreams
,
self
)
.
__init__
()
super
(
MRG_RandomStreams
,
self
)
.
__init__
()
...
@@ -1097,6 +1110,7 @@ class MRG_RandomStreams(object):
...
@@ -1097,6 +1110,7 @@ class MRG_RandomStreams(object):
self
.
use_cuda
=
use_cuda
self
.
use_cuda
=
use_cuda
def
set_rstate
(
self
,
seed
):
def
set_rstate
(
self
,
seed
):
# TODO : need description for method, parameter
if
isinstance
(
seed
,
int
):
if
isinstance
(
seed
,
int
):
if
seed
==
0
:
if
seed
==
0
:
raise
ValueError
(
'seed should not be 0'
,
seed
)
raise
ValueError
(
'seed should not be 0'
,
seed
)
...
@@ -1158,11 +1172,12 @@ class MRG_RandomStreams(object):
...
@@ -1158,11 +1172,12 @@ class MRG_RandomStreams(object):
start.
start.
"""
"""
#self.rstate = ff_2p134(self.rstate)
#
self.rstate = ff_2p134(self.rstate)
self
.
rstate
=
multMatVect
(
self
.
rstate
,
A1p134
,
M1
,
A2p134
,
M2
)
self
.
rstate
=
multMatVect
(
self
.
rstate
,
A1p134
,
M1
,
A2p134
,
M2
)
assert
self
.
rstate
.
dtype
==
numpy
.
int32
assert
self
.
rstate
.
dtype
==
numpy
.
int32
def
get_substream_rstates
(
self
,
n_streams
,
dtype
,
inc_rstate
=
True
):
def
get_substream_rstates
(
self
,
n_streams
,
dtype
,
inc_rstate
=
True
):
# TODO : need description for parameter and return
"""
"""
Initialize a matrix in which each row is a MRG stream state,
Initialize a matrix in which each row is a MRG stream state,
and they are spaced by 2**72 samples.
and they are spaced by 2**72 samples.
...
@@ -1186,7 +1201,7 @@ class MRG_RandomStreams(object):
...
@@ -1186,7 +1201,7 @@ class MRG_RandomStreams(object):
f
.
input_storage
[
5
]
.
storage
[
0
]
=
M2
f
.
input_storage
[
5
]
.
storage
[
0
]
=
M2
for
i
in
xrange
(
1
,
n_streams
):
for
i
in
xrange
(
1
,
n_streams
):
# Inline the following call to bypass Python overhead
# Inline the following call to bypass Python overhead
#rval[i] = ff_2p72(rval[i - 1])
#
rval[i] = ff_2p72(rval[i - 1])
v
=
rval
[
i
-
1
]
v
=
rval
[
i
-
1
]
f
.
input_storage
[
1
]
.
storage
[
0
]
=
v
[:
3
]
f
.
input_storage
[
1
]
.
storage
[
0
]
=
v
[:
3
]
f
.
input_storage
[
4
]
.
storage
[
0
]
=
v
[
3
:]
f
.
input_storage
[
4
]
.
storage
[
0
]
=
v
[
3
:]
...
@@ -1208,9 +1223,11 @@ class MRG_RandomStreams(object):
...
@@ -1208,9 +1223,11 @@ class MRG_RandomStreams(object):
return
rval
return
rval
def
n_streams
(
self
,
size
):
def
n_streams
(
self
,
size
):
# TODO : need description for method, parameter and return
return
guess_n_streams
(
size
)
return
guess_n_streams
(
size
)
def
pretty_return
(
self
,
node_rstate
,
new_rstate
,
sample
,
size
,
nstreams
):
def
pretty_return
(
self
,
node_rstate
,
new_rstate
,
sample
,
size
,
nstreams
):
# TODO : need description for method, parameter and return
sample
.
rstate
=
node_rstate
sample
.
rstate
=
node_rstate
sample
.
update
=
(
node_rstate
,
new_rstate
)
sample
.
update
=
(
node_rstate
,
new_rstate
)
self
.
state_updates
.
append
((
node_rstate
,
new_rstate
,
size
,
nstreams
))
self
.
state_updates
.
append
((
node_rstate
,
new_rstate
,
size
,
nstreams
))
...
@@ -1219,6 +1236,7 @@ class MRG_RandomStreams(object):
...
@@ -1219,6 +1236,7 @@ class MRG_RandomStreams(object):
def
uniform
(
self
,
size
,
low
=
0.0
,
high
=
1.0
,
ndim
=
None
,
dtype
=
None
,
def
uniform
(
self
,
size
,
low
=
0.0
,
high
=
1.0
,
ndim
=
None
,
dtype
=
None
,
nstreams
=
None
):
nstreams
=
None
):
# TODO : need description for parameter 'size', 'ndim', 'nstreams'
"""
"""
Sample a tensor of given size whose element from a uniform
Sample a tensor of given size whose element from a uniform
distribution between low and high.
distribution between low and high.
...
@@ -1306,6 +1324,7 @@ class MRG_RandomStreams(object):
...
@@ -1306,6 +1324,7 @@ class MRG_RandomStreams(object):
def
binomial
(
self
,
size
=
None
,
n
=
1
,
p
=
0.5
,
ndim
=
None
,
dtype
=
'int64'
,
def
binomial
(
self
,
size
=
None
,
n
=
1
,
p
=
0.5
,
ndim
=
None
,
dtype
=
'int64'
,
nstreams
=
None
):
nstreams
=
None
):
# TODO : need description for method, parameter and return
if
n
==
1
:
if
n
==
1
:
if
dtype
==
'float32'
and
self
.
use_cuda
:
if
dtype
==
'float32'
and
self
.
use_cuda
:
x
=
self
.
uniform
(
size
=
size
,
dtype
=
dtype
,
nstreams
=
nstreams
)
x
=
self
.
uniform
(
size
=
size
,
dtype
=
dtype
,
nstreams
=
nstreams
)
...
@@ -1317,6 +1336,7 @@ class MRG_RandomStreams(object):
...
@@ -1317,6 +1336,7 @@ class MRG_RandomStreams(object):
def
multinomial
(
self
,
size
=
None
,
n
=
1
,
pvals
=
None
,
ndim
=
None
,
dtype
=
'int64'
,
def
multinomial
(
self
,
size
=
None
,
n
=
1
,
pvals
=
None
,
ndim
=
None
,
dtype
=
'int64'
,
nstreams
=
None
):
nstreams
=
None
):
# TODO : need description for parameter and return
"""
"""
Sample `n` (`n` needs to be >= 1, default 1) times from a multinomial
Sample `n` (`n` needs to be >= 1, default 1) times from a multinomial
distribution defined by probabilities pvals.
distribution defined by probabilities pvals.
...
@@ -1347,15 +1367,15 @@ class MRG_RandomStreams(object):
...
@@ -1347,15 +1367,15 @@ class MRG_RandomStreams(object):
size
)
size
)
if
size
is
not
None
:
if
size
is
not
None
:
raise
ValueError
(
"Provided a size argument to "
raise
ValueError
(
"MRG_RandomStreams.multinomial, which does not use
"
"Provided a size argument to MRG_RandomStreams.multinomial,
"
"
the size argument."
)
"which does not use
the size argument."
)
if
ndim
is
not
None
:
if
ndim
is
not
None
:
raise
ValueError
(
"Provided an ndim argument to "
raise
ValueError
(
"MRG_RandomStreams.multinomial, which does not use
"
"Provided an ndim argument to MRG_RandomStreams.multinomial,
"
"
the ndim argument."
)
"which does not use
the ndim argument."
)
if
pvals
.
ndim
==
2
:
if
pvals
.
ndim
==
2
:
size
=
pvals
[:,
0
]
.
shape
*
n
size
=
pvals
[:,
0
]
.
shape
*
n
unis
=
self
.
uniform
(
size
=
size
,
ndim
=
1
,
nstreams
=
nstreams
)
unis
=
self
.
uniform
(
size
=
size
,
ndim
=
1
,
nstreams
=
nstreams
)
op
=
multinomial
.
MultinomialFromUniform
(
dtype
)
op
=
multinomial
.
MultinomialFromUniform
(
dtype
)
n_samples
=
as_tensor_variable
(
n
)
n_samples
=
as_tensor_variable
(
n
)
...
@@ -1364,8 +1384,9 @@ class MRG_RandomStreams(object):
...
@@ -1364,8 +1384,9 @@ class MRG_RandomStreams(object):
raise
NotImplementedError
((
"MRG_RandomStreams.multinomial only"
raise
NotImplementedError
((
"MRG_RandomStreams.multinomial only"
" implemented for pvals.ndim = 2"
))
" implemented for pvals.ndim = 2"
))
def
multinomial_wo_replacement
(
self
,
size
=
None
,
n
=
1
,
pvals
=
None
,
ndim
=
None
,
dtype
=
'int64'
,
def
multinomial_wo_replacement
(
self
,
size
=
None
,
n
=
1
,
pvals
=
None
,
nstreams
=
None
):
ndim
=
None
,
dtype
=
'int64'
,
nstreams
=
None
):
# TODO : need description for parameter
"""
"""
Sample `n` times *WITHOUT replacement* from a multinomial distribution
Sample `n` times *WITHOUT replacement* from a multinomial distribution
defined by probabilities pvals, and returns the indices of the sampled
defined by probabilities pvals, and returns the indices of the sampled
...
@@ -1395,25 +1416,27 @@ class MRG_RandomStreams(object):
...
@@ -1395,25 +1416,27 @@ class MRG_RandomStreams(object):
if
size
is
not
None
:
if
size
is
not
None
:
raise
ValueError
(
"Provided a size argument to "
raise
ValueError
(
"Provided a size argument to "
"MRG_RandomStreams.multinomial_wo_replacement,
which does not use
"
"MRG_RandomStreams.multinomial_wo_replacement, "
"the size argument."
)
"
which does not use
the size argument."
)
if
ndim
is
not
None
:
if
ndim
is
not
None
:
raise
ValueError
(
"Provided an ndim argument to "
raise
ValueError
(
"Provided an ndim argument to "
"MRG_RandomStreams.multinomial_wo_replacement,
which does not use
"
"MRG_RandomStreams.multinomial_wo_replacement, "
"the ndim argument."
)
"
which does not use
the ndim argument."
)
if
pvals
.
ndim
==
2
:
if
pvals
.
ndim
==
2
:
# size = [pvals.shape[0], as_tensor_variable(n)]
# size = [pvals.shape[0], as_tensor_variable(n)]
size
=
pvals
[:,
0
]
.
shape
*
n
size
=
pvals
[:,
0
]
.
shape
*
n
unis
=
self
.
uniform
(
size
=
size
,
ndim
=
1
,
nstreams
=
nstreams
)
unis
=
self
.
uniform
(
size
=
size
,
ndim
=
1
,
nstreams
=
nstreams
)
op
=
multinomial
.
MultinomialWOReplacementFromUniform
(
dtype
)
op
=
multinomial
.
MultinomialWOReplacementFromUniform
(
dtype
)
n_samples
=
as_tensor_variable
(
n
)
n_samples
=
as_tensor_variable
(
n
)
return
op
(
pvals
,
unis
,
n_samples
)
return
op
(
pvals
,
unis
,
n_samples
)
else
:
else
:
raise
NotImplementedError
((
"MRG_RandomStreams.multinomial_wo_replacement only"
raise
NotImplementedError
(
" implemented for pvals.ndim = 2"
))
"MRG_RandomStreams.multinomial_wo_replacement only implemented"
" for pvals.ndim = 2"
)
def
normal
(
self
,
size
,
avg
=
0.0
,
std
=
1.0
,
ndim
=
None
,
def
normal
(
self
,
size
,
avg
=
0.0
,
std
=
1.0
,
ndim
=
None
,
dtype
=
None
,
nstreams
=
None
):
dtype
=
None
,
nstreams
=
None
):
# TODO : need description for method
"""
"""
Parameters
Parameters
----------
----------
...
@@ -1443,7 +1466,8 @@ class MRG_RandomStreams(object):
...
@@ -1443,7 +1466,8 @@ class MRG_RandomStreams(object):
evened
=
False
evened
=
False
constant
=
False
constant
=
False
if
isinstance
(
size
,
tuple
)
and
all
([
isinstance
(
i
,
(
numpy
.
integer
,
int
))
for
i
in
size
]):
if
(
isinstance
(
size
,
tuple
)
and
all
([
isinstance
(
i
,
(
numpy
.
integer
,
int
))
for
i
in
size
])):
constant
=
True
constant
=
True
# Force dtype because it defaults to float when size is empty
# Force dtype because it defaults to float when size is empty
n_samples
=
numpy
.
prod
(
size
,
dtype
=
'int64'
)
n_samples
=
numpy
.
prod
(
size
,
dtype
=
'int64'
)
...
@@ -1464,16 +1488,18 @@ class MRG_RandomStreams(object):
...
@@ -1464,16 +1488,18 @@ class MRG_RandomStreams(object):
U1
=
flattened
[:
prod
(
flattened
.
shape
)
//
2
]
U1
=
flattened
[:
prod
(
flattened
.
shape
)
//
2
]
U2
=
flattened
[
prod
(
flattened
.
shape
)
//
2
:]
U2
=
flattened
[
prod
(
flattened
.
shape
)
//
2
:]
#normal_samples = zeros_like(flattened)
#
normal_samples = zeros_like(flattened)
sqrt_ln_U1
=
sqrt
(
-
2.0
*
log
(
U1
))
sqrt_ln_U1
=
sqrt
(
-
2.0
*
log
(
U1
))
# TypeError: 'TensorVariable' object does not support item assignment
# TypeError: 'TensorVariable' object does not support item assignment
# so this doesn't work...
# so this doesn't work...
#normal_samples[:n_samples/2] = sqrt_ln_U1 * cos(2.0*numpy.pi*U2)
#
normal_samples[:n_samples/2] = sqrt_ln_U1 * cos(2.0*numpy.pi*U2)
#normal_samples[n_samples/2:] = sqrt_ln_U1 * sin(2.0*numpy.pi*U2)
#
normal_samples[n_samples/2:] = sqrt_ln_U1 * sin(2.0*numpy.pi*U2)
# so trying this instead
# so trying this instead
first_half
=
sqrt_ln_U1
*
cos
(
numpy
.
array
(
2.0
*
numpy
.
pi
,
dtype
=
dtype
)
*
U2
)
first_half
=
sqrt_ln_U1
*
cos
(
second_half
=
sqrt_ln_U1
*
sin
(
numpy
.
array
(
2.0
*
numpy
.
pi
,
dtype
=
dtype
)
*
U2
)
numpy
.
array
(
2.0
*
numpy
.
pi
,
dtype
=
dtype
)
*
U2
)
second_half
=
sqrt_ln_U1
*
sin
(
numpy
.
array
(
2.0
*
numpy
.
pi
,
dtype
=
dtype
)
*
U2
)
normal_samples
=
join
(
0
,
first_half
,
second_half
)
normal_samples
=
join
(
0
,
first_half
,
second_half
)
final_samples
=
None
final_samples
=
None
...
@@ -1501,6 +1527,7 @@ from theano.sandbox.gpuarray.opt import (register_opt as register_gpua,
...
@@ -1501,6 +1527,7 @@ from theano.sandbox.gpuarray.opt import (register_opt as register_gpua,
@register_gpua
(
'fast_compile'
)
@register_gpua
(
'fast_compile'
)
@local_optimizer
([
mrg_uniform
])
@local_optimizer
([
mrg_uniform
])
def
local_gpua_mrg
(
node
):
def
local_gpua_mrg
(
node
):
# TODO : need description for function
if
(
type
(
node
.
op
)
==
mrg_uniform
and
if
(
type
(
node
.
op
)
==
mrg_uniform
and
isinstance
(
node
.
inputs
[
0
]
.
type
,
GpuArrayType
)):
isinstance
(
node
.
inputs
[
0
]
.
type
,
GpuArrayType
)):
outs
=
GPUA_mrg_uniform
.
new
(
node
.
inputs
[
0
],
outs
=
GPUA_mrg_uniform
.
new
(
node
.
inputs
[
0
],
...
@@ -1515,6 +1542,7 @@ MRG_RNGs = (mrg_uniform, GPU_mrg_uniform, GPUA_mrg_uniform)
...
@@ -1515,6 +1542,7 @@ MRG_RNGs = (mrg_uniform, GPU_mrg_uniform, GPUA_mrg_uniform)
@local_optimizer
(
MRG_RNGs
)
@local_optimizer
(
MRG_RNGs
)
def
mrg_random_make_inplace
(
node
):
def
mrg_random_make_inplace
(
node
):
op
=
node
.
op
op
=
node
.
op
if
isinstance
(
op
,
MRG_RNGs
)
and
not
op
.
inplace
:
if
isinstance
(
op
,
MRG_RNGs
)
and
not
op
.
inplace
:
# op might be gpu version
# op might be gpu version
...
...
theano/sandbox/softsign.py
浏览文件 @
842da1f2
...
@@ -4,6 +4,7 @@ import theano.tensor
...
@@ -4,6 +4,7 @@ import theano.tensor
class
ScalarSoftsign
(
theano
.
scalar
.
UnaryScalarOp
):
class
ScalarSoftsign
(
theano
.
scalar
.
UnaryScalarOp
):
# TODO : need description for class
@staticmethod
@staticmethod
def
static_impl
(
x
):
def
static_impl
(
x
):
return
x
/
(
1.0
+
abs
(
x
))
return
x
/
(
1.0
+
abs
(
x
))
...
...
theano/sandbox/solve.py
浏览文件 @
842da1f2
...
@@ -24,7 +24,8 @@ class Solve(gof.Op):
...
@@ -24,7 +24,8 @@ class Solve(gof.Op):
# sym_pos, lower, overwrite_a, overwrite_b
# sym_pos, lower, overwrite_a, overwrite_b
# TODO: Add C code that calls the underlying LAPACK routines
# TODO: Add C code that calls the underlying LAPACK routines
# and keeps a memory workspace from call to call as a non-default Op output
# and keeps a memory workspace from call to call as a non-default Op
# output
def
__eq__
(
self
,
other
):
def
__eq__
(
self
,
other
):
return
type
(
self
)
==
type
(
other
)
return
type
(
self
)
==
type
(
other
)
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论