Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
075ece1e
提交
075ece1e
authored
7月 05, 2012
作者:
Nicolas Bouchard
提交者:
Frederic
7月 06, 2012
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Move optimizations and tests.
上级
4791f726
显示空白字符变更
内嵌
并排
正在显示
5 个修改的文件
包含
230 行增加
和
242 行删除
+230
-242
opt.py
theano/sparse/opt.py
+145
-9
sp2.py
theano/sparse/sandbox/sp2.py
+7
-119
test_basic.py
theano/sparse/tests/test_basic.py
+2
-2
test_opt.py
theano/sparse/tests/test_opt.py
+76
-11
test_sp2.py
theano/sparse/tests/test_sp2.py
+0
-101
没有找到文件。
theano/sparse/opt.py
浏览文件 @
075ece1e
...
@@ -7,9 +7,9 @@ from theano.sparse import (CSC, CSR, csm_properties, Remove0,
...
@@ -7,9 +7,9 @@ from theano.sparse import (CSC, CSR, csm_properties, Remove0,
register_specialize
,
register_specialize
,
csm_grad
,
csm_grad_c
,
csm_grad
,
csm_grad_c
,
usmm_csc_dense
,
usmm
)
usmm_csc_dense
,
usmm
)
from
theano.sparse
import
basic
as
sparse
from
basic
import
(
_structured_dot
,
_dot
,
from
basic
import
_is_sparse_variable
_is_sparse_variable
,
usmm_csc_dense_inplace
)
# This is tested in tests/test_basic.py:UsmmTests
# This is tested in tests/test_basic.py:UsmmTests
...
@@ -18,7 +18,7 @@ local_usmm = gof.opt.PatternSub(
...
@@ -18,7 +18,7 @@ local_usmm = gof.opt.PatternSub(
(
theano
.
tensor
.
mul
,
(
theano
.
tensor
.
mul
,
{
'pattern'
:
'alpha'
,
{
'pattern'
:
'alpha'
,
'constraint'
:
lambda
expr
:
numpy
.
all
(
expr
.
type
.
broadcastable
)},
'constraint'
:
lambda
expr
:
numpy
.
all
(
expr
.
type
.
broadcastable
)},
(
_dot
,
'x'
,
'y'
))),
(
sparse
.
_dot
,
'x'
,
'y'
))),
(
usmm
,
(
theano
.
tensor
.
neg
,
'alpha'
),
'x'
,
'y'
,
'z'
))
(
usmm
,
(
theano
.
tensor
.
neg
,
'alpha'
),
'x'
,
'y'
,
'z'
))
register_specialize
(
local_usmm
,
name
=
"local_usmm"
)
register_specialize
(
local_usmm
,
name
=
"local_usmm"
)
...
@@ -48,7 +48,7 @@ def local_csm_properties_csm(node):
...
@@ -48,7 +48,7 @@ def local_csm_properties_csm(node):
return
ret_var
return
ret_var
return
False
return
False
register_specialize
(
local_csm_properties_csm
)
sparse
.
register_specialize
(
local_csm_properties_csm
)
# This is tested in tests/test_basic.py:test_remove0
# This is tested in tests/test_basic.py:test_remove0
...
@@ -57,7 +57,7 @@ def local_inplace_remove0(node):
...
@@ -57,7 +57,7 @@ def local_inplace_remove0(node):
"""
"""
Optimization to insert inplace versions of Remove0.
Optimization to insert inplace versions of Remove0.
"""
"""
if
isinstance
(
node
.
op
,
Remove0
)
and
not
node
.
op
.
inplace
:
if
isinstance
(
node
.
op
,
sparse
.
Remove0
)
and
not
node
.
op
.
inplace
:
new_op
=
node
.
op
.
__class__
(
inplace
=
True
)
new_op
=
node
.
op
.
__class__
(
inplace
=
True
)
new_node
=
new_op
(
*
node
.
inputs
)
new_node
=
new_op
(
*
node
.
inputs
)
return
[
new_node
]
return
[
new_node
]
...
@@ -70,9 +70,9 @@ theano.compile.optdb.register('local_inplace_remove0',
...
@@ -70,9 +70,9 @@ theano.compile.optdb.register('local_inplace_remove0',
# register a specialization to replace StructuredDot -> StructuredDotCSx
# register a specialization to replace StructuredDot -> StructuredDotCSx
# This is tested in tests/test_basic.py:792
# This is tested in tests/test_basic.py:792
@gof.local_optimizer
([
_structured_dot
])
@gof.local_optimizer
([
sparse
.
_structured_dot
])
def
local_structured_dot
(
node
):
def
local_structured_dot
(
node
):
if
node
.
op
==
_structured_dot
:
if
node
.
op
==
sparse
.
_structured_dot
:
a
,
b
=
node
.
inputs
a
,
b
=
node
.
inputs
if
a
.
type
.
format
==
'csc'
:
if
a
.
type
.
format
==
'csc'
:
a_val
,
a_ind
,
a_ptr
,
a_shape
=
csm_properties
(
a
)
a_val
,
a_ind
,
a_ptr
,
a_shape
=
csm_properties
(
a
)
...
@@ -95,10 +95,25 @@ def local_structured_dot(node):
...
@@ -95,10 +95,25 @@ def local_structured_dot(node):
@gof.local_optimizer
([
usmm_csc_dense
])
@gof.local_optimizer
([
usmm_csc_dense
])
def
local_usmm_csc_dense_inplace
(
node
):
def
local_usmm_csc_dense_inplace
(
node
):
if
node
.
op
==
usmm_csc_dense
:
if
node
.
op
==
usmm_csc_dense
:
return
[
usmm_csc_dense_inplace
(
*
node
.
inputs
)]
return
[
sparse
.
usmm_csc_dense_inplace
(
*
node
.
inputs
)]
register_specialize
(
local_usmm_csc_dense_inplace
,
'inplace'
)
register_specialize
(
local_usmm_csc_dense_inplace
,
'inplace'
)
@gof.local_optimizer
([
sparse
.
csm_properties
])
def
local_csm_properties_csm
(
node
):
"""if we find csm_properties(CSM(*args)), then we can replace that with the
*args directly"""
if
node
.
op
==
sparse
.
csm_properties
:
csm
,
=
node
.
inputs
if
csm
.
owner
and
(
csm
.
owner
.
op
==
sparse
.
CSC
or
csm
.
owner
.
op
==
sparse
.
CSR
):
# csm.owner.inputs could be broadcastable. In that case, we have
# to adjust the broadcasting flag here.
ret_var
=
[
theano
.
tensor
.
patternbroadcast
(
i
,
o
.
broadcastable
)
for
i
,
o
in
izip
(
csm
.
owner
.
inputs
,
node
.
outputs
)]
return
ret_var
# This is tested in tests/test_basic.py:UsmmTests
# This is tested in tests/test_basic.py:UsmmTests
@gof.local_optimizer
([
usmm
])
@gof.local_optimizer
([
usmm
])
def
local_usmm_csx
(
node
):
def
local_usmm_csx
(
node
):
...
@@ -124,4 +139,125 @@ def local_usmm_csx(node):
...
@@ -124,4 +139,125 @@ def local_usmm_csx(node):
return
[
usmm_csc_dense
(
alpha
,
x_val
,
x_ind
,
x_ptr
,
return
[
usmm_csc_dense
(
alpha
,
x_val
,
x_ind
,
x_ptr
,
x_nsparse
,
y
,
z
)]
x_nsparse
,
y
,
z
)]
return
False
return
False
register_specialize
(
local_usmm_csx
)
sparse
.
register_specialize
(
local_usmm_csx
)
# register a specialization to replace MulSD -> MulSDCSX
@gof.local_optimizer
([
sparse
.
mul_s_d
])
def
local_mul_s_d
(
node
):
if
node
.
op
==
sparse
.
mul_s_d
:
x
,
y
=
node
.
inputs
x_is_sparse_variable
=
_is_sparse_variable
(
x
)
if
x_is_sparse_variable
:
svar
=
x
dvar
=
y
else
:
svar
=
y
dvar
=
x
if
dvar
.
type
.
ndim
!=
2
:
return
False
if
svar
.
type
.
format
==
'csc'
:
CSx
=
sparse
.
CSC
mul_s_d_csx
=
sparse
.
mul_s_d_csc
elif
svar
.
type
.
format
==
'csr'
:
CSx
=
sparse
.
CSR
mul_s_d_csx
=
sparse
.
mul_s_d_csr
else
:
raise
NotImplemented
()
c_data
=
mul_s_d_csx
(
sparse
.
csm_data
(
svar
),
sparse
.
csm_indices
(
svar
),
sparse
.
csm_indptr
(
svar
),
dvar
)
return
[
CSx
(
c_data
,
sparse
.
csm_indices
(
svar
),
sparse
.
csm_indptr
(
svar
),
sparse
.
csm_shape
(
svar
))]
return
False
sparse
.
register_specialize
(
local_mul_s_d
)
@gof.local_optimizer
([
sparse
.
mul_s_v
])
def
local_mul_s_v
(
node
):
if
node
.
op
==
sparse
.
mul_s_v
:
x
,
y
=
node
.
inputs
x_is_sparse_variable
=
_is_sparse_variable
(
x
)
if
x_is_sparse_variable
:
svar
=
x
dvar
=
y
else
:
svar
=
y
dvar
=
x
if
dvar
.
type
.
ndim
!=
1
:
return
False
elif
svar
.
type
.
format
==
'csr'
:
CSx
=
sparse
.
CSR
mul_s_v_csx
=
sparse
.
mul_s_v_csr
else
:
return
False
s_val
,
s_ind
,
s_ptr
,
s_shape
=
sparse
.
csm_properties
(
svar
)
c_data
=
mul_s_v_csx
(
s_val
,
s_ind
,
s_ptr
,
dvar
)
return
[
CSx
(
c_data
,
s_ind
,
s_ptr
,
s_shape
)]
return
False
sparse
.
register_specialize
(
local_mul_s_v
)
@gof.local_optimizer
([
sparse
.
structured_add_s_v
])
def
local_structured_add_s_v
(
node
):
if
node
.
op
==
sparse
.
structured_add_s_v
:
x
,
y
=
node
.
inputs
x_is_sparse_variable
=
_is_sparse_variable
(
x
)
#y_is_sparse_variable = _is_sparse_variable(y)
if
x_is_sparse_variable
:
svar
=
x
dvar
=
y
else
:
svar
=
y
dvar
=
x
if
dvar
.
type
.
ndim
!=
1
:
return
False
elif
svar
.
type
.
format
==
'csr'
:
CSx
=
sparse
.
CSR
structured_add_s_v_csx
=
sparse
.
structured_add_s_v_csr
else
:
return
False
s_val
,
s_ind
,
s_ptr
,
s_shape
=
sparse
.
csm_properties
(
svar
)
c_data
=
structured_add_s_v_csx
(
s_val
,
s_ind
,
s_ptr
,
dvar
)
return
[
CSx
(
c_data
,
s_ind
,
s_ptr
,
s_shape
)]
return
False
sparse
.
register_specialize
(
local_structured_add_s_v
)
# register a specialization to replace SamplingDot -> SamplingDotCsr
@gof.local_optimizer
([
sparse
.
sampling_dot
])
def
local_sampling_dot_csr
(
node
):
if
node
.
op
==
sparse
.
sampling_dot
:
x
,
y
,
p
=
node
.
inputs
if
p
.
type
.
format
==
'csr'
:
p_data
,
p_ind
,
p_ptr
,
p_shape
=
sparse
.
csm_properties
(
p
)
z_data
,
z_ind
,
z_ptr
=
sparse
.
sampling_dot_csr
(
x
,
y
,
p_data
,
p_ind
,
p_ptr
,
p_shape
[
1
])
return
[
sparse
.
CSR
(
z_data
,
z_ind
,
z_ptr
,
p_shape
)]
return
False
sparse
.
register_specialize
(
local_sampling_dot_csr
,
name
=
'local_sampling_dot_csr'
)
theano/sparse/sandbox/sp2.py
浏览文件 @
075ece1e
...
@@ -12,8 +12,8 @@ from theano.sparse.basic import (
...
@@ -12,8 +12,8 @@ from theano.sparse.basic import (
_is_sparse_variable
,
_is_dense_variable
,
CSC
,
CSR
,
_is_sparse_variable
,
_is_dense_variable
,
CSC
,
CSR
,
csm_properties
,
csm_data
,
csm_indices
,
csm_indptr
,
csm_shape
,
csm_properties
,
csm_data
,
csm_indices
,
csm_indptr
,
csm_shape
,
_is_sparse
,
_is_sparse
,
Remove0
,
remove0
,
# To maintain compatibility
# To maintain compatibility
Remove0
,
remove0
,
Cast
,
bcast
,
wcast
,
icast
,
lcast
,
fcast
,
dcast
,
ccast
,
zcast
,
Cast
,
bcast
,
wcast
,
icast
,
lcast
,
fcast
,
dcast
,
ccast
,
zcast
,
HStack
,
hstack
,
VStack
,
vstack
,
HStack
,
hstack
,
VStack
,
vstack
,
AddSSData
,
add_s_s_data
,
AddSSData
,
add_s_s_data
,
...
@@ -27,48 +27,16 @@ from theano.sparse.basic import (
...
@@ -27,48 +27,16 @@ from theano.sparse.basic import (
StructuredAddSVCSR
,
structured_add_s_v_csr
,
StructuredAddSVCSR
,
structured_add_s_v_csr
,
SamplingDot
,
sampling_dot
,
SamplingDotCSR
,
sampling_dot_csr
)
SamplingDot
,
sampling_dot
,
SamplingDotCSR
,
sampling_dot_csr
)
# Also for compatibility
from
theano.sparse.opt
import
(
local_mul_s_d
,
local_mul_s_v
,
local_structured_add_s_v
,
local_sampling_dot_csr
)
# Alias to maintain compatibility
# Alias to maintain compatibility
EliminateZeros
=
Remove0
EliminateZeros
=
Remove0
eliminate_zeros
=
remove0
eliminate_zeros
=
remove0
# register a specialization to replace MulSD -> MulSDCSX
@gof.local_optimizer
([
mul_s_d
])
def
local_mul_s_d
(
node
):
if
node
.
op
==
mul_s_d
:
x
,
y
=
node
.
inputs
x_is_sparse_variable
=
_is_sparse_variable
(
x
)
# y_is_sparse_variable = _is_sparse_variable(y)
if
x_is_sparse_variable
:
svar
=
x
dvar
=
y
else
:
svar
=
y
dvar
=
x
if
dvar
.
type
.
ndim
!=
2
:
return
False
if
svar
.
type
.
format
==
'csc'
:
CSx
=
CSC
mul_s_d_csx
=
mul_s_d_csc
elif
svar
.
type
.
format
==
'csr'
:
CSx
=
CSR
mul_s_d_csx
=
mul_s_d_csr
else
:
raise
NotImplemented
()
c_data
=
mul_s_d_csx
(
csm_data
(
svar
),
csm_indices
(
svar
),
csm_indptr
(
svar
),
dvar
)
return
[
CSx
(
c_data
,
csm_indices
(
svar
),
csm_indptr
(
svar
),
csm_shape
(
svar
))]
return
False
register_specialize
(
local_mul_s_d
)
class
Binomial
(
gof
.
op
.
Op
):
class
Binomial
(
gof
.
op
.
Op
):
# TODO This op is not an equivalent of numpy.random.binomial. In
# TODO This op is not an equivalent of numpy.random.binomial. In
# facts, this does not follow a binomial distribution at all.
# facts, this does not follow a binomial distribution at all.
...
@@ -124,88 +92,8 @@ class Binomial(gof.op.Op):
...
@@ -124,88 +92,8 @@ class Binomial(gof.op.Op):
def
__str__
(
self
):
def
__str__
(
self
):
return
self
.
__class__
.
__name__
return
self
.
__class__
.
__name__
csr_fbinomial
=
Binomial
(
'csr'
,
'float32'
)
csr_fbinomial
=
Binomial
(
'csr'
,
'float32'
)
csc_fbinomial
=
Binomial
(
'csc'
,
'float32'
)
csc_fbinomial
=
Binomial
(
'csc'
,
'float32'
)
csr_dbinomial
=
Binomial
(
'csr'
,
'float64'
)
csr_dbinomial
=
Binomial
(
'csr'
,
'float64'
)
csc_dbinomial
=
Binomial
(
'csc'
,
'float64'
)
csc_dbinomial
=
Binomial
(
'csc'
,
'float64'
)
@gof.local_optimizer
([
mul_s_v
])
def
local_mul_s_v
(
node
):
if
node
.
op
==
mul_s_v
:
x
,
y
=
node
.
inputs
x_is_sparse_variable
=
_is_sparse_variable
(
x
)
if
x_is_sparse_variable
:
svar
=
x
dvar
=
y
else
:
svar
=
y
dvar
=
x
if
dvar
.
type
.
ndim
!=
1
:
return
False
elif
svar
.
type
.
format
==
'csr'
:
CSx
=
CSR
mul_s_v_csx
=
mul_s_v_csr
else
:
return
False
s_val
,
s_ind
,
s_ptr
,
s_shape
=
csm_properties
(
svar
)
c_data
=
mul_s_v_csx
(
s_val
,
s_ind
,
s_ptr
,
dvar
)
return
[
CSx
(
c_data
,
s_ind
,
s_ptr
,
s_shape
)]
return
False
register_specialize
(
local_mul_s_v
)
@gof.local_optimizer
([
structured_add_s_v
])
def
local_structured_add_s_v
(
node
):
if
node
.
op
==
structured_add_s_v
:
x
,
y
=
node
.
inputs
x_is_sparse_variable
=
_is_sparse_variable
(
x
)
#y_is_sparse_variable = _is_sparse_variable(y)
if
x_is_sparse_variable
:
svar
=
x
dvar
=
y
else
:
svar
=
y
dvar
=
x
if
dvar
.
type
.
ndim
!=
1
:
return
False
elif
svar
.
type
.
format
==
'csr'
:
CSx
=
CSR
structured_add_s_v_csx
=
structured_add_s_v_csr
else
:
return
False
s_val
,
s_ind
,
s_ptr
,
s_shape
=
csm_properties
(
svar
)
c_data
=
structured_add_s_v_csx
(
s_val
,
s_ind
,
s_ptr
,
dvar
)
return
[
CSx
(
c_data
,
s_ind
,
s_ptr
,
s_shape
)]
return
False
register_specialize
(
local_structured_add_s_v
)
# register a specialization to replace SamplingDot -> SamplingDotCsr
@gof.local_optimizer
([
sampling_dot
])
def
local_sampling_dot_csr
(
node
):
if
node
.
op
==
sampling_dot
:
x
,
y
,
p
=
node
.
inputs
if
p
.
type
.
format
==
'csr'
:
p_data
,
p_ind
,
p_ptr
,
p_shape
=
csm_properties
(
p
)
z_data
,
z_ind
,
z_ptr
=
sampling_dot_csr
(
x
,
y
,
p_data
,
p_ind
,
p_ptr
,
p_shape
[
1
])
return
[
CSR
(
z_data
,
z_ind
,
z_ptr
,
p_shape
)]
return
False
register_specialize
(
local_sampling_dot_csr
,
name
=
'local_sampling_dot_csr'
)
theano/sparse/tests/test_basic.py
浏览文件 @
075ece1e
...
@@ -941,8 +941,8 @@ class test_structureddot(unittest.TestCase):
...
@@ -941,8 +941,8 @@ class test_structureddot(unittest.TestCase):
theano_time
=
t1
-
t0
theano_time
=
t1
-
t0
scipy_time
=
t2
-
t1
scipy_time
=
t2
-
t1
#print 'theano took', theano_time,
#
print 'theano took', theano_time,
#print 'scipy took', scipy_time
#
print 'scipy took', scipy_time
overhead_tol
=
0.002
# seconds
overhead_tol
=
0.002
# seconds
overhead_rtol
=
1.1
# times as long
overhead_rtol
=
1.1
# times as long
self
.
assertTrue
(
numpy
.
allclose
(
theano_result
,
scipy_result
))
self
.
assertTrue
(
numpy
.
allclose
(
theano_result
,
scipy_result
))
...
...
theano/sparse/tests/test_opt.py
浏览文件 @
075ece1e
...
@@ -6,14 +6,12 @@ except ImportError:
...
@@ -6,14 +6,12 @@ except ImportError:
pass
# The variable enable_sparse will be used to disable the test file.
pass
# The variable enable_sparse will be used to disable the test file.
import
theano
import
theano
from
theano
import
config
,
tensor
from
theano
import
sparse
,
config
,
tensor
from
theano.sparse
import
enable_sparse
from
theano.sparse
import
enable_sparse
from
theano.gof.python25
import
any
from
theano.gof.python25
import
any
if
not
enable_sparse
:
if
not
enable_sparse
:
raise
SkipTest
(
'Optional package sparse disabled'
)
raise
SkipTest
(
'Optional package sparse disabled'
)
from
theano.sparse
import
(
CSM
,
CSMProperties
,
csm_properties
,
CSC
,
CSR
,
DenseFromSparse
,
CSMGrad
)
from
theano.sparse.tests.test_basic
import
random_lil
from
theano.sparse.tests.test_basic
import
random_lil
...
@@ -23,13 +21,15 @@ def test_local_csm_properties_csm():
...
@@ -23,13 +21,15 @@ def test_local_csm_properties_csm():
tensor
.
ivector
())
tensor
.
ivector
())
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_csm_properties_csm"
)
mode
=
mode
.
including
(
"specialize"
,
"local_csm_properties_csm"
)
for
CS
,
cast
in
[(
CSC
,
sp
.
csc_matrix
),
(
CSR
,
sp
.
csr_matrix
)]:
for
CS
,
cast
in
[(
sparse
.
CSC
,
sp
.
csc_matrix
),
(
sparse
.
CSR
,
sp
.
csr_matrix
)]:
f
=
theano
.
function
([
data
,
indices
,
indptr
,
shape
],
f
=
theano
.
function
([
data
,
indices
,
indptr
,
shape
],
csm_properties
(
CS
(
data
,
indices
,
indptr
,
shape
)),
sparse
.
csm_properties
(
CS
(
data
,
indices
,
indptr
,
shape
)),
mode
=
mode
)
mode
=
mode
)
#theano.printing.debugprint(f)
assert
not
any
(
assert
not
any
(
isinstance
(
node
.
op
,
(
CSM
,
CSMProperties
))
for
node
isinstance
(
node
.
op
,
(
sparse
.
CSM
,
sparse
.
CSMProperties
))
in
f
.
maker
.
env
.
toposort
())
for
node
in
f
.
maker
.
env
.
toposort
())
v
=
cast
(
random_lil
((
10
,
40
),
v
=
cast
(
random_lil
((
10
,
40
),
config
.
floatX
,
3
))
config
.
floatX
,
3
))
f
(
v
.
data
,
v
.
indices
,
v
.
indptr
,
v
.
shape
)
f
(
v
.
data
,
v
.
indices
,
v
.
indptr
,
v
.
shape
)
...
@@ -45,14 +45,79 @@ def test_local_csm_grad_c():
...
@@ -45,14 +45,79 @@ def test_local_csm_grad_c():
mode
=
theano
.
compile
.
Mode
(
linker
=
'c|py'
,
optimizer
=
'fast_compile'
)
mode
=
theano
.
compile
.
Mode
(
linker
=
'c|py'
,
optimizer
=
'fast_compile'
)
mode
=
mode
.
including
(
"specialize"
,
"local_csm_grad_c"
)
mode
=
mode
.
including
(
"specialize"
,
"local_csm_grad_c"
)
for
CS
,
cast
in
[(
CSC
,
sp
.
csc_matrix
),
(
CSR
,
sp
.
csr_matrix
)]:
for
CS
,
cast
in
[(
sparse
.
CSC
,
sp
.
csc_matrix
),
(
sparse
.
CSR
,
sp
.
csr_matrix
)]:
cost
=
tensor
.
sum
(
DenseFromSparse
()(
CS
(
data
,
indices
,
indptr
,
shape
)))
cost
=
tensor
.
sum
(
sparse
.
DenseFromSparse
()(
CS
(
data
,
indices
,
indptr
,
shape
)))
f
=
theano
.
function
(
f
=
theano
.
function
(
[
data
,
indices
,
indptr
,
shape
],
[
data
,
indices
,
indptr
,
shape
],
tensor
.
grad
(
cost
,
data
),
tensor
.
grad
(
cost
,
data
),
mode
=
mode
)
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
CSMGrad
)
for
node
assert
not
any
(
isinstance
(
node
.
op
,
sparse
.
CSMGrad
)
for
node
in
f
.
maker
.
env
.
toposort
())
in
f
.
maker
.
env
.
toposort
())
v
=
cast
(
random_lil
((
10
,
40
),
v
=
cast
(
random_lil
((
10
,
40
),
config
.
floatX
,
3
))
config
.
floatX
,
3
))
f
(
v
.
data
,
v
.
indices
,
v
.
indptr
,
v
.
shape
)
f
(
v
.
data
,
v
.
indices
,
v
.
indptr
,
v
.
shape
)
def
test_local_mul_s_d
():
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_mul_s_d"
)
for
sp_format
in
sparse
.
sparse_formats
:
inputs
=
[
getattr
(
theano
.
sparse
,
sp_format
+
'_matrix'
)(),
tensor
.
matrix
()]
f
=
theano
.
function
(
inputs
,
sparse
.
mul_s_d
(
*
inputs
),
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
sparse
.
MulSD
)
for
node
in
f
.
maker
.
env
.
toposort
())
def
test_local_mul_s_v
():
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_mul_s_v"
)
for
sp_format
in
[
'csr'
]:
# Not implemented for other format
inputs
=
[
getattr
(
theano
.
sparse
,
sp_format
+
'_matrix'
)(),
tensor
.
vector
()]
f
=
theano
.
function
(
inputs
,
sparse
.
mul_s_v
(
*
inputs
),
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
sparse
.
MulSV
)
for
node
in
f
.
maker
.
env
.
toposort
())
def
test_local_structured_add_s_v
():
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_structured_add_s_v"
)
for
sp_format
in
[
'csr'
]:
# Not implemented for other format
inputs
=
[
getattr
(
theano
.
sparse
,
sp_format
+
'_matrix'
)(),
tensor
.
vector
()]
f
=
theano
.
function
(
inputs
,
sparse
.
structured_add_s_v
(
*
inputs
),
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
sparse
.
StructuredAddSV
)
for
node
in
f
.
maker
.
env
.
toposort
())
def
test_local_sampling_dot_csr
():
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_sampling_dot_csr"
)
for
sp_format
in
[
'csr'
]:
# Not implemented for other format
inputs
=
[
tensor
.
matrix
(),
tensor
.
matrix
(),
getattr
(
theano
.
sparse
,
sp_format
+
'_matrix'
)()]
f
=
theano
.
function
(
inputs
,
sparse
.
sampling_dot
(
*
inputs
),
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
sparse
.
SamplingDot
)
for
node
in
f
.
maker
.
env
.
toposort
())
theano/sparse/tests/test_sp2.py
浏览文件 @
075ece1e
...
@@ -21,40 +21,6 @@ from theano.sparse.sandbox import sp2 as S2
...
@@ -21,40 +21,6 @@ from theano.sparse.sandbox import sp2 as S2
from
theano.tests
import
unittest_tools
as
utt
from
theano.tests
import
unittest_tools
as
utt
from
theano.sparse.basic
import
verify_grad_sparse
from
theano.sparse.basic
import
verify_grad_sparse
# Already in test_basic.py
# Not used here
# def as_sparse_format(data, format):
# if format == 'csc':
# return scipy.sparse.csc_matrix(data)
# elif format == 'csr':
# return scipy.sparse.csr_matrix(data)
# else:
# raise NotImplementedError()
# Already in test_basic.py
# Not used here
# def eval_outputs(outputs):
# return compile.function([], outputs)()[0]
# Already in test_basic.py
# Not used here
# def random_lil(shape, dtype, nnz):
# rval = sp.lil_matrix(shape, dtype=dtype)
# huge = 2 ** 30
# for k in range(nnz):
# # set non-zeros in random locations (row x, col y)
# idx = np.random.random_integers(huge, size=len(shape)) % shape
# value = np.random.rand()
# #if dtype *int*, value will always be zeros!
# if "int" in dtype:
# value = int(value * 100)
# rval.__setitem__(
# idx,
# value)
# return rval
class
BinomialTester
(
utt
.
InferShapeTester
):
class
BinomialTester
(
utt
.
InferShapeTester
):
n
=
tensor
.
scalar
()
n
=
tensor
.
scalar
()
...
@@ -96,72 +62,5 @@ class BinomialTester(utt.InferShapeTester):
...
@@ -96,72 +62,5 @@ class BinomialTester(utt.InferShapeTester):
self
.
op_class
)
self
.
op_class
)
# ##################
# Optimization tests
# ##################
def
test_local_mul_s_d
():
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_mul_s_d"
)
for
sp_format
in
sparse
.
sparse_formats
:
inputs
=
[
getattr
(
theano
.
sparse
,
sp_format
+
'_matrix'
)(),
tensor
.
matrix
()]
f
=
theano
.
function
(
inputs
,
sparse
.
mul_s_d
(
*
inputs
),
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
sparse
.
MulSD
)
for
node
in
f
.
maker
.
env
.
toposort
())
def
test_local_mul_s_v
():
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_mul_s_v"
)
for
sp_format
in
[
'csr'
]:
# Not implemented for other format
inputs
=
[
getattr
(
theano
.
sparse
,
sp_format
+
'_matrix'
)(),
tensor
.
vector
()]
f
=
theano
.
function
(
inputs
,
S2
.
mul_s_v
(
*
inputs
),
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
S2
.
MulSV
)
for
node
in
f
.
maker
.
env
.
toposort
())
def
test_local_structured_add_s_v
():
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_structured_add_s_v"
)
for
sp_format
in
[
'csr'
]:
# Not implemented for other format
inputs
=
[
getattr
(
theano
.
sparse
,
sp_format
+
'_matrix'
)(),
tensor
.
vector
()]
f
=
theano
.
function
(
inputs
,
S2
.
structured_add_s_v
(
*
inputs
),
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
S2
.
StructuredAddSV
)
for
node
in
f
.
maker
.
env
.
toposort
())
def
test_local_sampling_dot_csr
():
mode
=
theano
.
compile
.
mode
.
get_default_mode
()
mode
=
mode
.
including
(
"specialize"
,
"local_sampling_dot_csr"
)
for
sp_format
in
[
'csr'
]:
# Not implemented for other format
inputs
=
[
tensor
.
matrix
(),
tensor
.
matrix
(),
getattr
(
theano
.
sparse
,
sp_format
+
'_matrix'
)()]
f
=
theano
.
function
(
inputs
,
S2
.
sampling_dot
(
*
inputs
),
mode
=
mode
)
assert
not
any
(
isinstance
(
node
.
op
,
S2
.
SamplingDot
)
for
node
in
f
.
maker
.
env
.
toposort
())
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
unittest
.
main
()
unittest
.
main
()
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论