Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
87103eff
提交
87103eff
authored
4月 07, 2008
作者:
olivier@olivier-desktop
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
broadcast, reduce are working in python and c
上级
88f7858f
隐藏空白字符变更
内嵌
并排
正在显示
14 个修改的文件
包含
560 行增加
和
333 行删除
+560
-333
_test_scalar_ops.py
_test_scalar_ops.py
+5
-13
_test_tensor.py
_test_tensor.py
+1
-1
base_tensor.py
base_tensor.py
+27
-27
compile.py
compile.py
+2
-2
elemwise.py
elemwise.py
+28
-8
elemwise2.py
elemwise2.py
+289
-110
_test_cc.py
gof/_test_cc.py
+24
-24
cc.py
gof/cc.py
+57
-43
env.py
gof/env.py
+4
-4
op.py
gof/op.py
+12
-12
result.py
gof/result.py
+12
-12
scalar.py
scalar.py
+44
-46
scalar_ops.py
scalar_ops.py
+44
-24
tensor.py
tensor.py
+11
-7
没有找到文件。
sandbox/
_test_scalar_ops.py
→
_test_scalar_ops.py
浏览文件 @
87103eff
...
...
@@ -8,31 +8,23 @@ from scalar_ops import *
def
inputs
():
x
=
modes
.
build
_eval
(
as_scalar
(
1.0
,
'x'
))
y
=
modes
.
build
_eval
(
as_scalar
(
2.0
,
'y'
))
z
=
modes
.
build
_eval
(
as_scalar
(
3.0
,
'z'
))
x
=
modes
.
build
(
as_scalar
(
1.0
,
'x'
))
y
=
modes
.
build
(
as_scalar
(
2.0
,
'y'
))
z
=
modes
.
build
(
as_scalar
(
3.0
,
'z'
))
return
x
,
y
,
z
def
env
(
inputs
,
outputs
,
validate
=
True
,
features
=
[]):
# inputs = [input.r for input in inputs]
# outputs = [output.r for output in outputs]
return
Env
(
inputs
,
outputs
,
features
=
features
,
consistency_check
=
validate
)
class
_test_ScalarOps
(
unittest
.
TestCase
):
def
test_0
(
self
):
x
,
y
,
z
=
inputs
()
e
=
mul
(
add
(
x
,
y
),
div
(
x
,
y
))
assert
e
.
data
==
1.5
def
test_1
(
self
):
def
test_straightforward
(
self
):
x
,
y
,
z
=
inputs
()
e
=
mul
(
add
(
x
,
y
),
div
(
x
,
y
))
g
=
env
([
x
,
y
],
[
e
])
fn
=
gof
.
cc
.
C
Linker
(
g
)
.
make_function
()
fn
=
gof
.
Dual
Linker
(
g
)
.
make_function
()
assert
fn
(
1.0
,
2.0
)
==
1.5
assert
e
.
data
==
1.5
if
__name__
==
'__main__'
:
...
...
_test_tensor.py
浏览文件 @
87103eff
...
...
@@ -691,7 +691,7 @@ class t_gemm(unittest.TestCase):
self
.
rand
(
3
,
5
),
self
.
rand
(
5
,
4
),
1.0
)
def
test12
(
self
):
self
.
cmp
(
self
.
rand
(
3
,
4
),
-
1.0
,
self
.
rand
(
3
,
5
),
self
.
rand
(
5
,
4
),
-
1.0
)
t_gemm
=
None
if
__name__
==
'__main__'
:
unittest
.
main
()
base_tensor.py
浏览文件 @
87103eff
...
...
@@ -108,53 +108,53 @@ class BaseTensor(ResultBase):
#
# C codegen stubs
#
def
c_declare
(
self
):
def
c_declare
(
self
,
name
,
sub
):
return
"""
PyArrayObject*
%
%
(name)
s;
int type_num_
%
%
(name)
s;
typedef
%(dtype)
s dtype_
%
%
(name)
s;
"""
%
dict
(
dtype
=
self
.
dtype_specs
()[
1
])
PyArrayObject*
%(name)
s;
int type_num_
%(name)
s;
typedef
%(dtype)
s dtype_
%(name)
s;
"""
%
dict
(
sub
,
name
=
name
,
dtype
=
self
.
dtype_specs
()[
1
])
def
c_init
(
self
):
def
c_init
(
self
,
name
,
sub
):
return
"""
%
%
(name)
s = NULL;
type_num_
%
%
(name)
s =
%(type_num)
s;
"""
%
dict
(
type_num
=
self
.
dtype_specs
()[
2
])
%(name)
s = NULL;
type_num_
%(name)
s =
%(type_num)
s;
"""
%
dict
(
sub
,
name
=
name
,
type_num
=
self
.
dtype_specs
()[
2
])
def
c_extract
(
self
):
def
c_extract
(
self
,
name
,
sub
):
return
"""
%
%
(name)
s = NULL;
type_num_
%
%
(name)
s =
%(type_num)
s;
if (py_
%
%
(name)
s == Py_None) {
// We can either fail here or set
%
%
(name)
s to NULL and rely on Ops using
%(name)
s = NULL;
type_num_
%(name)
s =
%(type_num)
s;
if (py_
%(name)
s == Py_None) {
// We can either fail here or set
%(name)
s to NULL and rely on Ops using
// tensors to handle the NULL case, but if they fail to do so they'll end up
// with nasty segfaults, so this is public service.
PyErr_SetString(PyExc_ValueError, "expected an ndarray, not None");
%
%
(fail)
s
//
%
%
(name)
s = NULL;
%(fail)
s
//
%(name)
s = NULL;
}
else if (!PyArray_Check(py_
%
%
(name)
s)) {
else if (!PyArray_Check(py_
%(name)
s)) {
PyErr_SetString(PyExc_ValueError, "expected an ndarray");
%
%
(fail)
s
%(fail)
s
}
else if (((PyArrayObject*)py_
%
%
(name)
s)->descr->type_num !=
%(type_num)
s) {
else if (((PyArrayObject*)py_
%(name)
s)->descr->type_num !=
%(type_num)
s) {
PyErr_SetString(PyExc_ValueError, "expected
%(type_num)
s");
%
%
(fail)
s
%(fail)
s
}
else {
%
%(name)
s = (PyArrayObject*)(py_
%
%(name)
s);
Py_XINCREF(
%
%
(name)
s);
%
(name)
s = (PyArrayObject*)(py_
%(name)
s);
Py_XINCREF(
%(name)
s);
}
"""
%
dict
(
type_num
=
self
.
dtype_specs
()[
2
])
"""
%
dict
(
sub
,
name
=
name
,
type_num
=
self
.
dtype_specs
()[
2
])
def
c_cleanup
(
self
):
def
c_cleanup
(
self
,
name
,
sub
):
return
"""
if (
%(name)
s) {
Py_XDECREF(
%(name)
s);
}
"""
"""
%
locals
()
def
c_sync
(
self
):
def
c_sync
(
self
,
name
,
sub
):
return
"""
if (!
%(name)
s) {
Py_XDECREF(py_
%(name)
s);
...
...
@@ -165,7 +165,7 @@ class BaseTensor(ResultBase):
py_
%(name)
s = (PyObject*)
%(name)
s;
Py_XINCREF(py_
%(name)
s);
}
"""
"""
%
locals
()
def
c_headers
(
self
):
return
[]
...
...
compile.py
浏览文件 @
87103eff
...
...
@@ -16,8 +16,8 @@ exec_opt.optimizer = None
def
default_optimizer
(
env
):
#TODO: pass tests with these un-commented
default_optimizer
.
const
(
env
)
default_optimizer
.
merge
(
env
)
#
default_optimizer.const(env)
#
default_optimizer.merge(env)
pass
default_optimizer
.
merge
=
gof
.
opt
.
MergeOptimizer
()
default_optimizer
.
const
=
gof
.
opt
.
ConstantFinder
()
...
...
elemwise.py
浏览文件 @
87103eff
...
...
@@ -61,21 +61,41 @@ class Elemwise(Op):
return
ret
def
c_validate_update
(
self
):
def
c_validate_update
(
self
,
input_names
,
output_names
,
sub
):
sub
=
dict
(
sub
)
icvn
,
ocvn
=
self
.
c_var_names
()
for
real
,
tosub
in
zip
(
input_names
+
output_names
,
icvn
+
ocvn
):
sub
[
tosub
]
=
real
(
valupd
,
valupd_cleanup
),
(
code
,
code_cleanup
)
=
self
.
__c_code
()
return
valupd
return
valupd
%
sub
def
c_validate_update_cleanup
(
self
,
input_names
,
output_names
,
sub
):
sub
=
dict
(
sub
)
icvn
,
ocvn
=
self
.
c_var_names
()
for
real
,
tosub
in
zip
(
input_names
+
output_names
,
icvn
+
ocvn
):
sub
[
tosub
]
=
real
def
c_validate_update_cleanup
(
self
):
(
valupd
,
valupd_cleanup
),
(
code
,
code_cleanup
)
=
self
.
__c_code
()
return
valupd_cleanup
return
valupd_cleanup
%
sub
def
c_code
(
self
,
input_names
,
output_names
,
sub
):
sub
=
dict
(
sub
)
icvn
,
ocvn
=
self
.
c_var_names
()
for
real
,
tosub
in
zip
(
input_names
+
output_names
,
icvn
+
ocvn
):
sub
[
tosub
]
=
real
def
c_code
(
self
):
(
valupd
,
valupd_cleanup
),
(
code
,
code_cleanup
)
=
self
.
__c_code
()
return
code
return
code
%
sub
def
c_code_cleanup
(
self
,
input_names
,
output_names
,
sub
):
sub
=
dict
(
sub
)
icvn
,
ocvn
=
self
.
c_var_names
()
for
real
,
tosub
in
zip
(
input_names
+
output_names
,
icvn
+
ocvn
):
sub
[
tosub
]
=
real
def
c_code_cleanup
(
self
):
(
valupd
,
valupd_cleanup
),
(
code
,
code_cleanup
)
=
self
.
__c_code
()
return
code_cleanup
return
code_cleanup
%
sub
@classmethod
def
inplace_version
(
cls
,
dmap
=
{
0
:
0
}):
...
...
sandbox/
elemwise2.py
→
elemwise2.py
浏览文件 @
87103eff
import
elemwise_cgen
as
cgen
# foldl(f, fold_inputs, init) =>
# fold_inputs = init;
# for loop_inputs in c_order(difference(inputs, fold_inputs)):
# fold_inputs = f(fold_inputs, loop_inputs)
# a+b+c+d => ((a+b)+c)+d
# foldr(f, fold_inputs, init) =>
# fold_inputs = init;
# for loop_inputs in reversed_c_order(difference(inputs, fold_inputs)):
# fold_inputs = f(fold_inputs, loop_inputs)
# a**b**c**d => a**(b**(c**d))
# foldx(f, fold_inputs, init) =>
# fold_inputs = init;
# for loop_inputs in any_order(difference(inputs, fold_inputs)):
# fold_inputs = f(fold_inputs, loop_inputs)
# a+b+c+d => ((a+b)+c)+d
# a+b+c+d => a+(b+(c+d))
# a+b+c+d => (a+b)+(c+d)
# foldx <=> f.associative
# f.associative => (foldl => foldx) and (foldr => foldx)
# z = a*b + b*c + c*d + d*e
# z: (0, 0, 0, 0)
# a: (0, 0, 0, 0) => (0, 0, 1, 0, 0, 1) => loop order: 1, 2, 3, 4, x, x
# b: (0, 0) => (1, 1, 1, 0, 0, 1) => loop order: x, x, x, 1, 2, x
# c: (0, 0, S, 0, 0, S) => (0, 0, S, 0, 0, S) => loop order: 1, 2, 4, 5, 3, 6
# d: (1, 0, 1) => (1, 1, 1, 0, 1, 1) => loop order: x, x, x, 2, x, x
# e: (S, 0, 0, S, 0, 0) => => loop order: 2, 3, 5, 6, 1, 4
# strategy: (broadcasted, folded, fold_method)
# (2, 1, 1, 3, 1), (1, 7, 1, 1, 4)
# (2, 7, 1, 3, 4), (1, 1, 8, 1, 4)
# (2, 7, 8, 3, 4)
# (2, 3, 4, 5), (7, 3, 4, 8)
# (2, 3, 4), (3, 4)
# (2, 3, 4)
class
ElemwiseGroup
:
def
__init__
(
self
):
self
.
def
compile_env
(
env
):
mappings
=
{}
order
=
env
.
io_toposort
()
for
op
in
reversed
(
order
):
if
not
isinstance
(
op
,
Elemwise
):
raise
TypeError
(
"Unsupported op type for the Elemwise compiler."
,
op
)
for
input
in
op
.
input_policy
:
strategies
.
setdefault
()
def
elemwise_op_gen
(
op
,
modalities
):
"""
* op: z = x + y
modalities: {z: foldx(0, x, y)}
result: Z = sum(Y)
* op: z = x + y
"""
def
broadcasting_cgen
(
op
):
template
=
op
.
c_foreach
()
import
numpy
from
gof
import
Op
,
Viewer
,
Destroyer
from
tensor
import
Tensor
from
scalar
import
upcast
,
Scalar
import
scalar_ops
import
gof
##################
### DimShuffle ###
##################
class
DimShuffle
(
Op
,
Viewer
):
...
...
@@ -108,11 +37,14 @@ class DimShuffle(Op, Viewer):
self
.
inplace
=
inplace
self
.
numorder
=
[
x
for
x
in
new_order
if
type
(
x
)
==
int
]
self
.
is_transposition
=
sorted
(
new_order
)
==
range
(
len
gth
(
ib
))
self
.
is_transposition
=
sorted
(
new_order
)
==
range
(
len
(
ib
))
self
.
dup_dims
=
len
(
set
(
self
.
numorder
))
!=
len
(
self
.
numorder
)
self
.
all_dims
=
len
(
set
(
self
.
numorder
))
==
len
(
ib
)
if
self
.
dup_dims
or
not
self
.
all_dims
:
raise
NotImplementedError
(
"You must provide a permutation of *all* the input dimensions with *no duplicates*."
)
def
clone_with_new_inputs
(
self
,
*
new_inputs
):
return
DimShuffle
(
new_inputs
[
0
],
self
.
new_order
,
self
.
inplace
)
def
view_map
(
self
):
if
self
.
inplace
:
...
...
@@ -124,13 +56,13 @@ class DimShuffle(Op, Viewer):
res
=
self
.
inputs
[
0
]
.
data
.
transpose
(
self
.
numorder
)
shape
=
list
(
res
.
shape
)
new_shape
=
[]
for
entry
in
new_order
:
for
entry
in
self
.
new_order
:
if
entry
==
'x'
:
new_shape
.
append
(
1
)
else
:
new_shape
.
append
(
shape
.
pop
())
new_shape
.
append
(
shape
.
pop
(
0
))
res
=
res
.
reshape
(
new_shape
)
if
not
inplace
:
if
not
self
.
inplace
:
res
=
numpy
.
copy
(
res
)
self
.
outputs
[
0
]
.
data
=
res
...
...
@@ -141,20 +73,40 @@ class DimShuffle(Op, Viewer):
class
Transpose
(
DimShuffle
):
def
__init__
(
self
,
input
):
DimShuffle
.
__init__
(
self
,
input
,
range
(
len
(
input
.
broadcastable
)
-
1
,
-
1
,
-
1
))
DimShuffle
.
__init__
(
self
,
input
,
range
(
len
(
input
.
broadcastable
)
-
1
,
-
1
,
-
1
)
,
False
)
def
clone_with_new_inputs
(
self
,
*
new_inputs
):
return
Transpose
(
new_inputs
[
0
])
def
__str__
(
self
):
return
"
%
s(
%
s)"
%
(
self
.
__class__
.
__name__
,
str
(
self
.
inputs
[
0
]))
#################
### Broadcast ###
#################
class
Broadcast
(
Op
,
Destroyer
):
def
__init__
(
self
,
scalar_opclass
,
inputs
,
inplace_pattern
):
def
__init__
(
self
,
scalar_opclass
,
inputs
,
inplace_pattern
=
{}
):
try
:
assert
len
(
set
([
len
(
input
.
broadcastable
)
for
input
in
inputs
])
==
1
)
assert
len
(
set
([
len
(
input
.
broadcastable
)
for
input
in
inputs
])
)
==
1
except
(
AssertionError
,
AttributeError
):
raise
TypeError
(
"All inputs to a Broadcast subclass must be Tensor instances and their broadcastable fields must all have the same length."
,
self
.
__class__
)
self
.
nin
=
scalar_opclass
.
nin
self
.
nout
=
scalar_opclass
.
nout
out_broadcastables
=
[[
1
*
all
(
bcast
)
for
bcast
in
zip
(
*
[
input
.
broadcastable
for
input
in
inputs
])]]
*
self
.
nout
if
inplace_pattern
:
for
overwriter
,
overwritten
in
inplace_pattern
.
items
():
for
ob
,
ib
in
zip
(
out_broadcastables
[
overwriter
],
inputs
[
overwritten
]
.
broadcastable
):
if
ib
and
not
ob
:
raise
ValueError
(
"Operation cannot be done inplace on an input with broadcasted dimensions."
)
upcasted
=
upcast
(
*
[
input
.
dtype
for
input
in
inputs
])
def
get_dtype
(
i
):
input_idx
=
inplace_pattern
.
get
(
i
,
[
None
]
)
input_idx
=
inplace_pattern
.
get
(
i
,
None
)
if
input_idx
is
not
None
:
return
inputs
[
input_idx
]
.
dtype
else
:
...
...
@@ -164,8 +116,11 @@ class Broadcast(Op, Destroyer):
self
.
outputs
=
[
Tensor
(
dtype
=
dtype
,
broadcastable
=
broadcastable
)
for
dtype
,
broadcastable
in
zip
(
out_dtypes
,
out_broadcastables
)]
self
.
inplace_pattern
=
inplace_pattern
self
.
scalar_opclass
=
scalar_opclass
self
.
shadow
=
scalar_opclass
([
Scalar
(
dtype
=
t
.
dtype
)
for
t
in
self
.
inputs
])
self
.
ufunc
=
numpy
.
frompyfunc
(
scalar_opclass
.
impl
,
scalar_opclass
.
nin
,
scalar_opclass
.
nout
)
self
.
shadow
=
scalar_opclass
(
*
[
Scalar
(
dtype
=
t
.
dtype
)
for
t
in
self
.
inputs
])
self
.
ufunc
=
numpy
.
frompyfunc
(
self
.
shadow
.
impl
,
scalar_opclass
.
nin
,
scalar_opclass
.
nout
)
def
clone_with_new_inputs
(
self
,
*
new_inputs
):
return
Broadcast
(
self
.
scalar_opclass
,
new_inputs
,
self
.
inplace_pattern
)
def
id
(
self
):
return
(
self
.
__class__
,
self
.
scalar_opclass
,
self
.
inplace_pattern
)
...
...
@@ -194,7 +149,8 @@ class Broadcast(Op, Destroyer):
r
=
transform
(
scalar_igrad
)
to_sum
=
[
i
for
i
,
bcast
in
enumerate
(
input
.
broadcastable
)
if
bcast
]
if
to_sum
:
ret
.
append
(
Sum
(
r
,
to_sum
))
sr
=
Sum
(
r
,
axis
=
to_sum
)
.
out
ret
.
append
(
sr
)
else
:
ret
.
append
(
r
)
return
ret
...
...
@@ -204,24 +160,114 @@ class Broadcast(Op, Destroyer):
if
not
self
.
inplace_pattern
:
for
output
in
self
.
outputs
:
odat
=
output
.
data
shape
=
[
max
(
values
)
for
values
in
zip
(
*
[
input
.
data
.
shape
for
input
in
self
.
inputs
])]
if
odat
is
not
None
:
odat
.
resize
(
s
elf
.
inputs
[
0
]
.
data
.
s
hape
)
odat
.
resize
(
shape
)
else
:
odat
=
numpy
.
ndarray
(
s
elf
.
inputs
[
0
]
.
data
.
s
hape
,
dtype
=
output
.
dtype
)
odat
=
numpy
.
ndarray
(
shape
,
dtype
=
output
.
dtype
)
output_storage
.
append
(
odat
)
output
.
data
=
odat
else
:
for
i
,
output
in
enumerate
(
self
.
outputs
):
if
i
in
self
.
inplace_pattern
:
odat
=
self
.
inputs
[
self
.
inplace_pattern
[
i
]]
.
data
else
:
odat
=
output
.
data
shape
=
[
max
(
values
)
for
values
in
zip
(
*
[
input
.
data
.
shape
for
input
in
self
.
inputs
])]
if
odat
is
not
None
:
odat
.
resize
(
s
elf
.
inputs
[
0
]
.
data
.
s
hape
)
odat
.
resize
(
shape
)
else
:
odat
=
numpy
.
ndarray
(
s
elf
.
inputs
[
0
]
.
data
.
s
hape
,
dtype
=
output
.
dtype
)
odat
=
numpy
.
ndarray
(
shape
,
dtype
=
output
.
dtype
)
output_storage
.
append
(
odat
)
output
.
data
=
odat
self
.
ufunc
(
*
([
input
.
data
for
input
in
self
.
inputs
]
+
output_storage
))
def
_c_all
(
self
,
inames
,
onames
,
sub
):
defines
=
""
undefs
=
""
dmap
=
self
.
destroy_map
()
idtypes
=
[
input
.
dtype_specs
()[
1
]
for
input
in
self
.
inputs
]
real
=
zip
(
*
[(
r
,
s
,
r
.
dtype_specs
()[
1
])
for
r
,
s
in
zip
(
self
.
outputs
,
onames
)
if
r
not
in
dmap
])
if
real
:
real_outputs
,
real_onames
,
real_odtypes
=
real
else
:
real_outputs
,
real_onames
,
real_odtypes
=
[],
[],
[]
aliased
=
zip
(
*
[(
r
,
s
)
for
(
r
,
s
)
in
zip
(
self
.
outputs
,
onames
)
if
r
in
dmap
])
if
aliased
:
aliased_outputs
,
aliased_onames
=
aliased
else
:
aliased_outputs
,
aliased_onames
=
[],
[]
orders
=
[[
x
and
'x'
or
i
for
i
,
x
in
enumerate
(
input
.
broadcastable
)]
for
input
in
self
.
inputs
]
nnested
=
len
(
orders
[
0
])
sub
=
dict
(
sub
)
for
i
,
(
input
,
iname
)
in
enumerate
(
zip
(
self
.
inputs
,
inames
)):
sub
[
'lv
%
i'
%
i
]
=
iname
decl
=
cgen
.
make_declare
(
orders
,
idtypes
,
sub
)
checks
=
cgen
.
make_checks
(
orders
,
idtypes
,
sub
)
alloc
=
""
for
output
,
oname
,
odtype
in
zip
(
real_outputs
,
real_onames
,
real_odtypes
):
i
+=
1
sub
[
'lv
%
i'
%
i
]
=
oname
sub
[
'olv'
]
=
oname
alloc
+=
cgen
.
make_declare
([
range
(
nnested
)],
[
odtype
],
dict
(
sub
,
lv0
=
oname
))
alloc
+=
cgen
.
make_alloc
(
orders
,
odtype
,
sub
)
alloc
+=
cgen
.
make_checks
([
range
(
nnested
)],
[
odtype
],
dict
(
sub
,
lv0
=
oname
))
for
output
,
oname
in
zip
(
aliased_outputs
,
aliased_onames
):
iname
=
inames
[
self
.
inputs
.
index
(
dmap
[
output
][
0
])]
alloc
+=
"""
if (
%(oname)
s) {
Py_XDECREF(
%(oname)
s);
}
%(oname)
s =
%(iname)
s;
Py_XINCREF(
%(oname)
s);
"""
%
locals
()
defines
+=
"#define
%(oname)
s_i
%(iname)
s_i"
%
locals
()
undefs
+=
"#undef
%(oname)
s_i"
%
locals
()
task_code
=
self
.
shadow
.
c_code
([
"
%
s_i"
%
s
for
s
in
inames
],
[
"
%
s_i"
%
s
for
s
in
onames
],
sub
)
task_decl
=
""
.
join
([
"
%(dtype)
s&
%(name)
s_i = *
%(name)
s_iter;
\n
"
%
locals
()
for
name
,
dtype
in
zip
(
inames
+
list
(
real_onames
),
idtypes
+
list
(
real_odtypes
))])
code
=
"""
{
%(defines)
s
%(task_decl)
s
%(task_code)
s
%(undefs)
s
}
"""
%
locals
()
if
nnested
:
all_code
=
[(
""
,
""
)]
*
(
nnested
-
1
)
+
[(
""
,
code
)]
+
[
""
]
else
:
all_code
=
[
code
]
loop
=
cgen
.
make_loop
(
orders
+
[
range
(
nnested
)]
*
len
(
real_onames
),
idtypes
+
list
(
real_odtypes
),
all_code
,
sub
)
return
decl
,
checks
,
alloc
,
loop
def
c_code
(
self
,
inames
,
onames
,
sub
):
code
=
"
\n
"
.
join
(
self
.
_c_all
(
inames
,
onames
,
sub
))
return
code
def
make_broadcast
(
scalar_opclass
,
inplace_pattern
=
{},
name
=
None
):
class
New
(
Broadcast
):
def
__init__
(
self
,
*
inputs
):
Broadcast
.
__init__
(
self
,
scalar_opclass
,
inputs
,
inplace_pattern
)
def
clone_with_new_inputs
(
self
,
*
new_inputs
):
return
New
(
*
new_inputs
)
if
name
is
not
None
:
New
.
__name__
=
name
else
:
New
.
__name__
=
"Tensor"
+
scalar_opclass
.
__name__
return
New
def
broadcast
(
op
):
def
instantiate
(
*
inputs
):
...
...
@@ -234,8 +280,14 @@ def broadcast(op):
else
:
args
.
append
(
DimShuffle
(
input
,
[
'x'
]
*
difference
+
range
(
length
)))
return
op
(
*
args
)
return
instantiate
################
### CAReduce ###
################
class
CAReduce
(
Op
):
"""
CAReduce(scalar_op, inputs, dimensions_to_reduce = None, init = None, shortcut = False)
...
...
@@ -269,8 +321,148 @@ class CAReduce(Op):
def
__init__
(
self
,
scalar_opclass
,
inputs
,
dimensions_to_reduce
=
None
):
if
scalar_opclass
.
nin
!=
2
or
scalar_opclass
.
nout
!=
1
:
raise
NotImplementedError
(
"CAReduce only supports binary functions with a single output."
)
if
len
(
inputs
)
!=
1
:
raise
TypeError
(
"Only one argument expected."
)
if
dimensions_to_reduce
is
None
:
dimensions_to_reduce
=
range
(
len
(
inputs
[
0
]
.
broadcastable
))
self
.
nin
=
1
self
.
nout
=
1
self
.
inputs
=
inputs
self
.
outputs
=
[
Tensor
(
dtype
=
inputs
[
0
]
.
dtype
,
broadcastable
=
[
x
for
i
,
x
in
enumerate
(
inputs
[
0
]
.
broadcastable
)
if
i
not
in
dimensions_to_reduce
])]
self
.
dimensions_to_reduce
=
dimensions_to_reduce
self
.
scalar_opclass
=
scalar_opclass
self
.
shadow
=
scalar_opclass
(
*
[
Scalar
(
dtype
=
inputs
[
0
]
.
dtype
)
for
i
in
xrange
(
scalar_opclass
.
nin
)])
self
.
ufunc
=
numpy
.
frompyfunc
(
self
.
shadow
.
impl
,
scalar_opclass
.
nin
,
scalar_opclass
.
nout
)
def
id
(
self
):
return
(
self
.
__class__
,
self
.
scalar_opclass
,
self
.
dimensions_to_reduce
)
def
clone_with_new_inputs
(
self
,
*
new_inputs
):
return
CAReduce
(
self
.
scalar_opclass
,
new_inputs
,
self
.
dimensions_to_reduce
)
def
perform
(
self
):
result
=
self
.
inputs
[
0
]
.
data
for
dimension
in
reversed
(
sorted
(
self
.
dimensions_to_reduce
)):
result
=
self
.
ufunc
.
reduce
(
result
,
dimension
)
self
.
outputs
[
0
]
.
data
=
result
def
_c_all
(
self
,
inames
,
onames
,
sub
):
input
=
self
.
inputs
[
0
]
output
=
self
.
outputs
[
0
]
iname
=
inames
[
0
]
oname
=
onames
[
0
]
idtype
=
input
.
dtype_specs
()[
1
]
odtype
=
output
.
dtype_specs
()[
1
]
tosum
=
self
.
dimensions_to_reduce
order1
=
[
i
for
i
in
xrange
(
len
(
input
.
broadcastable
))
if
i
not
in
tosum
]
order
=
order1
+
list
(
tosum
)
nnested
=
len
(
order1
)
sub
=
dict
(
sub
)
for
i
,
(
input
,
iname
)
in
enumerate
(
zip
(
self
.
inputs
,
inames
)):
sub
[
'lv
%
i'
%
i
]
=
iname
decl
=
cgen
.
make_declare
([
order
],
[
idtype
],
sub
)
checks
=
cgen
.
make_checks
([
order
],
[
idtype
],
sub
)
alloc
=
""
i
+=
1
sub
[
'lv
%
i'
%
i
]
=
oname
sub
[
'olv'
]
=
oname
alloc
+=
cgen
.
make_declare
([
range
(
nnested
)
+
[
'x'
]
*
len
(
tosum
)],
[
odtype
],
dict
(
sub
,
lv0
=
oname
))
alloc
+=
cgen
.
make_alloc
([
order1
],
odtype
,
sub
)
alloc
+=
cgen
.
make_checks
([
range
(
nnested
)
+
[
'x'
]
*
len
(
tosum
)],
[
odtype
],
dict
(
sub
,
lv0
=
oname
))
task0_decl
=
"
%(dtype)
s&
%(name)
s_i = *
%(name)
s_iter;
\n
%(name)
s_i =
%(identity)
s;"
%
dict
(
dtype
=
odtype
,
name
=
onames
[
0
],
identity
=
self
.
shadow
.
identity
)
task1_decl
=
"
%(dtype)
s&
%(name)
s_i = *
%(name)
s_iter;
\n
"
%
dict
(
dtype
=
idtype
,
name
=
inames
[
0
])
task1_code
=
self
.
shadow
.
c_code
([
"
%
s_i"
%
onames
[
0
],
"
%
s_i"
%
inames
[
0
]],
[
"
%
s_i"
%
onames
[
0
]],
sub
)
code1
=
"""
{
%(task1_decl)
s
%(task1_code)
s
}
"""
%
locals
()
if
len
(
tosum
)
==
1
:
all_code
=
[(
""
,
""
)]
*
nnested
+
[(
task0_decl
,
code1
),
""
]
else
:
all_code
=
[(
""
,
""
)]
*
nnested
+
[(
task0_decl
,
""
)]
+
[(
""
,
""
)]
*
(
len
(
tosum
)
-
2
)
+
[(
""
,
code1
),
""
]
# if nnested:
# all_code = [("", "")] * (nnested - 1) + [("", code)] + [""]
# else:
# all_code = [code]
def
reduce
(
op
,
dimensions_to_reduce
):
# print [order, range(nnested) + ['x'] * len(tosum)]
loop
=
cgen
.
make_loop
([
order
,
range
(
nnested
)
+
[
'x'
]
*
len
(
tosum
)],
[
idtype
,
odtype
],
all_code
,
sub
)
return
decl
,
checks
,
alloc
,
loop
def
c_code
(
self
,
inames
,
onames
,
sub
):
code
=
"
\n
"
.
join
(
self
.
_c_all
(
inames
,
onames
,
sub
))
# print code
return
code
def
__str__
(
self
):
input
=
self
.
inputs
[
0
]
if
len
(
input
.
broadcastable
)
==
len
(
self
.
dimensions_to_reduce
):
return
"
%
s:
%
s(
%
s)"
%
(
self
.
__class__
.
__name__
,
self
.
scalar_opclass
.
__name__
,
str
(
input
))
else
:
return
"
%
s:
%
s(
%
s, axis =
%
s)"
%
(
self
.
__class__
.
__name__
,
self
.
scalar_opclass
.
__name__
,
str
(
input
),
self
.
dimensions_to_reduce
)
def
make_reduce
(
scalar_opclass
,
name
=
None
):
if
getattr
(
scalar_opclass
,
'commutative'
,
True
)
\
and
getattr
(
scalar_opclass
,
'associative'
,
True
):
reducer
=
CAReduce
else
:
raise
NotImplementedError
(
"The scalar op class to reduce must be commutative and associative."
)
class
New
(
reducer
):
def
__init__
(
self
,
*
inputs
,
**
kwargs
):
reducer
.
__init__
(
self
,
scalar_opclass
,
inputs
,
kwargs
.
get
(
'axis'
,
None
))
def
clone_with_new_inputs
(
self
,
*
new_inputs
):
return
New
(
*
new_inputs
,
**
dict
(
axis
=
self
.
dimensions_to_reduce
))
def
__str__
(
self
):
input
=
self
.
inputs
[
0
]
if
len
(
input
.
broadcastable
)
==
len
(
self
.
dimensions_to_reduce
):
return
"
%
s(
%
s)"
%
(
self
.
__class__
.
__name__
,
str
(
input
))
else
:
return
"
%
s(
%
s, axis =
%
s)"
%
(
self
.
__class__
.
__name__
,
str
(
input
),
self
.
dimensions_to_reduce
)
if
name
is
not
None
:
New
.
__name__
=
name
else
:
New
.
__name__
=
"Reduce"
+
scalar_opclass
.
__name__
return
New
Sum
=
make_reduce
(
scalar_ops
.
Add
,
name
=
'Sum'
)
def
reduce
(
op
):
if
getattr
(
op
,
'commutative'
,
True
)
and
getattr
(
op
,
'associative'
,
True
):
reducer
=
CAReduce
else
:
...
...
@@ -281,16 +473,3 @@ def reduce(op, dimensions_to_reduce):
# class Elemwise(TensorOp):
# def propagate_dtype(self, idtypes):
# raise AbstractFunctionError
# def propagate_broadcastable(self, ibroadcastables):
# raise AbstractFunctionError
# def _calculate_elemwise_strategy(self, input_strategies):
# raise AbstractFunctionError
gof/_test_cc.py
浏览文件 @
87103eff
...
...
@@ -25,20 +25,20 @@ class Double(ResultBase):
# def c_is_simple(self): return True
def
c_declare
(
self
):
return
"double
%(name)
s; void*
%(name)
s_bad_thing;"
def
c_declare
(
self
,
name
,
sub
):
return
"double
%(name)
s; void*
%(name)
s_bad_thing;"
%
locals
()
def
c_init
(
self
):
def
c_init
(
self
,
name
,
sub
):
return
"""
%(name)
s = 0;
%(name)
s_bad_thing = malloc(100000);
//printf("Initializing
%(name)
s
\\
n");
"""
"""
%
locals
()
def
c_literal
(
self
):
return
str
(
self
.
data
)
def
c_extract
(
self
):
def
c_extract
(
self
,
name
,
sub
):
return
"""
if (!PyFloat_Check(py_
%(name)
s)) {
PyErr_SetString(PyExc_TypeError, "not a double!");
...
...
@@ -47,23 +47,23 @@ class Double(ResultBase):
%(name)
s = PyFloat_AsDouble(py_
%(name)
s);
%(name)
s_bad_thing = NULL;
//printf("Extracting
%(name)
s
\\
n");
"""
"""
%
dict
(
locals
(),
**
sub
)
def
c_sync
(
self
):
def
c_sync
(
self
,
name
,
sub
):
return
"""
Py_XDECREF(py_
%(name)
s);
py_
%(name)
s = PyFloat_FromDouble(
%(name)
s);
if (!py_
%(name)
s)
py_
%(name)
s = Py_None;
//printf("Syncing
%(name)
s
\\
n");
"""
"""
%
locals
()
def
c_cleanup
(
self
):
def
c_cleanup
(
self
,
name
,
sub
):
return
"""
//printf("Cleaning up
%(name)
s
\\
n");
if (
%(name)
s_bad_thing)
free(
%(name)
s_bad_thing);
"""
"""
%
locals
()
class
MyOp
(
Op
):
...
...
@@ -80,43 +80,43 @@ class MyOp(Op):
class
Unary
(
MyOp
):
nin
=
1
def
c_var_names
(
self
):
return
[[
'x'
],
[
'z'
]]
#
def c_var_names(self):
#
return [['x'], ['z']]
class
Binary
(
MyOp
):
nin
=
2
def
c_var_names
(
self
):
return
[[
'x'
,
'y'
],
[
'z'
]]
#
def c_var_names(self):
#
return [['x', 'y'], ['z']]
class
Add
(
Binary
):
def
c_code
(
self
):
return
"
%(z)
s =
%(x)
s +
%(y)
s;"
def
c_code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s +
%(y)
s;"
%
locals
()
def
perform
(
self
):
self
.
outputs
[
0
]
.
data
=
self
.
inputs
[
0
]
.
data
+
self
.
inputs
[
1
]
.
data
class
Sub
(
Binary
):
def
c_code
(
self
):
return
"
%(z)
s =
%(x)
s -
%(y)
s;"
def
c_code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s -
%(y)
s;"
%
locals
()
def
perform
(
self
):
self
.
outputs
[
0
]
.
data
=
-
10
# erroneous
class
Mul
(
Binary
):
def
c_code
(
self
):
return
"
%(z)
s =
%(x)
s *
%(y)
s;"
def
c_code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s *
%(y)
s;"
%
locals
()
def
perform
(
self
):
self
.
outputs
[
0
]
.
data
=
self
.
inputs
[
0
]
.
data
*
self
.
inputs
[
1
]
.
data
class
Div
(
Binary
):
def
c_validate_update
(
self
):
def
c_validate_update
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"""
if (
%(y)
s == 0.0) {
PyErr_SetString(PyExc_ZeroDivisionError, "division by zero");
%(fail)
s
}
"""
def
c_code
(
self
):
return
"
%(z)
s =
%(x)
s /
%(y)
s;"
"""
%
dict
(
locals
(),
**
sub
)
def
c_code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s /
%(y)
s;"
%
locals
()
def
perform
(
self
):
self
.
outputs
[
0
]
.
data
=
self
.
inputs
[
0
]
.
data
/
self
.
inputs
[
1
]
.
data
...
...
gof/cc.py
浏览文件 @
87103eff
...
...
@@ -66,15 +66,19 @@ class CodeBlock:
to jump to. It should also contain a key called 'failure_var' that contains
the name of the variable that contains the error code.
"""
self
.
declare
=
declare
%
sub
behavior_sub
=
copy
(
sub
)
behavior_sub
[
'fail'
]
=
"{
%(failure_var)
s =
%(id)
s; goto __label_
%(id)
i;}"
%
sub
self
.
behavior
=
behavior
%
behavior_sub
self
.
declare
=
declare
#
% sub
#
behavior_sub = copy(sub)
#
behavior_sub['fail'] = "{%(failure_var)s = %(id)s; goto __label_%(id)i;}" % sub
self
.
behavior
=
behavior
#
% behavior_sub
# the dummy is because gcc throws an error when a label's right next to a closing
# brace (maybe there's an ignore flag for that...)
# we need the label even if cleanup is empty because the behavior block jumps there
# on failure
self
.
cleanup
=
(
"__label_
%(id)
i:
\n
"
+
cleanup
+
"
\n
double __DUMMY_
%(id)
i;
\n
"
)
%
sub
self
.
cleanup
=
(
"__label_
%(id)
i:
\n
"
%
sub
+
cleanup
+
"
\n
double __DUMMY_
%(id)
i;
\n
"
%
sub
)
#% sub
def
failure_code
(
sub
):
return
"{
%(failure_var)
s =
%(id)
s; goto __label_
%(id)
i;}"
%
sub
def
code_gen
(
blocks
):
...
...
@@ -192,14 +196,14 @@ def struct_gen(args, struct_builders, blocks, sub):
# TODO: add some error checking to make sure storage_<x> are 1-element lists
# and __ERROR is a 3-elements list.
struct_code
=
"""
struct
%
%
(name)
s {
struct
%(name)
s {
PyObject* __ERROR;
%(storage_decl)
s
%(struct_decl)
s
%
%
(name)
s() {}
~
%
%
(name)
s(void) {
%(name)
s() {}
~
%(name)
s(void) {
cleanup();
}
...
...
@@ -232,47 +236,47 @@ def struct_gen(args, struct_builders, blocks, sub):
# The get_<x> functions complete the return value of r.get_<x>()
# with handling of the py_<name> variable.
def
get_nothing
(
r
):
def
get_nothing
(
r
,
name
,
sub
):
""
return
""
def
get_c_declare
(
r
):
def
get_c_declare
(
r
,
name
,
sub
):
pre
=
"""
PyObject* py_
%(name)
s;
"""
return
pre
+
r
.
c_declare
()
"""
%
locals
()
return
pre
+
r
.
c_declare
(
name
,
sub
)
def
get_c_init
(
r
):
def
get_c_init
(
r
,
name
,
sub
):
pre
=
""
"""
py_
%(name)
s = Py_None;
"""
return
pre
+
r
.
c_init
()
"""
%
locals
()
return
pre
+
r
.
c_init
(
name
,
sub
)
def
get_c_extract
(
r
):
def
get_c_extract
(
r
,
name
,
sub
):
pre
=
"""
py_
%(name)
s = PyList_GET_ITEM(storage_
%(name)
s, 0);
Py_XINCREF(py_
%(name)
s);
"""
return
pre
+
r
.
c_extract
()
"""
%
locals
()
return
pre
+
r
.
c_extract
(
name
,
sub
)
def
get_c_cleanup
(
r
):
def
get_c_cleanup
(
r
,
name
,
sub
):
post
=
"""
Py_XDECREF(py_
%(name)
s);
"""
return
r
.
c_cleanup
()
+
post
"""
%
locals
()
return
r
.
c_cleanup
(
name
,
sub
)
+
post
def
get_c_sync
(
r
):
def
get_c_sync
(
r
,
name
,
sub
):
return
"""
if (!
%
%
(failure_var)
s) {
if (!
%(failure_var)
s) {
%(sync)
s
PyObject* old = PyList_GET_ITEM(storage_
%
%
(name)
s, 0);
Py_XINCREF(py_
%
%
(name)
s);
PyList_SET_ITEM(storage_
%
%(name)
s, 0, py_
%
%(name)
s);
PyObject* old = PyList_GET_ITEM(storage_
%(name)
s, 0);
Py_XINCREF(py_
%(name)
s);
PyList_SET_ITEM(storage_
%
(name)
s, 0, py_
%(name)
s);
Py_XDECREF(old);
}
"""
%
dict
(
sync
=
r
.
c_sync
(
)
)
"""
%
dict
(
sync
=
r
.
c_sync
(
name
,
sub
),
name
=
name
,
**
sub
)
def
apply_policy
(
policy
,
r
):
def
apply_policy
(
policy
,
r
,
name
,
sub
):
"""
policy -> list of functions that map a Result to a string,
or a single such function
...
...
@@ -282,8 +286,8 @@ def apply_policy(policy, r):
if
isinstance
(
r
,
(
list
,
tuple
)):
ret
=
""
for
sub_policy
in
policy
:
ret
+=
sub_policy
(
r
)
return
policy
(
r
)
ret
+=
sub_policy
(
r
,
name
,
sub
)
return
policy
(
r
,
name
,
sub
)
...
...
@@ -304,11 +308,15 @@ def struct_result_codeblocks(result, policies, id, symbol_table, sub):
name
=
"V
%
i"
%
id
symbol_table
[
result
]
=
name
sub
=
copy
(
sub
)
sub
[
'name'
]
=
name
#
sub['name'] = name
sub
[
'id'
]
=
id
struct_builder
=
CodeBlock
(
*
[
apply_policy
(
policy
,
result
)
for
policy
in
policies
[
0
]]
+
[
sub
])
# struct_declare, struct_behavior, struct_cleanup, sub)
sub
[
'fail'
]
=
failure_code
(
sub
)
struct_builder
=
CodeBlock
(
*
[
apply_policy
(
policy
,
result
,
name
,
sub
)
for
policy
in
policies
[
0
]]
+
[
sub
])
# struct_declare, struct_behavior, struct_cleanup, sub)
sub
[
'id'
]
=
id
+
1
block
=
CodeBlock
(
*
[
apply_policy
(
policy
,
result
)
for
policy
in
policies
[
1
]]
+
[
sub
])
# run_declare, run_behavior, run_cleanup, sub)
sub
[
'fail'
]
=
failure_code
(
sub
)
block
=
CodeBlock
(
*
[
apply_policy
(
policy
,
result
,
name
,
sub
)
for
policy
in
policies
[
1
]]
+
[
sub
])
# run_declare, run_behavior, run_cleanup, sub)
return
struct_builder
,
block
...
...
@@ -453,33 +461,39 @@ class CLinker(Linker):
# We populate sub with a mapping from the variable names specified by the op's c_var_names
# method to the actual variable names that we will use.
ivnames
,
ovnames
=
op
.
c_var_names
()
##
ivnames, ovnames = op.c_var_names()
sub
=
dict
(
failure_var
=
failure_var
)
for
result
,
vname
in
zip
(
op
.
inputs
+
op
.
outputs
,
ivnames
+
ovnames
):
sub
[
vname
]
=
symbol
[
result
]
## for result, vname in zip(op.inputs + op.outputs, ivnames + ovnames):
## sub[vname] = symbol[result]
isyms
,
osyms
=
[
symbol
[
r
]
for
r
in
op
.
inputs
],
[
symbol
[
r
]
for
r
in
op
.
outputs
]
# Make the CodeBlock for c_validate_update
try
:
validate_behavior
=
op
.
c_validate_update
()
sub
[
'id'
]
=
id
sub
[
'fail'
]
=
failure_code
(
sub
)
try
:
validate_behavior
=
op
.
c_validate_update
(
isyms
,
osyms
,
sub
)
except
AbstractFunctionError
:
validate_behavior
=
""
try
:
validate_cleanup
=
op
.
c_validate_update_cleanup
()
try
:
validate_cleanup
=
op
.
c_validate_update_cleanup
(
isyms
,
osyms
,
sub
)
except
AbstractFunctionError
:
validate_cleanup
=
""
sub
[
'id'
]
=
id
blocks
.
append
(
CodeBlock
(
""
,
validate_behavior
,
validate_cleanup
,
sub
))
tasks
.
append
((
op
,
'validate_update'
,
id
))
id
+=
1
# Make the CodeBlock for c_code
behavior
=
op
.
c_code
()
# this one must be implemented!
sub
[
'id'
]
=
id
sub
[
'fail'
]
=
failure_code
(
sub
)
behavior
=
op
.
c_code
(
isyms
,
osyms
,
sub
)
# this one must be implemented!
try
:
cleanup
=
op
.
c_code_cleanup
()
try
:
cleanup
=
op
.
c_code_cleanup
(
isyms
,
osyms
,
sub
)
except
AbstractFunctionError
:
cleanup
=
""
sub
[
'id'
]
=
id
blocks
.
append
(
CodeBlock
(
""
,
behavior
,
cleanup
,
sub
))
tasks
.
append
((
op
,
'code'
,
id
))
id
+=
1
...
...
@@ -489,7 +503,7 @@ class CLinker(Linker):
args
=
[]
args
+=
[
"storage_
%
s"
%
symbol
[
result
]
for
result
in
utils
.
uniq
(
self
.
inputs
+
self
.
outputs
+
self
.
orphans
)]
struct_code
=
struct_gen
(
args
,
init_blocks
,
blocks
,
dict
(
failure_var
=
failure_var
))
struct_code
=
struct_gen
(
args
,
init_blocks
,
blocks
,
dict
(
failure_var
=
failure_var
,
name
=
"
%(name)
s"
))
# The hash calculated on the code identifies it so weave can cache properly.
# (the hash has to be used outside of the support code because weave does not consider changes in the support code)
...
...
gof/env.py
浏览文件 @
87103eff
...
...
@@ -78,16 +78,13 @@ class Env(graph.Graph):
# The inputs and outputs set bound the subgraph this Env operates on.
self
.
inputs
=
list
(
inputs
)
self
.
outputs
=
list
(
outputs
)
for
feature_class
in
uniq_features
(
features
):
self
.
add_feature
(
feature_class
,
False
)
# All ops in the subgraph defined by inputs and outputs are cached in _ops
self
.
_ops
=
set
()
# Ditto for results
self
.
_results
=
set
(
self
.
inputs
)
# Set of all the results that are not an output of an op in the subgraph but
# are an input of an op in the subgraph.
# e.g. z for inputs=(x, y) and outputs=(x + (y - z),)
...
...
@@ -95,6 +92,9 @@ class Env(graph.Graph):
# it will be removed from the set of orphans.
self
.
_orphans
=
set
(
outputs
)
for
feature_class
in
uniq_features
(
features
):
self
.
add_feature
(
feature_class
,
False
)
# Maps results to ops that use them:
# if op.inputs[i] == v then (op, i) in self._clients[v]
self
.
_clients
=
{}
...
...
gof/op.py
浏览文件 @
87103eff
...
...
@@ -179,15 +179,15 @@ class Op(object):
# C code generators
#
def
c_var_names
(
self
):
"""
Returns ([list of input names], [list of output names]) for
use as C variables.
"""
return
[[
"i
%
i"
%
i
for
i
in
xrange
(
len
(
self
.
inputs
))],
[
"o
%
i"
%
i
for
i
in
xrange
(
len
(
self
.
outputs
))]]
def
c_validate_update
(
self
):
#
def c_var_names(self):
#
"""
#
Returns ([list of input names], [list of output names]) for
#
use as C variables.
#
"""
#
return [["i%i" % i for i in xrange(len(self.inputs))],
#
["o%i" % i for i in xrange(len(self.outputs))]]
def
c_validate_update
(
self
,
inputs
,
outputs
,
sub
):
"""
Returns templated C code that checks that the inputs to this
function can be worked on. If a failure occurs, set an
...
...
@@ -198,13 +198,13 @@ class Op(object):
"""
raise
AbstractFunctionError
()
def
c_validate_update_cleanup
(
self
):
def
c_validate_update_cleanup
(
self
,
inputs
,
outputs
,
sub
):
"""
Clean up things allocated by c_validate().
"""
raise
AbstractFunctionError
()
def
c_code
(
self
):
def
c_code
(
self
,
inputs
,
outputs
,
sub
):
"""
Returns templated C code that does the computation associated
to this Op. You may assume that input validation and output
...
...
@@ -215,7 +215,7 @@ class Op(object):
"""
raise
AbstractFunctionError
()
def
c_code_cleanup
(
self
):
def
c_code_cleanup
(
self
,
inputs
,
outputs
,
sub
):
"""
Clean up things allocated by c_code().
"""
...
...
gof/result.py
浏览文件 @
87103eff
...
...
@@ -175,13 +175,13 @@ class ResultBase(object):
"""
return
False
def
c_declare
(
self
):
def
c_declare
(
self
,
name
,
sub
):
"""
Declares variables that will be instantiated by c_data_extract.
"""
raise
AbstractFunctionError
()
def
c_extract
(
self
):
def
c_extract
(
self
,
name
,
sub
):
"""
The code returned from this function must be templated using
"
%(name)
s", representing the name that the caller wants to
...
...
@@ -193,7 +193,7 @@ class ResultBase(object):
"""
raise
AbstractFunctionError
()
def
c_cleanup
(
self
):
def
c_cleanup
(
self
,
name
,
sub
):
"""
This returns C code that should deallocate whatever
c_data_extract allocated or decrease the reference counts. Do
...
...
@@ -201,7 +201,7 @@ class ResultBase(object):
"""
raise
AbstractFunctionError
()
def
c_sync
(
self
):
def
c_sync
(
self
,
name
,
sub
):
"""
The code returned from this function must be templated using "
%(name)
s",
representing the name that the caller wants to call this Result.
...
...
@@ -297,28 +297,28 @@ class PythonResult(ResultBase):
through
%(name)
s.
"""
def
c_declare
(
self
):
def
c_declare
(
self
,
name
,
sub
):
return
"""
PyObject*
%(name)
s;
"""
"""
%
locals
()
def
c_extract
(
self
):
def
c_extract
(
self
,
name
,
sub
):
return
"""
Py_XINCREF(py_
%(name)
s);
%(name)
s = py_
%(name)
s;
"""
"""
%
locals
()
def
c_cleanup
(
self
):
def
c_cleanup
(
self
,
name
,
sub
):
return
"""
Py_XDECREF(
%(name)
s);
"""
"""
%
locals
()
def
c_sync
(
self
):
def
c_sync
(
self
,
name
,
sub
):
return
"""
Py_XDECREF(py_
%(name)
s);
py_
%(name)
s =
%(name)
s;
Py_XINCREF(py_
%(name)
s);
"""
"""
%
locals
()
def
same_properties
(
self
,
other
):
return
False
...
...
s
andbox/s
calar.py
→
scalar.py
浏览文件 @
87103eff
...
...
@@ -18,10 +18,9 @@ def as_scalar(x, name = None):
class
Scalar
(
ResultBase
):
def
__init__
(
self
,
dtype
,
data
=
None
,
name
=
None
):
def
__init__
(
self
,
dtype
,
name
=
None
):
ResultBase
.
__init__
(
self
,
role
=
None
,
name
=
name
)
self
.
dtype
=
dtype
self
.
constant
=
False
ResultBase
.
__init__
(
self
,
role
=
None
,
data
=
data
,
name
=
name
)
def
__get_constant
(
self
):
return
self
.
_constant
...
...
@@ -40,60 +39,59 @@ class Scalar(ResultBase):
def
same_properties
(
self
,
other
):
return
other
.
dtype
==
self
.
dtype
def
mergeable
(
self
,
other
):
return
getattr
(
self
,
'constant'
,
False
)
\
and
getattr
(
other
,
'constant'
,
False
)
\
and
self
.
data
==
other
.
data
#
def mergeable(self, other):
#
return getattr(self, 'constant', False) \
#
and getattr(other, 'constant', False) \
#
and self.data == other.data
def
dtype_specs
(
self
):
return
{
'float64'
:
(
float
,
'double'
,
'PyFloat_Check'
,
'PyFloat_AsDouble'
,
'PyFloat_FromDouble'
)}[
self
.
dtype
]
# def py_type(self):
# return {'float64': float}[self.dtype]
# def c_type(self):
# return {'float64': 'double'}[self.dtype]
# def c_from(self):
# return {'float64': 'PyFloat_FromDouble'}[self.dtype]
# def c_as(self):
# return {'float64': 'PyFloat_AsDouble'}[self.dtype]
def
c_declare
(
self
):
def
c_declare
(
self
,
name
,
sub
):
return
"""
%(dtype)
s
%
%
(name)
s;
typedef
%(dtype)
s
%
%
(name)
s_dtype;
"""
%
dict
(
dtype
=
self
.
dtype_specs
()[
1
])
%(dtype)
s
%(name)
s;
typedef
%(dtype)
s
%(name)
s_dtype;
"""
%
dict
(
name
=
name
,
dtype
=
self
.
dtype_specs
()[
1
])
def
c_init
(
self
):
def
c_init
(
self
,
name
,
sub
):
return
"""
%(name)
s = 0;
"""
"""
%
locals
()
def
c_extract
(
self
):
def
c_extract
(
self
,
name
,
sub
):
specs
=
self
.
dtype_specs
()
return
"""
if (!
%(check)
s(py_
%%(name)
s))
%%(fail)
s
%%(name)
s = (
%(dtype)
s)
%(conv)
s(py_
%%(name)
s);
"""
%
dict
(
dtype
=
specs
[
1
],
if (!
%(check)
s(py_
%(name)
s))
%(fail)
s
%(name)
s = (
%(dtype)
s)
%(conv)
s(py_
%(name)
s);
"""
%
dict
(
sub
,
name
=
name
,
dtype
=
specs
[
1
],
check
=
specs
[
2
],
conv
=
specs
[
3
])
def
c_sync
(
self
):
def
c_sync
(
self
,
name
,
sub
):
specs
=
self
.
dtype_specs
()
return
"""
Py_XDECREF(py_
%%(name)
s);
py_
%%(name)
s =
%(conv)
s((
%(dtype)
s)
%%(name)
s);
if (!py_
%%(name)
s)
py_
%%(name)
s = Py_None;
"""
%
dict
(
dtype
=
specs
[
1
],
Py_XDECREF(py_
%(name)
s);
py_
%(name)
s =
%(conv)
s((
%(dtype)
s)
%(name)
s);
if (!py_
%(name)
s)
py_
%(name)
s = Py_None;
"""
%
dict
(
name
=
name
,
dtype
=
specs
[
1
],
conv
=
specs
[
4
])
def
c_cleanup
(
self
):
def
c_cleanup
(
self
,
name
,
sub
):
return
""
def
__copy__
(
self
):
"""
Return a copy of this instance (with its own attributes)
"""
cpy
=
self
.
__class__
(
self
.
dtype
,
self
.
name
)
cpy
.
data
=
self
.
data
return
cpy
class
ScalarMixedOp
(
GuardedOp
):
...
...
@@ -104,8 +102,8 @@ class ScalarMixedOp(GuardedOp):
def
__init__
(
self
,
*
inputs
):
if
self
.
nin
>=
0
:
if
len
(
inputs
)
!=
self
.
nin
:
raise
TypeError
(
"Wrong number of inputs for
%
s (got
%
i, expected
%
i)"
)
\
%
(
self
,
len
(
inputs
),
self
.
nin
)
raise
TypeError
(
"Wrong number of inputs for
%
s (got
%
i, expected
%
i)"
\
%
(
self
.
__class__
.
__name__
,
len
(
inputs
),
self
.
nin
)
)
i_dtypes
=
[
getattr
(
input
,
'dtype'
,
None
)
for
input
in
inputs
]
o_dtypes
=
utils
.
from_return_values
(
self
.
propagate_dtypes
(
*
i_dtypes
))
...
...
@@ -125,14 +123,14 @@ class ScalarMixedOp(GuardedOp):
def
perform
(
self
):
self
.
outputs
[
0
]
.
data
=
self
.
impl
(
*
[
input
.
data
for
input
in
self
.
inputs
])
def
c_var_names
(
self
):
(
self
,
inames
,
onames
),
_1
,
_2
,
_3
=
inspect
.
getargspec
(
self
.
c_impl
)
inames
=
utils
.
from_return_values
(
inames
)
onames
=
utils
.
from_return_values
(
onames
)
return
[
inames
,
onames
]
#
def c_var_names(self):
#
(self, inames, onames), _1, _2, _3 = inspect.getargspec(self.c_impl)
#
inames = utils.from_return_values(inames)
#
onames = utils.from_return_values(onames)
#
return [inames, onames]
def
c_code
(
self
):
return
self
.
c_impl
(
self
.
inputs
,
self
.
outputs
)
#
def c_code(self):
#
return self.c_impl(self.inputs, self.outputs)
def
upcast
(
dtype
,
*
dtypes
):
...
...
s
andbox/s
calar_ops.py
→
scalar_ops.py
浏览文件 @
87103eff
...
...
@@ -4,71 +4,91 @@ import math
class
Add
(
BinaryScalarOp
):
identity
=
0
def
impl
(
self
,
x
,
y
):
return
x
+
y
def
c_
impl
(
self
,
(
x
,
y
),
z
):
return
"
%(z)
s =
%(x)
s +
%(y)
s;"
def
grad
(
self
,
(
x
,
y
),
gz
):
def
c_
code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s +
%(y)
s;"
%
locals
()
def
grad
(
self
,
(
x
,
y
),
(
gz
,
)
):
return
gz
,
gz
class
Sub
(
BinaryScalarOp
):
def
impl
(
self
,
x
,
y
):
return
x
-
y
def
c_
impl
(
self
,
(
x
,
y
),
z
):
return
"
%(z)
s =
%(x)
s -
%(y)
s;"
def
grad
(
self
,
(
x
,
y
),
gz
):
def
c_
code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s -
%(y)
s;"
%
locals
()
def
grad
(
self
,
(
x
,
y
),
(
gz
,
)
):
return
gz
,
-
gz
class
Mul
(
BinaryScalarOp
):
def
impl
(
self
,
x
,
y
):
return
x
*
y
def
c_
impl
(
self
,
(
x
,
y
),
z
):
return
"
%(z)
s =
%(x)
s *
%(y)
s;"
def
grad
(
self
,
(
x
,
y
),
gz
):
def
c_
code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s *
%(y)
s;"
%
locals
()
def
grad
(
self
,
(
x
,
y
),
(
gz
,
)
):
return
mul
(
y
,
gz
),
mul
(
x
,
gz
)
class
Div
(
BinaryScalarOp
):
def
impl
(
self
,
x
,
y
):
return
x
/
y
def
c_
impl
(
self
,
(
x
,
y
),
z
):
return
"
%(z)
s =
%(x)
s /
%(y)
s;"
def
grad
(
self
,
(
x
,
y
),
gz
):
def
c_
code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s /
%(y)
s;"
%
locals
()
def
grad
(
self
,
(
x
,
y
),
(
gz
,
)
):
return
div
(
gz
,
y
),
-
div
(
mul
(
x
,
gz
),
y
*
y
)
class
Pow
(
BinaryScalarOp
):
def
impl
(
self
,
x
,
y
):
return
x
**
y
def
c_impl
(
self
,
(
x
,
y
),
z
):
return
"
%(z)
s = pow(
%(x)
s,
%(y)
s);"
def
c_code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s = pow(
%(x)
s,
%(y)
s);"
%
locals
()
class
First
(
BinaryScalarOp
):
def
impl
(
self
,
x
,
y
):
return
x
def
c_code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s;"
%
locals
()
class
Second
(
BinaryScalarOp
):
def
impl
(
self
,
x
,
y
):
return
y
def
c_code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(y)
s;"
%
locals
()
class
SquareDiff
(
BinaryScalarOp
):
def
impl
(
self
,
x
,
y
):
diff
=
(
x
-
y
)
return
diff
*
diff
def
c_code
(
self
,
(
x
,
y
),
(
z
,
),
sub
):
return
"
%(z)
s =
%(x)
s -
%(y)
s;
%(z)
s *=
%(z)
s;"
%
locals
()
class
Neg
(
UnaryScalarOp
):
def
impl
(
self
,
x
):
return
-
x
def
grad
(
self
,
x
,
gz
):
def
grad
(
self
,
(
x
,
),
(
gz
,
)
):
return
-
gz
def
c_
impl
(
self
,
x
,
z
):
return
"
%(z)
s = -
%(x)
s;"
def
c_
code
(
self
,
(
x
,
),
(
z
,
),
sub
):
return
"
%(z)
s = -
%(x)
s;"
%
locals
()
class
Inv
(
UnaryScalarOp
):
def
impl
(
self
,
x
):
return
1
/
x
def
grad
(
self
,
x
,
gz
):
def
grad
(
self
,
(
x
,
),
(
gz
,
)
):
return
-
gz
/
(
x
*
x
)
def
c_
impl
(
self
,
x
,
z
):
return
"
%(z)
s = 1 /
%(x)
s;"
def
c_
code
(
self
,
(
x
,
),
(
z
,
),
sub
):
return
"
%(z)
s = 1 /
%(x)
s;"
%
locals
()
class
Log
(
UnaryScalarOp
):
def
impl
(
self
,
x
):
return
math
.
log
(
x
)
def
c_
impl
(
self
,
x
,
z
):
return
"
%(z)
s = log(
%(x)
s);"
def
c_
code
(
self
,
(
x
,
),
(
z
,
),
sub
):
return
"
%(z)
s = log(
%(x)
s);"
%
locals
()
class
Exp
(
UnaryScalarOp
):
def
impl
(
self
,
x
):
return
math
.
exp
(
x
)
def
c_
impl
(
self
,
x
,
z
):
return
"
%(z)
s = exp(
%(x)
s);"
def
c_
code
(
self
,
(
x
,
),
(
z
,
),
sub
):
return
"
%(z)
s = exp(
%(x)
s);"
%
locals
()
# class Sigmoid(UnaryComposite):
...
...
tensor.py
浏览文件 @
87103eff
...
...
@@ -136,8 +136,12 @@ class _Op(BaseTensorOp):
onames
=
utils
.
from_return_values
(
onames
)
return
[
inames
,
onames
]
def
c_code
(
self
):
return
self
.
c_impl
(
self
.
inputs
,
self
.
outputs
)
def
c_code
(
self
,
input_names
,
output_names
,
sub
):
sub
=
dict
(
sub
)
icvn
,
ocvn
=
self
.
c_var_names
()
for
real
,
tosub
in
zip
(
input_names
+
output_names
,
icvn
+
ocvn
):
sub
[
tosub
]
=
real
return
self
.
c_impl
(
self
.
inputs
,
self
.
outputs
)
%
sub
def
c_impl
(
self
,
inputs
,
outputs
):
raise
AbstractFunctionError
()
...
...
@@ -759,7 +763,7 @@ class Gemm(_Op):
return
blas
.
ldflags
()
def
c_var_names
(
self
):
return
[[
'_z'
,
'_a'
,
'_x'
,
'_y'
,
'_b'
],
[
'_zout'
]]
def
c_validate_update
(
self
):
def
c_validate_update
(
self
,
(
_z
,
_a
,
_x
,
_y
,
_b
),
(
_zout
,
),
sub
):
return
"""
if (
%(_zout)
s)
{
...
...
@@ -770,10 +774,10 @@ class Gemm(_Op):
%(_zout)
s =
%(_z)
s;
Py_INCREF(
%(_zout)
s);
}
"""
def
c_validate_update_cleanup
(
self
):
"""
%
locals
()
def
c_validate_update_cleanup
(
self
,
ignore
,
_ignore
,
__ignore
):
return
""
def
c_code
(
self
):
def
c_code
(
self
,
(
_z
,
_a
,
_x
,
_y
,
_b
),
(
_zout
,
),
sub
):
return
"""
int unit = 0;
...
...
@@ -913,7 +917,7 @@ class Gemm(_Op):
break;
}
"""
"""
%
dict
(
locals
(),
**
sub
)
gemm
=
gof
.
op
.
constructor
(
Gemm
)
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论