Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
a7e53331
提交
a7e53331
authored
3月 23, 2008
作者:
james@mackie
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
moved tensorOp impl and perform, as well as _constructor to gof.op
上级
a86d558a
隐藏空白字符变更
内嵌
并排
正在显示
2 个修改的文件
包含
67 行增加
和
60 行删除
+67
-60
op.py
gof/op.py
+34
-5
tensor.py
tensor.py
+33
-55
没有找到文件。
gof/op.py
浏览文件 @
a7e53331
...
...
@@ -15,6 +15,16 @@ __all__ = ['Op',
]
def
constructor
(
op_cls
):
"""Make an Op look like a Result-valued function."""
def
f
(
*
args
,
**
kwargs
):
op
=
op_cls
(
*
args
,
**
kwargs
)
if
len
(
op
.
outputs
)
>
1
:
return
op
.
outputs
else
:
return
op
.
outputs
[
0
]
return
f
class
Op
(
object
):
"""
Op represents a computation on the storage in its 'inputs' slot,
...
...
@@ -41,9 +51,8 @@ class Op(object):
doc
=
"Same as self.outputs[0] if this Op's has_default_output field is True."
)
def
__init__
(
self
,
*
inputs
):
# this might be a bit brainless
raise
AbstractFunctionError
(
"Op is an abstract class. Its constructor does nothing, you must override it."
)
def
__init__
(
self
,
**
kwargs
):
pass
def
get_input
(
self
,
i
):
return
self
.
_inputs
[
i
]
...
...
@@ -114,13 +123,33 @@ class Op(object):
#
# perform
#
def
impl
(
self
,
*
args
):
"""Return output data [tuple], given input data
If this Op has a single output (len(self.outputs)==1) then the return
value of this function will be assigned to self.outputs[0].data.
If this Op has multiple otuputs, then this function should return a
tuple with the data for outputs[0], outputs[1], outputs[2], etc.
"""
raise
AbstractFunctionError
()
def
perform
(
self
):
"""
Performs the computation associated to this Op and places the
result(s) in the output Results.
TODO: consider moving this function to the python linker.
"""
raise
AbstractFunctionError
()
res
=
self
.
impl
(
*
[
input
.
data
for
input
in
self
.
inputs
])
if
self
.
nout
==
1
:
self
.
outputs
[
0
]
.
data
=
res
else
:
assert
len
(
res
)
==
len
(
self
.
outputs
)
for
output
,
value
in
zip
(
self
.
outputs
,
res
):
output
.
data
=
value
#
...
...
@@ -196,7 +225,7 @@ class Op(object):
raise
AbstractFunctionError
()
#TODO: consider adding a flag to the base class that toggles this behaviour
class
GuardedOp
(
Op
):
"""An Op that disallows input properties to change after construction"""
...
...
tensor.py
浏览文件 @
a7e53331
...
...
@@ -6,6 +6,7 @@ import numpy
from
gof
import
ResultBase
,
Op
,
utils
,
Destroyer
,
Viewer
,
AbstractFunctionError
import
gof.result
import
gof.op
from
base_tensor
import
BaseTensor
,
BaseTensorOp
from
elemwise
import
Elemwise
...
...
@@ -130,17 +131,6 @@ class _Op(BaseTensorOp):
def
input_wrapper
(
cls
,
obj
):
return
_as_tensor
(
obj
)
def
impl
(
self
,
*
inputs
):
raise
AbstractFunctionError
()
def
perform
(
self
):
res
=
self
.
impl
(
*
[
input
.
data
for
input
in
self
.
inputs
])
if
self
.
nout
==
1
:
self
.
outputs
[
0
]
.
data
=
res
else
:
for
output
,
value
in
zip
(
self
.
outputs
,
res
):
output
.
data
=
value
def
c_var_names
(
self
):
(
self
,
inames
,
onames
),
_1
,
_2
,
_3
=
inspect
.
getargspec
(
self
.
c_impl
)
inames
=
utils
.
from_return_values
(
inames
)
...
...
@@ -231,18 +221,6 @@ class TensorScalarOp(_Elemwise):
def
c_code_foreach
(
self
):
return
"
%%(z)
s_i =
%
s;"
%
self
.
c_expr
def
_constructor
(
op_cls
):
"""Return a function that calls op_cls(*input)
and returns the outputs of the op (with single outputs unpacked)
"""
def
f
(
*
args
,
**
kwargs
):
op
=
op_cls
(
*
args
,
**
kwargs
)
if
len
(
op
.
outputs
)
>
1
:
return
op
.
outputs
else
:
return
op
.
outputs
[
0
]
return
f
##########################
# Unary Operations
...
...
@@ -276,7 +254,7 @@ class Argmax(Op):
x
=
self
.
inputs
[
0
]
.
data
self
.
outputs
[
0
]
.
data
=
numpy
.
max
(
x
,
axis
)
self
.
outputs
[
1
]
.
data
=
numpy
.
argmax
(
x
,
axis
)
argmax
=
_
constructor
(
Argmax
)
argmax
=
gof
.
op
.
constructor
(
Argmax
)
def
max
(
x
,
axis
=
None
):
"""Return maximum elements obtained by iterating over given axis
...
...
@@ -292,7 +270,7 @@ class Exp(_Elemwise):
def
impl
(
self
,
x
):
return
numpy
.
exp
(
x
)
def
grad
(
self
,
x
,
gz
):
return
gz
*
exp
(
x
)
def
c_foreach
(
self
,
(
x_i
,
),
(
z_i
,
)):
return
"z_i = exp(x_i);"
exp
=
_
constructor
(
Exp
)
exp
=
gof
.
op
.
constructor
(
Exp
)
class
Neg
(
_Elemwise
):
...
...
@@ -308,13 +286,13 @@ class Log(_Elemwise):
def
impl
(
self
,
x
):
return
numpy
.
log
(
x
)
def
grad
(
self
,
x
,
gz
):
return
gz
/
x
def
c_foreach
(
self
,
(
x_i
,
),
(
z_i
,
)):
return
"z_i = log(x_i);"
log
=
_
constructor
(
Log
)
log
=
gof
.
op
.
constructor
(
Log
)
class
Log2
(
_Elemwise
):
def
impl
(
self
,
x
):
return
numpy
.
log2
(
x
)
def
grad
(
self
,
x
,
gz
):
return
gz
/
(
x
*
numpy
.
log
(
2.0
))
def
c_foreach
(
self
,
(
x_i
,
),
(
z_i
,
)):
return
"
%(z)
s_i = log2(
%(x)
s_i);"
log2
=
_
constructor
(
Log2
)
log2
=
gof
.
op
.
constructor
(
Log2
)
class
Sgn
(
_Elemwise
):
def
impl
(
self
,
x
):
...
...
@@ -323,19 +301,19 @@ class Sgn(_Elemwise):
return
[
None
]
def
c_foreach
(
self
,
(
x_i
,
),
(
z_i
,
)):
return
"
%(z)
s_i =
%(x)
s_i/abs(
%(x)
s_i);"
# TODO: C use copysign
sgn
=
_
constructor
(
Sgn
)
sgn
=
gof
.
op
.
constructor
(
Sgn
)
class
Sqr
(
_Elemwise
):
def
impl
(
self
,
x
):
return
x
*
x
def
grad
(
self
,
x
,
gz
):
return
2.0
*
x
*
gz
def
c_foreach
(
self
,
(
x_i
,
),
(
z_i
,
)):
return
"
%(z)
s_i =
%(x)
s_i *
%(x)
s_i;"
sqr
=
_
constructor
(
Sqr
)
sqr
=
gof
.
op
.
constructor
(
Sqr
)
class
Sqrt
(
_Elemwise
):
def
impl
(
self
,
x
):
return
numpy
.
sqrt
(
x
)
def
grad
(
self
,
x
,
gz
):
return
0.5
*
gz
/
sqrt
(
x
)
def
c_foreach
(
self
,
(
x_i
,
),
(
z_i
,
)):
return
"
%(z)
s_i = sqrt(
%(x)
s_i);"
sqrt
=
_
constructor
(
Sqrt
)
sqrt
=
gof
.
op
.
constructor
(
Sqrt
)
class
Sum
(
_Elemwise
):
def
impl
(
self
,
x
):
...
...
@@ -348,7 +326,7 @@ class Sum(_Elemwise):
return
"dtype_
%(sum)
s*
%(sum)
sp = ((dtype_
%(sum)
s*)PyArray_DATA(
%(sum)
s));
%(sum)
sp[0] = 0;"
def
c_foreach
(
self
,
(
x_i
,
),
(
sum
,
)):
return
"
%(sum)
sp[0] +=
%(x)
s_i;"
sum
=
_
constructor
(
Sum
)
sum
=
gof
.
op
.
constructor
(
Sum
)
class
Fill
(
_Elemwise
):
def
impl
(
self
,
model
,
value
):
...
...
@@ -359,7 +337,7 @@ class Fill(_Elemwise):
return
"dtype_
%(value)
s
%(value)
s0 = ((dtype_
%(value)
s*)PyArray_DATA(
%(value)
s))[0];"
def
c_foreach
(
self
,
(
model_i
,
value
),
(
z_i
,
)):
return
"
%(z)
s_i =
%(value)
s0;"
fill
=
_
constructor
(
Fill
)
fill
=
gof
.
op
.
constructor
(
Fill
)
def
ones_like
(
model
):
return
fill
(
model
,
1.0
)
def
zeros_like
(
model
):
...
...
@@ -373,7 +351,7 @@ class TensorCopy(_Elemwise):
return
gz
def
c_foreach
(
self
,
(
x_i
,
),
(
z_i
,
)):
return
"
%(z)
s_i =
%(x)
s_i;"
tensor_copy
=
_
constructor
(
TensorCopy
)
tensor_copy
=
gof
.
op
.
constructor
(
TensorCopy
)
##########################
# View Operations
...
...
@@ -399,7 +377,7 @@ class Transpose(_Op, Viewer):
}
%(z)
s = transposed;
"""
transpose
=
_
constructor
(
Transpose
)
transpose
=
gof
.
op
.
constructor
(
Transpose
)
class
Subtensor
(
Op
,
Viewer
):
nin
=
2
...
...
@@ -461,7 +439,7 @@ class Subtensor(Op, Viewer):
# - option: return gz, but think about how to include a special addition
# function that works on a corresponding view of the original data
raise
NotImplementedError
()
subtensor
=
_
constructor
(
Subtensor
)
subtensor
=
gof
.
op
.
constructor
(
Subtensor
)
##########################
...
...
@@ -481,14 +459,14 @@ class AddElemwise(_Elemwise):
return
gz
,
gz
def
c_foreach
(
self
,
(
x_i
,
y_i
),
(
z_i
,
)):
return
"
%(z)
s_i =
%(x)
s_i +
%(y)
s_i;"
add_elemwise
=
_
constructor
(
AddElemwise
)
add_elemwise
=
gof
.
op
.
constructor
(
AddElemwise
)
class
AddElemwiseInplace
(
AddElemwise
.
inplace_version
()):
def
impl
(
self
,
x
,
y
):
_assert_same_shapes
(
x
,
y
)
x
+=
y
return
x
add_elemwise_inplace
=
_
constructor
(
AddElemwiseInplace
)
add_elemwise_inplace
=
gof
.
op
.
constructor
(
AddElemwiseInplace
)
# Scalar #
class
AddScalar
(
TensorScalarOp
):
...
...
@@ -498,14 +476,14 @@ class AddScalar(TensorScalarOp):
def
grad
(
self
,
(
x
,
a
),
gz
):
return
gz
,
sum
(
gz
)
c_expr
=
"x_i + a"
add_scalar
=
_
constructor
(
AddScalar
)
add_scalar
=
gof
.
op
.
constructor
(
AddScalar
)
class
AddScalarInplace
(
AddScalar
.
inplace_version
()):
def
impl
(
self
,
x
,
a
):
_assert_tensor_scalar
(
x
,
a
)
x
+=
a
return
x
add_scalar_inplace
=
_
constructor
(
AddScalarInplace
)
add_scalar_inplace
=
gof
.
op
.
constructor
(
AddScalarInplace
)
add
=
_scalar_switch
(
add_elemwise
,
add_scalar
,
add_scalar
)
add_inplace
=
_scalar_switch
(
add_elemwise_inplace
,
add_scalar_inplace
)
...
...
@@ -524,14 +502,14 @@ class SubElemwise(_Elemwise):
return
gz
,
-
gz
def
c_foreach
(
self
,
(
x_i
,
y_i
),
(
z_i
,
)):
return
"
%(z)
s_i =
%(x)
s_i -
%(y)
s_i;"
sub_elemwise
=
_
constructor
(
SubElemwise
)
sub_elemwise
=
gof
.
op
.
constructor
(
SubElemwise
)
class
SubElemwiseInplace
(
SubElemwise
.
inplace_version
()):
def
impl
(
self
,
x
,
y
):
_assert_same_shapes
(
x
,
y
)
x
-=
y
return
x
sub_elemwise_inplace
=
_
constructor
(
SubElemwiseInplace
)
sub_elemwise_inplace
=
gof
.
op
.
constructor
(
SubElemwiseInplace
)
# Scalar #
def
sub_scalar_r
(
x
,
a
):
...
...
@@ -559,14 +537,14 @@ class MulElemwise(_Elemwise):
return
mul
(
y
,
gz
),
mul
(
x
,
gz
)
def
c_foreach
(
self
,
(
x_i
,
y_i
),
(
z_i
,
)):
return
"
%(z)
s_i =
%(x)
s_i *
%(y)
s_i;"
mul_elemwise
=
_
constructor
(
MulElemwise
)
mul_elemwise
=
gof
.
op
.
constructor
(
MulElemwise
)
class
MulElemwiseInplace
(
MulElemwise
.
inplace_version
()):
def
impl
(
self
,
x
,
y
):
_assert_same_shapes
(
x
,
y
)
x
*=
y
return
x
mul_elemwise_inplace
=
_
constructor
(
MulElemwiseInplace
)
mul_elemwise_inplace
=
gof
.
op
.
constructor
(
MulElemwiseInplace
)
# Scalar #
class
Scale
(
TensorScalarOp
):
...
...
@@ -576,14 +554,14 @@ class Scale(TensorScalarOp):
def
grad
(
self
,
(
x
,
a
),
gz
):
return
scale
(
a
,
gz
),
sum
(
mul_elemwise
(
x
,
gz
))
c_expr
=
"
%(x)
s_i * _
%(a)
s"
scale
=
_
constructor
(
Scale
)
scale
=
gof
.
op
.
constructor
(
Scale
)
class
ScaleInplace
(
Scale
.
inplace_version
()):
def
impl
(
self
,
x
,
a
):
_assert_tensor_scalar
(
x
,
a
)
x
*=
a
return
x
scale_inplace
=
_
constructor
(
ScaleInplace
)
scale_inplace
=
gof
.
op
.
constructor
(
ScaleInplace
)
mul
=
_scalar_switch
(
mul_elemwise
,
scale
,
scale
)
mul_inplace
=
_scalar_switch
(
mul_elemwise_inplace
,
scale_inplace
)
...
...
@@ -602,14 +580,14 @@ class DivElemwise(_Elemwise):
return
div
(
gz
,
y
),
-
div
(
mul
(
x
,
gz
),
(
y
*
y
))
def
c_foreach
(
self
,
(
x_i
,
y_i
),
(
z_i
,
)):
return
"
%(z)
s_i =
%(x)
s_i /
%(y)
s_i;"
div_elemwise
=
_
constructor
(
DivElemwise
)
div_elemwise
=
gof
.
op
.
constructor
(
DivElemwise
)
class
DivElemwiseInplace
(
DivElemwise
.
inplace_version
()):
def
impl
(
self
,
x
,
y
):
_assert_same_shapes
(
x
,
y
)
x
/=
y
return
x
div_elemwise_inplace
=
_
constructor
(
DivElemwiseInplace
)
div_elemwise_inplace
=
gof
.
op
.
constructor
(
DivElemwiseInplace
)
class
InvElemwise
(
_Elemwise
):
def
impl
(
self
,
x
):
...
...
@@ -619,7 +597,7 @@ class InvElemwise(_Elemwise):
return
-
gz
*
(
ix
*
ix
)
def
c_foreach
(
self
,
(
x_i
,
),
(
z_i
,
)):
return
"
%(z)
s_i = 1.0 /
%(x)
s_i;"
#TODO: cast 1.0 to the dtype of x
inv_elemwise
=
_
constructor
(
InvElemwise
)
inv_elemwise
=
gof
.
op
.
constructor
(
InvElemwise
)
# Scalar #
def
div_scalar_r
(
x
,
a
):
...
...
@@ -653,14 +631,14 @@ class PowElemwise(_Elemwise):
return
gx
,
gy
def
c_foreach
(
self
,
(
x_i
,
y_i
),
(
z_i
,
)):
return
"
%(z)
s_i = pow(
%(x)
s_i,
%(y)
s_i);"
pow_elemwise
=
_
constructor
(
PowElemwise
)
pow_elemwise
=
gof
.
op
.
constructor
(
PowElemwise
)
class
PowElemwiseInplace
(
PowElemwise
.
inplace_version
()):
def
impl
(
self
,
x
,
y
):
_assert_same_shapes
(
x
,
y
)
x
**=
y
return
x
pow_elemwise_inplace
=
_
constructor
(
PowElemwiseInplace
)
pow_elemwise_inplace
=
gof
.
op
.
constructor
(
PowElemwiseInplace
)
# Scalar #
class
PowScalarL
(
TensorScalarOp
):
...
...
@@ -672,7 +650,7 @@ class PowScalarL(TensorScalarOp):
gy
=
gz
*
log
(
x
)
*
x
**
y
return
gy
,
gx
c_expr
=
"pow(
%(a)
s,
%(x)
s_i)"
pow_scalar_l
=
_
constructor
(
PowScalarL
)
pow_scalar_l
=
gof
.
op
.
constructor
(
PowScalarL
)
class
PowScalarR
(
TensorScalarOp
):
def
impl
(
self
,
x
,
a
):
...
...
@@ -683,14 +661,14 @@ class PowScalarR(TensorScalarOp):
gs
=
sum
(
mul_elemwise
(
mul_elemwise
(
gz
,
pow_scalar_r
(
x
,
s
)),
log
(
x
)))
return
gx
,
gs
c_expr
=
"pow(
%(x)
s_i, _
%(a)
s)"
pow_scalar_r
=
_
constructor
(
PowScalarR
)
pow_scalar_r
=
gof
.
op
.
constructor
(
PowScalarR
)
class
PowScalarRInplace
(
PowScalarR
.
inplace_version
()):
def
impl
(
self
,
x
,
a
):
_assert_tensor_scalar
(
x
,
a
)
x
**=
a
return
x
pow_scalar_r_inplace
=
_
constructor
(
PowScalarRInplace
)
pow_scalar_r_inplace
=
gof
.
op
.
constructor
(
PowScalarRInplace
)
pow
=
_scalar_switch
(
pow_elemwise
,
pow_scalar_r
,
pow_scalar_l
)
pow_inplace
=
_scalar_switch
(
pow_elemwise_inplace
,
pow_scalar_r_inplace
)
...
...
@@ -729,7 +707,7 @@ class Dot(_Op):
return
blas
.
ldflags
()
def
c_impl
(
self
,
(
_x
,
_y
),
(
_z
,
)):
return
blas
.
gemm_code
(
''
,
'1.0'
,
'0.0'
)
dot
=
_
constructor
(
Dot
)
dot
=
gof
.
op
.
constructor
(
Dot
)
class
Gemm
(
_Op
):
nin
=
5
...
...
@@ -788,7 +766,7 @@ class Gemm(_Op):
return
blas
.
gemm_code
(
check_ab
,
'(_a->descr->type_num == PyArray_FLOAT) ? (REAL)(((float*)_a->data)[0]) : (REAL)(((double*)_a->data)[0])'
,
'(_b->descr->type_num == PyArray_FLOAT) ? (REAL)(((float*)_b->data)[0]) : (REAL)(((double*)_b->data)[0])'
)
gemm
=
_
constructor
(
Gemm
)
gemm
=
gof
.
op
.
constructor
(
Gemm
)
if
0
:
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论