Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
14753530
提交
14753530
authored
1月 07, 2008
作者:
olivier@olivier-desktop
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
more files
上级
7ef88361
显示空白字符变更
内嵌
并排
正在显示
4 个修改的文件
包含
613 行增加
和
111 行删除
+613
-111
core.py
core.py
+461
-111
grad.py
grad.py
+44
-0
opt.py
opt.py
+70
-0
rand.py
rand.py
+38
-0
没有找到文件。
core.py
浏览文件 @
14753530
...
@@ -2,72 +2,81 @@
...
@@ -2,72 +2,81 @@
import
gof
import
gof
import
numpy
import
numpy
from
copy
import
copy
from
copy
import
copy
as
pycopy
# __all__ = ['set_mode', 'get_mode', 'NumpyR', 'NumpyOp']
# __all__ = ['set_mode', 'get_mode', 'NumpyR', 'NumpyOp']
_mode
=
'eval'
_mode
=
[
'eval'
]
def
set_mode
(
mode
):
def
set_mode
(
mode
):
global
_mode
_mode
.
append
(
mode
)
_mode
=
mode
def
ge
t_mode
():
def
curren
t_mode
():
return
_mode
return
_mode
def
start_build
():
def
build_mode
():
set_mode
(
'build'
)
set_mode
(
'build'
)
def
end_build
():
def
eval_mode
():
set_mode
(
'eval'
)
def
pop_mode
():
if
len
(
_mode
)
==
1
:
raise
Exception
(
"There's only one mode left on the stack."
)
else
:
_mode
.
pop
()
def
end_eval
():
set_mode
(
'eval'
)
set_mode
(
'eval'
)
def
build
(
f
,
*
args
,
**
kwargs
):
def
build
(
f
,
*
args
,
**
kwargs
):
start_build
()
build_mode
()
r
=
f
(
*
args
,
**
kwargs
)
r
=
f
(
*
args
,
**
kwargs
)
end_build
()
pop_mode
()
return
r
return
r
class
Keyword
:
# class Proxy(object
):
def
__init__
(
self
,
name
,
nonzero
=
True
):
self
.
name
=
name
# __slots__ = ['_obj']
self
.
nonzero
=
nonzero
# def __init__(self, obj = None
):
def
__nonzero__
(
self
):
# self._obj = obj
return
self
.
nonzero
# def __getattribute__(self, attr):
def
__str__
(
self
):
# if attr in ['__class__', '_obj']:
return
"<
%
s>"
%
self
.
name
# return object.__getattribute__(self, attr)
# return getattr(self._obj, attr)
# def __setattr__(self, attr, value):
def
__repr__
(
self
):
# if attr in ['_obj']:
return
str
(
self
)
# object.__setattr__(self, attr, value)
# else:
# setattr(self._obj, attr, value)
# def __delattr__(self, attr):
UNCOMPUTED
=
Keyword
(
"UNCOMPUTED"
,
False
)
# delattr(self._obj, attr
)
UNDEFINED
=
Keyword
(
"UNDEFINED"
,
False
)
# def __str__(self):
# return str(self._obj)
# def __iadd__(self, y):
class
Proxy
(
object
):
# newobj = self._obj.__iadd__(y)
# if isinstance(newobj, Proxy):
# newobj = newobj._obj
# self._obj = newobj
# return self
__slots__
=
[
'_obj'
]
def
__init__
(
self
,
obj
=
None
):
self
.
_obj
=
obj
# class NumpyR(numpy.ndarray):
def
__getattribute__
(
self
,
attr
):
if
attr
in
[
'__class__'
,
'_obj'
]:
return
object
.
__getattribute__
(
self
,
attr
)
else
:
return
getattr
(
object
.
__getattribute__
(
self
,
'_obj'
),
attr
)
# __add__ = wrapper(numpy.ndarray.__add__)
def
__setattr__
(
self
,
attr
,
value
):
# __radd__ = wrapper(numpy.ndarray.__radd__)
if
attr
in
[
'_obj'
]:
object
.
__setattr__
(
self
,
attr
,
value
)
else
:
setattr
(
self
.
_obj
,
attr
,
value
)
def
__delattr__
(
self
,
attr
):
delattr
(
self
.
_obj
,
attr
)
class
IViewer
(
gof
.
ext
.
Viewer
):
class
IViewer
(
gof
.
ext
.
Viewer
):
...
@@ -75,8 +84,11 @@ class IViewer(gof.ext.Viewer):
...
@@ -75,8 +84,11 @@ class IViewer(gof.ext.Viewer):
def
view_map
(
self
):
def
view_map
(
self
):
rval
=
{}
rval
=
{}
for
i
,
o
in
self
.
_v_map
.
items
():
for
output
,
inputs
in
self
.
_v_map
.
items
():
rval
[
self
.
inputs
[
i
]]
=
self
.
outputs
[
o
]
if
isinstance
(
inputs
,
(
list
,
tuple
)):
rval
[
self
.
outputs
[
output
]]
=
[
self
.
inputs
[
i
]
for
i
in
inputs
]
else
:
rval
[
self
.
outputs
[
output
]]
=
self
.
inputs
[
inputs
]
return
rval
return
rval
...
@@ -85,135 +97,473 @@ class IDestroyer(gof.ext.Destroyer):
...
@@ -85,135 +97,473 @@ class IDestroyer(gof.ext.Destroyer):
def
destroy_map
(
self
):
def
destroy_map
(
self
):
rval
=
{}
rval
=
{}
for
i
,
o
in
self
.
_d_map
.
items
():
for
output
,
inputs
in
self
.
_d_map
.
items
():
rval
[
self
.
inputs
[
i
]]
=
self
.
outputs
[
o
]
if
isinstance
(
inputs
,
(
list
,
tuple
)):
rval
[
self
.
outputs
[
output
]]
=
[
self
.
inputs
[
i
]
for
i
in
inputs
]
else
:
rval
[
self
.
outputs
[
output
]]
=
self
.
inputs
[
inputs
]
return
rval
return
rval
class
PythonR
(
gof
.
HolderResult
):
def
__init__
(
self
,
a
=
None
):
if
a
is
None
:
self
.
storage
=
UNCOMPUTED
else
:
self
.
storage
=
a
def
set_value
(
self
,
value
):
self
.
storage
=
value
def
__str__
(
self
):
return
str
(
self
.
storage
)
def
__repr__
(
self
):
return
repr
(
self
.
storage
)
class
NumpyR
(
gof
.
HolderResult
):
def
__init__
(
self
,
a
=
None
):
self
.
set_value
(
a
)
def
set_value
(
self
,
value
):
if
value
is
None
or
value
is
UNCOMPUTED
:
self
.
storage
=
UNCOMPUTED
elif
isinstance
(
value
,
numpy
.
ndarray
):
self
.
storage
=
value
else
:
self
.
storage
=
numpy
.
array
(
value
)
def
__add__
(
self
,
y
):
return
add
(
self
,
y
)
def
__radd__
(
self
,
x
):
return
add
(
x
,
self
)
def
__iadd__
(
self
,
y
):
return
iadd
(
self
,
y
)
def
__sub__
(
self
,
y
):
return
sub
(
self
,
y
)
def
__rsub__
(
self
,
x
):
return
sub
(
x
,
self
)
def
__isub__
(
self
,
y
):
return
isub
(
self
,
y
)
def
__mul__
(
self
,
y
):
return
dot
(
self
,
y
)
def
__rmul__
(
self
,
x
):
return
dot
(
x
,
self
)
def
__imul__
(
self
,
y
):
return
imul
(
self
,
y
)
def
__div__
(
self
,
y
):
return
div
(
self
,
y
)
def
__rdiv__
(
self
,
x
):
return
div
(
x
,
self
)
def
__idiv__
(
self
,
y
):
return
idiv
(
self
,
y
)
def
__mod__
(
self
,
y
):
return
mod
(
self
,
y
)
def
__rmod__
(
self
,
x
):
return
mod
(
x
,
self
)
def
__pow__
(
self
,
y
):
return
pow
(
self
,
y
)
def
__rpow__
(
self
,
x
):
return
pow
(
x
,
self
)
def
__ipow__
(
self
,
y
):
return
ipow
(
self
,
y
)
def
__neg__
(
self
):
return
neg
(
self
)
T
=
property
(
lambda
self
:
transpose
(
self
))
Tc
=
property
(
lambda
self
:
transpose_copy
(
self
))
def
__copy__
(
self
):
return
array_copy
(
self
)
def
__str__
(
self
):
return
str
(
self
.
storage
)
def
__repr__
(
self
):
return
repr
(
self
.
storage
)
#[iadd(iadd(iadd(iadd(<UNCOMPUTED>, itwice(<UNCOMPUTED>)), <UNCOMPUTED>), 1.0), dot(<UNCOMPUTED>, <UNCOMPUTED>))]
#[iadd(iadd(iadd(iadd(<UNCOMPUTED>, itwice(<UNCOMPUTED>)), <UNCOMPUTED>), 1.0), dot(<UNCOMPUTED>, <UNCOMPUTED>))]
def
wrap
(
x
):
# try:
# return to_numpyr(x)
# except TypeError:
# if isinstance(x, PythonR):
# return x
# else:
# return PythonR(x)
# def to_numpyr(x):
if
isinstance
(
x
,
NumpyR
):
return
x
elif
isinstance
(
x
,
PythonR
):
return
x
elif
isinstance
(
x
,
NumpyOp
):
return
x
.
out
elif
isinstance
(
x
,
Proxy
):
return
wrap
(
x
.
_obj
)
elif
isinstance
(
x
,
numpy
.
ndarray
):
return
NumpyR
(
x
)
else
:
return
PythonR
(
x
)
# else:
# raise TypeError("%s cannot be converted to or encapsulated in a NumpyR instance." % x)
class
NumpyOp
(
gof
.
Op
,
gof
.
ext
.
BuildableFromInputs
):
nout
=
1
def
__init__
(
self
,
*
args
):
inputs
=
[
wrap
(
arg
)
for
arg
in
args
]
outputs
=
[
NumpyR
()
for
i
in
xrange
(
self
.
nout
)]
gof
.
Op
.
__init__
(
self
,
inputs
,
outputs
)
@classmethod
def
from_inputs
(
cls
,
*
inputs
):
return
cls
(
*
inputs
)
def
gen_outputs
(
self
):
return
[
NumpyR
()
for
i
in
xrange
(
self
.
nout
)]
class
wrapper
:
class
wrapper
:
__slots__
=
[
'f'
,
'opclass'
]
__slots__
=
[
'f'
,
'opclass'
]
def
__init__
(
self
,
f
,
vmap
=
None
,
dmap
=
None
):
def
__init__
(
self
,
name
,
f
,
grad
,
vmap
=
None
,
dmap
=
None
,
optype
=
NumpyOp
):
self
.
f
=
f
self
.
f
=
f
if
not
callable
(
f
):
if
not
callable
(
f
):
raise
TypeError
(
"Can only wrap a callable."
)
raise
TypeError
(
"Can only wrap a callable."
)
bases
=
[
NumpyOp
]
bases
=
[
optype
]
if
vmap
:
if
vmap
:
bases
.
append
(
IViewer
)
bases
.
append
(
IViewer
)
if
dmap
:
bases
.
append
(
IDestroyer
)
if
dmap
:
bases
.
append
(
IDestroyer
)
if
hasattr
(
f
,
'__module__'
):
mod
=
f
.
__module__
elif
hasattr
(
f
,
'__objclass__'
):
mod
=
f
.
__objclass__
.
__name__
else
:
mod
=
""
Wrapper
=
type
(
"Op:"
+
mod
+
":"
+
f
.
__name__
,
tuple
(
bases
),
{})
Wrapper
=
type
(
name
,
tuple
(
bases
),
{})
if
vmap
:
Wrapper
.
_v_map
=
vmap
if
vmap
:
Wrapper
.
_v_map
=
vmap
if
dmap
:
Wrapper
.
_d_map
=
dmap
if
dmap
:
Wrapper
.
_d_map
=
dmap
def
thunk
(
self
):
def
thunk
(
self
):
def
ret
():
def
ret
():
self
.
outputs
[
0
]
.
s
torage
=
f
(
*
[
input
.
storage
for
input
in
self
.
inputs
]
)
self
.
outputs
[
0
]
.
s
et_value
(
f
(
*
[
input
.
storage
for
input
in
self
.
inputs
])
)
return
ret
return
ret
Wrapper
.
thunk
=
thunk
Wrapper
.
thunk
=
thunk
#.impl = staticmethod(impl)
if
grad
is
UNDEFINED
:
grad
=
lambda
*
_
:
UNDEFINED
Wrapper
.
grad
=
staticmethod
(
grad
)
self
.
opclass
=
Wrapper
self
.
opclass
=
Wrapper
def
__call__
(
self
,
*
args
):
def
__call__
(
self
,
*
args
):
op
=
self
.
opclass
(
*
args
)
op
=
self
.
opclass
(
*
args
)
if
_mode
==
'eval'
:
if
current_mode
()
==
'eval'
:
op
.
thunk
()()
op
.
thunk
()()
# outputs = [Proxy(o) for o in op.outputs]
outputs
=
pycopy
(
op
.
outputs
)
outputs
=
copy
(
op
.
outputs
)
# outputs = [Proxy(output) for output in op.outputs]
if
op
.
nout
==
1
:
if
op
.
nout
==
1
:
return
outputs
[
0
]
return
outputs
[
0
]
else
:
else
:
return
outputs
return
outputs
# if '_mode' in kwargs:
# def wrap_producer(f):
# mode = kwargs['_mode']
# def ret(*args, **kwargs):
# del kwargs['_mode']
# result = f(*args, **kwargs)
# else:
# if not isinstance(result, numpy.ndarray):
# mode = _mode
# result = numpy.array(result)
# return NumpyR(result)
# return ret
# if mode == 'eval':
# return f(*[to_ndarray(arg) for arg in args])
# elif mode == 'build':
# o = self.op(*args)
# elif mode == 'build_eval':
def
wrap_producer
(
f
):
# return f(*args, **kwargs)
wrapped_f
=
wrapper
(
f
.
__name__
,
f
,
UNDEFINED
)
def
ret
(
dim
,
dtype
=
'float'
,
order
=
'C'
):
return
wrapped_f
(
dim
,
dtype
,
order
)
return
ret
class
NumpyR
(
gof
.
HolderResult
):
def
__init__
(
self
,
a
=
None
):
ndarray
=
wrap_producer
(
numpy
.
ndarray
)
if
a
is
None
:
array
=
wrap_producer
(
numpy
.
array
)
self
.
storage
=
None
zeros
=
wrap_producer
(
numpy
.
zeros
)
elif
isinstance
(
a
,
numpy
.
ndarray
):
ones
=
wrap_producer
(
numpy
.
ones
)
self
.
storage
=
a
inplace
=
gof
.
ext
.
Destroyer
view
=
gof
.
ext
.
Viewer
class
omega_op_metaclass
(
type
):
def
__init__
(
cls
,
name
,
bases
,
dct
):
type
.
__init__
(
cls
,
name
,
bases
,
dct
)
cls
.
__clsinit__
(
name
,
bases
,
dct
)
class
omega_op
(
gof
.
Op
,
gof
.
ext
.
BuildableFromInputs
):
__metaclass__
=
omega_op_metaclass
nout
=
1
@classmethod
def
__clsinit__
(
cls
,
name
,
bases
,
dct
):
# make grad and impl static methods
grad
=
cls
.
grad
if
hasattr
(
grad
,
'im_func'
):
grad
=
grad
.
im_func
impl
=
cls
.
impl
if
hasattr
(
cls
.
impl
,
'im_func'
):
impl
=
impl
.
im_func
cls
.
grad
=
staticmethod
(
grad
)
cls
.
impl
=
staticmethod
(
impl
)
def
__new__
(
cls
,
*
inputs
):
op
=
gof
.
Op
.
__new__
(
cls
)
op
.
__init__
(
*
[
wrap
(
input
)
for
input
in
inputs
])
if
current_mode
()
==
'eval'
:
op
.
thunk
()()
if
op
.
nout
==
1
:
return
op
.
out
else
:
else
:
r
aise
TypeError
(
"NumpyR constructor expects a ndarray instance."
)
r
eturn
op
.
outputs
def
__add__
(
self
,
y
):
def
__init__
(
self
,
*
inputs
):
return
add
(
self
,
y
)
for
input
in
inputs
:
assert
isinstance
(
input
,
gof
.
HolderResult
)
gof
.
Op
.
__init__
(
self
,
inputs
,
self
.
gen_outputs
())
def
__mul__
(
self
,
y
):
@classmethod
return
dot
(
self
,
y
)
def
from_inputs
(
cls
,
*
inputs
):
build_mode
()
r
=
cls
(
*
inputs
)
pop_mode
()
return
r
.
owner
def
__iadd__
(
self
,
y
):
def
gen_outputs
(
self
):
return
iadd
(
self
,
y
)
return
[
NumpyR
()
for
i
in
xrange
(
self
.
nout
)]
def
__str__
(
self
):
def
thunk
(
self
):
return
str
(
self
.
storage
)
def
ret
():
results
=
self
.
impl
(
*
[
input
.
storage
for
input
in
self
.
inputs
])
if
self
.
nout
==
1
:
self
.
out
.
set_value
(
results
)
else
:
assert
self
.
nout
==
len
(
results
)
for
result
,
output
in
zip
(
results
,
self
.
outputs
):
output
.
set_value
(
result
)
return
ret
def
__repr__
(
self
):
def
update_gradient
(
self
,
grad_d
):
return
repr
(
self
.
storage
)
inputgs
=
self
.
grad
(
*
(
self
.
inputs
+
[
grad_d
[
output
]
for
output
in
self
.
outputs
]))
if
not
isinstance
(
inputgs
,
(
list
,
tuple
)):
inputgs
=
[
inputgs
]
*
len
(
self
.
inputs
)
for
input
,
inputg
in
zip
(
self
.
inputs
,
inputgs
):
grad_d
.
add
(
input
,
inputg
)
def
grad
(
*
args
):
return
UNDEFINED
def
impl
(
*
args
):
raise
NotImplementedError
(
"This op has no implementation."
)
def
to_numpyr
(
x
):
if
isinstance
(
x
,
NumpyR
):
## Addition ##
class
proto_add
(
omega_op
):
def
grad
(
x
,
y
,
gz
):
return
gz
class
add
(
proto_add
):
impl
=
numpy
.
ndarray
.
__add__
class
iadd
(
proto_add
,
inplace
):
impl
=
numpy
.
ndarray
.
__iadd__
class
proto_twice
(
omega_op
):
def
grad
(
x
,
gz
):
return
scal
(
gz
,
2.0
)
class
twice
(
proto_twice
):
def
impl
(
x
):
return
x
+
x
class
itwice
(
proto_twice
,
inplace
):
def
impl
(
x
):
x
+=
x
return
x
return
x
elif
isinstance
(
x
,
NumpyOp
):
return
x
.
out
# elif isinstance(x, Proxy):
# return to_numpyr(x._obj)
elif
isinstance
(
x
,
numpy
.
ndarray
):
return
NumpyR
(
x
)
else
:
raise
TypeError
(
"
%
s cannot be converted to or encapsulated in a NumpyR instance."
%
x
)
## Subtraction ##
class
NumpyOp
(
gof
.
Op
):
class
proto_sub
(
omega_op
):
def
grad
(
x
,
y
,
gz
):
return
gz
,
-
gz
nout
=
1
class
sub
(
proto_sub
):
impl
=
numpy
.
ndarray
.
__sub__
def
__init__
(
self
,
*
args
):
class
isub
(
proto_sub
,
inplace
):
inputs
=
[
to_numpyr
(
arg
)
for
arg
in
args
]
impl
=
numpy
.
ndarray
.
__isub__
outputs
=
[
NumpyR
()
for
i
in
xrange
(
self
.
nout
)]
gof
.
Op
.
__init__
(
self
,
inputs
,
outputs
)
## Element-wise multiplication ##
def
mul_grad
(
x
,
y
,
gz
):
return
mul
(
y
,
gz
),
mul
(
x
,
gz
)
mul
=
wrapper
(
"mul"
,
numpy
.
ndarray
.
__mul__
,
mul_grad
)
imul
=
wrapper
(
"imul"
,
numpy
.
ndarray
.
__imul__
,
mul_grad
,
dmap
=
{
0
:
0
})
def
sqr_grad
(
x
,
gz
):
return
scal
(
mul
(
x
,
gz
),
2.0
)
sqr
=
wrapper
(
"sqr"
,
lambda
x
:
numpy
.
multiply
(
x
,
x
),
sqr_grad
)
isqr
=
wrapper
(
"isqr"
,
lambda
x
:
x
.
__imul__
(
x
),
sqr_grad
,
dmap
=
{
0
:
0
})
def
sqrt_grad
(
x
,
gz
):
return
scal
(
div
(
gz
,
sqrt
(
x
)),
0.5
)
sqrt
=
wrapper
(
"sqrt"
,
lambda
x
:
numpy
.
sqrt
(
x
),
sqrt_grad
)
isqrt
=
wrapper
(
"isqrt"
,
lambda
x
:
x
.
__ipow__
(
0.5
),
sqrt_grad
,
dmap
=
{
0
:
0
})
## Element-wise division ##
def
div_grad
(
x
,
y
,
gz
):
return
div
(
gz
,
y
),
-
div
(
mul
(
x
,
gz
),
sqr
(
y
))
div
=
wrapper
(
"div"
,
numpy
.
ndarray
.
__div__
,
div_grad
)
idiv
=
wrapper
(
"idiv"
,
numpy
.
ndarray
.
__idiv__
,
div_grad
,
dmap
=
{
0
:
0
})
## Scaling ##
def
scal_grad
(
x
,
a
,
gz
):
return
scal
(
a
,
gz
),
sum
(
mul
(
x
,
gz
))
scal
=
wrapper
(
"scal"
,
numpy
.
ndarray
.
__mul__
,
scal_grad
)
iscal
=
wrapper
(
"iscal"
,
numpy
.
ndarray
.
__imul__
,
scal_grad
,
dmap
=
{
0
:
0
})
neg
=
wrapper
(
"neg"
,
numpy
.
ndarray
.
__neg__
,
lambda
x
,
gz
:
-
gz
)
ineg
=
wrapper
(
"ineg"
,
lambda
x
:
x
.
__imul__
(
-
1
),
lambda
x
,
gz
:
-
gz
,
dmap
=
{
0
:
0
})
## Dot product ##
dot
=
wrapper
(
"dot"
,
numpy
.
dot
,
lambda
x
,
y
,
gz
:
(
dot
(
gz
,
transpose
(
y
)),
dot
(
transpose
(
x
),
gz
)))
## Transposition ##
class
transpose
(
omega_op
,
view
):
impl
=
numpy
.
transpose
def
grad
(
x
,
gz
):
return
transpose_copy
(
gz
)
# transpose = wrapper("transpose",
# numpy.transpose,
# lambda x, z, gz: transpose_copy(gz),
# vmap = {0: 0})
def
transpose_copy
(
x
):
return
array_copy
(
transpose
(
x
))
## Copy ##
class
array_copy
(
omega_op
):
impl
=
numpy
.
array
grad
=
lambda
x
,
gz
:
gz
## Others ##
class
minmax
(
omega_op
):
nout
=
2
def
impl
(
x
):
return
x
.
min
,
x
.
max
# array_copy = wrapper("copy",
# numpy.array,
# lambda x, gz: gz)
# def __validate__(self):
# for input in self.inputs:
# assert isinstance(input, NumpyR)
# for output in self.outputs:
# assert isinstance(output, NumpyR)
# array slicing
add
=
wrapper
(
numpy
.
ndarray
.
__add__
)
iadd
=
wrapper
(
numpy
.
ndarray
.
__iadd__
,
None
,
{
0
:
0
})
dot
=
wrapper
(
numpy
.
dot
)
grad.py
0 → 100644
浏览文件 @
14753530
import
gof
import
core
class
_GradD
(
dict
):
"""A dictionary-like class, into which derivative expressions may be added"""
def
add
(
self
,
r
,
dr
):
"""Add dv to the sum of gradients associated with v"""
if
r
is
core
.
UNDEFINED
:
self
[
r
]
=
core
.
UNDEFINED
elif
r
in
self
:
self
[
r
]
=
self
[
r
]
+
dr
else
:
self
[
r
]
=
dr
def
expand_grad
(
i
,
o
,
cost_derivs
):
grad_d
=
_GradD
(
cost_derivs
)
core
.
build_mode
()
for
op
in
gof
.
graph
.
io_toposort
(
i
,
o
)
.
__reversed__
():
op
.
update_gradient
(
grad_d
)
# inputgs = op.grad(*(op.inputs + [grad_d[output] for output in op.outputs]))
# if not isinstance(inputgs, (list, tuple)):
# inputgs = [inputgs] * len(op.inputs)
# for input, inputg in zip(op.inputs, inputgs):
# grad_d.add(input, inputg)
core
.
pop_mode
()
return
grad_d
def
grad
(
cost
,
wrt
,
cost_grad
=
1.0
):
# cost, wrt = core.wrap(cost), core.wrap(wrt)
cost_derivs
=
expand_grad
([
wrt
],
[
cost
],
{
cost
:
core
.
wrap
(
cost_grad
)})
# print wrt
# for k, v in cost_derivs.items():
# print k, v
ret
=
cost_derivs
.
get
(
wrt
,
None
)
if
ret
is
core
.
UNDEFINED
:
raise
Exception
(
"The gradient wrt
%
s is undefined."
%
wrt
)
return
ret
opt.py
0 → 100644
浏览文件 @
14753530
from
core
import
*
import
gof
def
pattern_opt
(
in_pattern
,
out_pattern
):
def
parse
(
x
):
if
isinstance
(
x
,
(
list
,
tuple
)):
return
[
parse
(
y
)
for
y
in
x
]
elif
isinstance
(
x
,
wrapper
):
return
x
.
opclass
elif
isinstance
(
x
,
str
)
or
(
hasattr
(
x
,
'__bases__'
)
and
issubclass
(
x
,
gof
.
op
.
Op
)):
return
x
else
:
raise
TypeError
(
"Bad input type for pattern_opt."
)
return
gof
.
opt
.
PatternOptimizer
(
parse
(
in_pattern
),
parse
(
out_pattern
))
def
op_sub
(
op1
,
op2
):
if
isinstance
(
op1
,
wrapper
):
op1
=
op1
.
opclass
if
isinstance
(
op2
,
wrapper
):
op2
=
op2
.
opclass
return
gof
.
opt
.
OpSubOptimizer
(
op1
,
op2
)
#def make_patterns(patterns):
# return [name, pattern_opt(inp, outp) for name, inp, outp in patterns]
def
export_opts
(
opts
):
for
name
,
opt
in
opts
:
if
name
:
globals
()[
name
]
=
opt
# double_transpose_eliminator = pattern_opt((transpose, (transpose, 'x')), 'x')
# patterns = make_patterns(patterns)
# export_patterns(patterns)
# List of optimizations to perform. They are listed in the order they are applied.
opts
=
[
[
'double_transpose_eliminator'
,
pattern_opt
((
transpose
,
(
transpose
,
'x'
)),
'x'
)],
[
'addxx_to_twice'
,
pattern_opt
((
add
,
'x'
,
'x'
),
(
twice
,
'x'
))],
[
'twice_to_itwice'
,
op_sub
(
twice
,
itwice
)],
[
'mulxx_to_twice'
,
pattern_opt
((
mul
,
'x'
,
'x'
),
(
sqr
,
'x'
))],
[
'sqr_to_isqr'
,
op_sub
(
sqr
,
isqr
)],
[
'add_to_iadd'
,
op_sub
(
add
,
iadd
)],
[
'add_to_iadd_reverse'
,
pattern_opt
((
add
,
'x'
,
'y'
),
(
iadd
,
'y'
,
'x'
))],
[
'remove_copies'
,
gof
.
opt
.
OpRemover
(
array_copy
)],
[
None
,
gof
.
lib
.
DummyRemover
]
# has to be at the end
]
export_opts
(
opts
)
# publish the optimizations performed under individual names
optimizer
=
gof
.
opt
.
MergeOptMerge
(
gof
.
opt
.
SeqOptimizer
([
opt
for
name
,
opt
in
opts
]))
rand.py
0 → 100644
浏览文件 @
14753530
import
core
import
gof
from
numpy
import
random
as
r
# def rwrap(f):
# wrapped =
# def ret(self, *args):
class
RandomState
(
gof
.
Op
,
gof
.
ext
.
IONames
):
input_names
=
[
'seed'
]
def
__init__
(
self
,
seed
):
inputs
=
[
wrap
(
seed
)]
outputs
=
[
PythonR
()]
gof
.
Op
.
__init__
(
self
,
inputs
,
outputs
)
def
thunk
(
self
):
def
f
():
self
.
out
.
storage
=
r
.
RandomState
(
self
.
seed
.
storage
)
return
f
class
Random
(
object
):
def
__init__
(
seed
):
self
.
state
=
core
.
wrap
(
seed
)
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论