Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
0c1711f5
提交
0c1711f5
authored
8月 05, 2015
作者:
Iban Harlouchet
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
numpydoc for theano/tensor/blas.py
上级
47bf7427
隐藏空白字符变更
内嵌
并排
正在显示
1 个修改的文件
包含
81 行增加
和
60 行删除
+81
-60
blas.py
theano/tensor/blas.py
+81
-60
没有找到文件。
theano/tensor/blas.py
浏览文件 @
0c1711f5
...
@@ -17,10 +17,12 @@ There are four kinds of BLAS Ops in Theano:
...
@@ -17,10 +17,12 @@ There are four kinds of BLAS Ops in Theano:
- C-based (blas_c)
- C-based (blas_c)
- CUDA-based (theano.sandbox.cuda.blas)
- CUDA-based (theano.sandbox.cuda.blas)
:note: Unfortunately (because it's confusing) this file currently contains Ops
Notes
that contain both Python and C versions. I think it would be better to
-----
move the C implementations to blas_c so that this file is pure Python.
Unfortunately (because it's confusing) this file currently contains Ops
-JB
that contain both Python and C versions. I think it would be better to
move the C implementations to blas_c so that this file is pure Python.
-JB
Ops
Ops
...
@@ -121,7 +123,6 @@ Specialize Gemm to Gemv
...
@@ -121,7 +123,6 @@ Specialize Gemm to Gemv
If arguments to GEMM are dimshuffled vectors, then we can use GEMV
If arguments to GEMM are dimshuffled vectors, then we can use GEMV
instead. This optimization is `local_gemm_to_gemv`.
instead. This optimization is `local_gemm_to_gemv`.
"""
"""
from
__future__
import
print_function
from
__future__
import
print_function
import
copy
import
copy
...
@@ -359,7 +360,9 @@ class Gemv(Op):
...
@@ -359,7 +360,9 @@ class Gemv(Op):
x, y are vectors
x, y are vectors
alpha, beta are scalars
alpha, beta are scalars
output is a vector that can be inplace on y
output is a vector that can be inplace on y
"""
"""
__props__
=
(
"inplace"
,)
__props__
=
(
"inplace"
,)
def
__init__
(
self
,
inplace
):
def
__init__
(
self
,
inplace
):
...
@@ -443,12 +446,13 @@ class Ger(Op):
...
@@ -443,12 +446,13 @@ class Ger(Op):
for matrix A, scalar alpha, vectors x and y.
for matrix A, scalar alpha, vectors x and y.
This interface to GER allows non-destructive operation on A via the
This interface to GER allows non-destructive operation on A via the
`destructive`
`destructive` argument to the constructor.
argument to the constructor.
:TODO: Create better classes ScipyGer and CGer that inherit from this class
:TODO: Create better classes ScipyGer and CGer that inherit from this class
and override the make_thunk() method to use Scipy and C respectively.
and override the make_thunk() method to use Scipy and C respectively.
"""
"""
__props__
=
(
"destructive"
,)
__props__
=
(
"destructive"
,)
def
__init__
(
self
,
destructive
):
def
__init__
(
self
,
destructive
):
...
@@ -508,16 +512,22 @@ def ldflags(libs=True, flags=False, libs_dir=False, include_dir=False):
...
@@ -508,16 +512,22 @@ def ldflags(libs=True, flags=False, libs_dir=False, include_dir=False):
It returns a list of libraries against which an Op's object file
It returns a list of libraries against which an Op's object file
should be linked to benefit from a BLAS implementation.
should be linked to benefit from a BLAS implementation.
:type libs: bool, defaults to True
Parameters
:param libs: extract flags starting with "-l"
----------
:type libs_dir: bool, defaults to False
libs : bool, optional
:param libs_dir: extract flags starting with "-L"
Extract flags starting with "-l" (the default is True).
:type include_dir: bool, defaults to False
libs_dir : bool, optional
:param include_dir: extract flags starting with "-I"
Extract flags starting with "-L" (the default is False).
:type flags: bool, defaults to False
include_dir : bool, optional
:param flags: extract all the other flags
Extract flags starting with "-I" (the default is False).
:rtype: list of strings
flags: bool, optional
:returns: extracted flags
Extract all the other flags (the default is False).
Returns
-------
list of strings
Extracted flags.
"""
"""
ldflags_str
=
theano
.
config
.
blas
.
ldflags
ldflags_str
=
theano
.
config
.
blas
.
ldflags
return
_ldflags
(
ldflags_str
=
ldflags_str
,
return
_ldflags
(
ldflags_str
=
ldflags_str
,
...
@@ -533,19 +543,25 @@ def _ldflags(ldflags_str, libs, flags, libs_dir, include_dir):
...
@@ -533,19 +543,25 @@ def _ldflags(ldflags_str, libs, flags, libs_dir, include_dir):
Depending on the options, different type of flags will be kept.
Depending on the options, different type of flags will be kept.
:type ldflags_str: string
Parameters
:param ldflags_str: the string to process. Typically, this will
----------
be the content of `theano.config.blas.ldflags`
ldflags_str : string
:type libs: bool
The string to process. Typically, this will be the content of
:param libs: extract flags starting with "-l"
`theano.config.blas.ldflags`.
:type libs_dir: bool
libs : bool
:param libs_dir: extract flags starting with "-L"
Extract flags starting with "-l".
:type include_dir: bool
flags: bool
:param include_dir: extract flags starting with "-I"
Extract all the other flags.
:type flags: bool
libs_dir: bool
:param flags: extract all the other flags
Extract flags starting with "-L".
:rtype: list of strings
include_dir: bool
:returns: extracted flags
Extract flags starting with "-I".
Returns
-------
list of strings
Extracted flags.
"""
"""
rval
=
[]
rval
=
[]
if
libs_dir
:
if
libs_dir
:
...
@@ -598,10 +614,12 @@ def _ldflags(ldflags_str, libs, flags, libs_dir, include_dir):
...
@@ -598,10 +614,12 @@ def _ldflags(ldflags_str, libs, flags, libs_dir, include_dir):
class
GemmRelated
(
Op
):
class
GemmRelated
(
Op
):
"""Base class for Gemm and Dot22
"""Base class for Gemm and Dot22
.
This class provides a kind of templated gemm Op.
This class provides a kind of templated gemm Op.
"""
"""
__props__
=
()
__props__
=
()
def
c_support_code
(
self
):
def
c_support_code
(
self
):
...
@@ -915,7 +933,7 @@ class GemmRelated(Op):
...
@@ -915,7 +933,7 @@ class GemmRelated(Op):
class
Gemm
(
GemmRelated
):
class
Gemm
(
GemmRelated
):
"""In-place version of matrix-matrix multiplication (with accumulation)
:
"""In-place version of matrix-matrix multiplication (with accumulation)
.
When a and b are scalars and x, y, and z are matrices, then
When a and b are scalars and x, y, and z are matrices, then
...
@@ -936,6 +954,7 @@ class Gemm(GemmRelated):
...
@@ -936,6 +954,7 @@ class Gemm(GemmRelated):
optimized linear algebra operations.)
optimized linear algebra operations.)
"""
"""
E_rank
=
'gemm only works for rank 2'
E_rank
=
'gemm only works for rank 2'
E_scalar
=
'gemm requires scalar argument'
E_scalar
=
'gemm requires scalar argument'
E_z_uniq
=
'argument z aliased to x or y'
# TODO: justify / delete this
E_z_uniq
=
'argument z aliased to x or y'
# TODO: justify / delete this
...
@@ -1430,9 +1449,7 @@ def _factor_canonicalized(lst):
...
@@ -1430,9 +1449,7 @@ def _factor_canonicalized(lst):
def
_gemm_from_factored_list
(
lst
):
def
_gemm_from_factored_list
(
lst
):
"""Returns None, or a list to replace node.outputs
"""Returns None, or a list to replace node.outputs."""
"""
lst2
=
[]
lst2
=
[]
# Remove the tuple that can't be cast correctly.
# Remove the tuple that can't be cast correctly.
# This can happen when we try to cast a complex to a real
# This can happen when we try to cast a complex to a real
...
@@ -1524,7 +1541,7 @@ def _gemm_from_node2(node):
...
@@ -1524,7 +1541,7 @@ def _gemm_from_node2(node):
class
GemmOptimizer
(
Optimizer
):
class
GemmOptimizer
(
Optimizer
):
"""Graph optimizer for inserting Gemm operations"""
"""Graph optimizer for inserting Gemm operations
.
"""
def
__init__
(
self
):
def
__init__
(
self
):
Optimizer
.
__init__
(
self
)
Optimizer
.
__init__
(
self
)
self
.
warned
=
False
self
.
warned
=
False
...
@@ -1645,8 +1662,11 @@ class GemmOptimizer(Optimizer):
...
@@ -1645,8 +1662,11 @@ class GemmOptimizer(Optimizer):
class
Dot22
(
GemmRelated
):
class
Dot22
(
GemmRelated
):
"""Compute a matrix-matrix product.
"""Compute a matrix-matrix product.
This is a specialization of the more general Dot()
This is a specialization of the more general Dot().
"""
"""
def
make_node
(
self
,
x
,
y
):
def
make_node
(
self
,
x
,
y
):
dtypes
=
(
'float32'
,
'float64'
,
'complex64'
,
'complex128'
)
dtypes
=
(
'float32'
,
'float64'
,
'complex64'
,
'complex128'
)
if
x
.
type
.
ndim
!=
2
or
x
.
type
.
dtype
not
in
dtypes
:
if
x
.
type
.
ndim
!=
2
or
x
.
type
.
dtype
not
in
dtypes
:
...
@@ -1780,8 +1800,7 @@ def local_inplace_ger(node):
...
@@ -1780,8 +1800,7 @@ def local_inplace_ger(node):
@local_optimizer
([
gemm_no_inplace
])
@local_optimizer
([
gemm_no_inplace
])
def
local_gemm_to_gemv
(
node
):
def
local_gemm_to_gemv
(
node
):
"""GEMM acting on row or column matrices -> GEMV
"""GEMM acting on row or column matrices -> GEMV."""
"""
if
node
.
op
==
gemm_no_inplace
:
if
node
.
op
==
gemm_no_inplace
:
z
,
a
,
x
,
y
,
b
=
node
.
inputs
z
,
a
,
x
,
y
,
b
=
node
.
inputs
if
z
.
broadcastable
==
x
.
broadcastable
==
(
True
,
False
):
if
z
.
broadcastable
==
x
.
broadcastable
==
(
True
,
False
):
...
@@ -1794,8 +1813,7 @@ def local_gemm_to_gemv(node):
...
@@ -1794,8 +1813,7 @@ def local_gemm_to_gemv(node):
@local_optimizer
([
gemm_no_inplace
])
@local_optimizer
([
gemm_no_inplace
])
def
local_gemm_to_ger
(
node
):
def
local_gemm_to_ger
(
node
):
"""GEMM computing an outer-product -> GER
"""GEMM computing an outer-product -> GER."""
"""
if
node
.
op
==
gemm_no_inplace
:
if
node
.
op
==
gemm_no_inplace
:
z
,
a
,
x
,
y
,
b
=
node
.
inputs
z
,
a
,
x
,
y
,
b
=
node
.
inputs
if
x
.
broadcastable
[
1
]
and
y
.
broadcastable
[
0
]:
if
x
.
broadcastable
[
1
]
and
y
.
broadcastable
[
0
]:
...
@@ -1825,8 +1843,7 @@ def local_gemm_to_ger(node):
...
@@ -1825,8 +1843,7 @@ def local_gemm_to_ger(node):
# working
# working
@local_optimizer
([
_dot22
])
@local_optimizer
([
_dot22
])
def
local_dot22_to_ger_or_gemv
(
node
):
def
local_dot22_to_ger_or_gemv
(
node
):
"""dot22 computing an outer-product -> GER
"""dot22 computing an outer-product -> GER."""
"""
if
node
.
op
==
_dot22
:
if
node
.
op
==
_dot22
:
x
,
y
=
node
.
inputs
x
,
y
=
node
.
inputs
xb
=
x
.
broadcastable
xb
=
x
.
broadcastable
...
@@ -1904,11 +1921,14 @@ optdb.register('InplaceBlasOpt',
...
@@ -1904,11 +1921,14 @@ optdb.register('InplaceBlasOpt',
class
Dot22Scalar
(
GemmRelated
):
class
Dot22Scalar
(
GemmRelated
):
"""Compute a matrix-matrix product.
"""Compute a matrix-matrix product.
This is a specialization of the more general Dot()
This is a specialization of the more general Dot()
Used to call optimized gemm implementation.
Used to call optimized gemm implementation.
Also used to generate a gemm later.
Also used to generate a gemm later.
compute scalar*dot(x,y)
compute scalar*dot(x,y).
"""
"""
def
make_node
(
self
,
x
,
y
,
a
):
def
make_node
(
self
,
x
,
y
,
a
):
if
a
.
ndim
!=
0
:
if
a
.
ndim
!=
0
:
raise
TypeError
(
Gemm
.
E_scalar
,
a
)
raise
TypeError
(
Gemm
.
E_scalar
,
a
)
...
@@ -1996,25 +2016,27 @@ _dot22scalar = Dot22Scalar()
...
@@ -1996,25 +2016,27 @@ _dot22scalar = Dot22Scalar()
@local_optimizer
([
T
.
mul
])
@local_optimizer
([
T
.
mul
])
def
local_dot22_to_dot22scalar
(
node
):
def
local_dot22_to_dot22scalar
(
node
):
"""
"""
:note: Previous attempts to alter this optimization to replace dot22 with
Notes
gemm instead of dot22scalar resulted in some Scan nodes being
-----
duplicated and the ScanSaveMem optimization never running on them,
Previous attempts to alter this optimization to replace dot22 with
resulting in highly increased memory usage. Until this issue is
gemm instead of dot22scalar resulted in some Scan nodes being
resolved, this optimization should keep using dot22scalar instead of
duplicated and the ScanSaveMem optimization never running on them,
gemm.
resulting in highly increased memory usage. Until this issue is
resolved, this optimization should keep using dot22scalar instead of
:note: we upcast the scalar if after the multiplication with the
gemm.
dot this give the same type.
We upcast the scalar if after the multiplication with the dot this give
.. note: We execute this optimizer after the gemm optimizer. This
the same type.
allow to give more priority to gemm that give more speed up
then this optimizer, but allow the gemm optimizer to ignore
We execute this optimizer after the gemm optimizer. This
this op.
allow to give more priority to gemm that give more speed up
then this optimizer, but allow the gemm optimizer to ignore
this op.
TODO: support when we can reorder the mul to generate a
TODO: support when we can reorder the mul to generate a
dot22scalar or fix the canonizer to merge them(1 mul with multiple
dot22scalar or fix the canonizer to merge them(1 mul with multiple
inputs)
inputs)
"""
"""
if
node
.
op
!=
T
.
mul
:
if
node
.
op
!=
T
.
mul
:
return
False
return
False
...
@@ -2102,7 +2124,6 @@ def local_dot22_to_dot22scalar(node):
...
@@ -2102,7 +2124,6 @@ def local_dot22_to_dot22scalar(node):
return
[
T
.
mul
(
_dot22scalar
(
d
.
owner
.
inputs
[
0
],
return
[
T
.
mul
(
_dot22scalar
(
d
.
owner
.
inputs
[
0
],
d
.
owner
.
inputs
[
1
],
a
),
*
o
)]
d
.
owner
.
inputs
[
1
],
a
),
*
o
)]
# must happen after gemm as the gemm optimizer don't understant
# must happen after gemm as the gemm optimizer don't understant
# dot22scalar and gemm give more speed up then dot22scalar
# dot22scalar and gemm give more speed up then dot22scalar
blas_optdb
.
register
(
'local_dot22_to_dot22scalar'
,
blas_optdb
.
register
(
'local_dot22_to_dot22scalar'
,
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论