Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
2b73732a
提交
2b73732a
authored
7月 10, 2015
作者:
Iban Harlouchet
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
flake8 of theano/scalar/basic.py
上级
62c810f3
隐藏空白字符变更
内嵌
并排
正在显示
2 个修改的文件
包含
76 行增加
和
78 行删除
+76
-78
basic.py
theano/scalar/basic.py
+76
-77
test_flake8.py
theano/tests/test_flake8.py
+0
-1
没有找到文件。
theano/scalar/basic.py
浏览文件 @
2b73732a
...
...
@@ -242,22 +242,21 @@ class Scalar(Type):
print(dtype, np.zeros(1, dtype=dtype).dtype.num)
"""
return
{
# dtype: (py_type, c_type, cls_name)
'float16'
:
(
numpy
.
float16
,
'npy_float16'
,
'Float16'
),
'float32'
:
(
numpy
.
float32
,
'npy_float32'
,
'Float32'
),
'float64'
:
(
numpy
.
float64
,
'npy_float64'
,
'Float64'
),
'complex128'
:
(
numpy
.
complex128
,
'theano_complex128'
,
'Complex128'
),
'complex64'
:
(
numpy
.
complex64
,
'theano_complex64'
,
'Complex64'
),
'uint8'
:
(
numpy
.
uint8
,
'npy_uint8'
,
'UInt8'
),
'int8'
:
(
numpy
.
int8
,
'npy_int8'
,
'Int8'
),
'uint16'
:
(
numpy
.
uint16
,
'npy_uint16'
,
'UInt16'
),
'int16'
:
(
numpy
.
int16
,
'npy_int16'
,
'Int16'
),
'uint32'
:
(
numpy
.
uint32
,
'npy_uint32'
,
'UInt32'
),
'int32'
:
(
numpy
.
int32
,
'npy_int32'
,
'Int32'
),
'uint64'
:
(
numpy
.
uint64
,
'npy_uint64'
,
'UInt64'
),
'int64'
:
(
numpy
.
int64
,
'npy_int64'
,
'Int64'
)
}[
self
.
dtype
]
'float16'
:
(
numpy
.
float16
,
'npy_float16'
,
'Float16'
),
'float32'
:
(
numpy
.
float32
,
'npy_float32'
,
'Float32'
),
'float64'
:
(
numpy
.
float64
,
'npy_float64'
,
'Float64'
),
'complex128'
:
(
numpy
.
complex128
,
'theano_complex128'
,
'Complex128'
),
'complex64'
:
(
numpy
.
complex64
,
'theano_complex64'
,
'Complex64'
),
'uint8'
:
(
numpy
.
uint8
,
'npy_uint8'
,
'UInt8'
),
'int8'
:
(
numpy
.
int8
,
'npy_int8'
,
'Int8'
),
'uint16'
:
(
numpy
.
uint16
,
'npy_uint16'
,
'UInt16'
),
'int16'
:
(
numpy
.
int16
,
'npy_int16'
,
'Int16'
),
'uint32'
:
(
numpy
.
uint32
,
'npy_uint32'
,
'UInt32'
),
'int32'
:
(
numpy
.
int32
,
'npy_int32'
,
'Int32'
),
'uint64'
:
(
numpy
.
uint64
,
'npy_uint64'
,
'UInt64'
),
'int64'
:
(
numpy
.
int64
,
'npy_int64'
,
'Int64'
)
}[
self
.
dtype
]
except
KeyError
:
raise
TypeError
(
"Unsupported dtype for
%
s:
%
s"
%
(
self
.
__class__
.
__name__
,
self
.
dtype
))
...
...
@@ -348,7 +347,7 @@ class Scalar(Type):
# 'npy_intX', some C code may not compile, e.g. when assigning
# the value 0 (cast to 'int' in C) to a theano_complex64.
if
(
numpy
.
dtype
(
'intc'
)
.
num
not
in
[
numpy
.
dtype
(
d
[
4
:])
.
num
for
d
in
real_types
]):
[
numpy
.
dtype
(
d
[
4
:])
.
num
for
d
in
real_types
]):
# In that case we add the 'int' type to the real types.
real_types
.
append
(
'int'
)
...
...
@@ -424,9 +423,9 @@ class Scalar(Type):
operator_eq
=
''
.
join
(
operator_eq_real
(
ctype
,
rtype
)
for
ctype
in
cplx_types
for
rtype
in
real_types
)
\
+
''
.
join
(
operator_eq_cplx
(
ctype1
,
ctype2
)
for
ctype1
in
cplx_types
for
ctype2
in
cplx_types
)
+
''
.
join
(
operator_eq_cplx
(
ctype1
,
ctype2
)
for
ctype1
in
cplx_types
for
ctype2
in
cplx_types
)
# We are not using C++ generic templating here, because this would
# generate two different functions for adding a complex64 and a
...
...
@@ -474,11 +473,11 @@ class Scalar(Type):
for
rtype
in
real_types
)
return
template
%
dict
(
nbits
=
64
,
half_nbits
=
32
)
\
+
template
%
dict
(
nbits
=
128
,
half_nbits
=
64
)
\
+
operator_eq
\
+
operator_plus
\
+
operator_minus
\
+
operator_mul
+
template
%
dict
(
nbits
=
128
,
half_nbits
=
64
)
\
+
operator_eq
\
+
operator_plus
\
+
operator_minus
\
+
operator_mul
else
:
return
""
...
...
@@ -544,9 +543,9 @@ class _scalar_py_operators:
return
neg
(
self
)
# CASTS
#def __int__(self): return AsInt(self).out
#def __float__(self): return AsDouble(self).out
#def __complex__(self): return AsComplex(self).out
#
def __int__(self): return AsInt(self).out
#
def __float__(self): return AsDouble(self).out
#
def __complex__(self): return AsComplex(self).out
# BITWISE
def
__invert__
(
self
):
...
...
@@ -583,7 +582,7 @@ class _scalar_py_operators:
def
__ge__
(
self
,
other
):
return
ge
(
self
,
other
)
#ARITHMETIC - NORMAL
#
ARITHMETIC - NORMAL
def
__add__
(
self
,
other
):
return
add
(
self
,
other
)
...
...
@@ -609,7 +608,7 @@ class _scalar_py_operators:
def
__pow__
(
self
,
other
):
return
pow
(
self
,
other
)
#ARITHMETIC - RIGHT-OPERAND
#
ARITHMETIC - RIGHT-OPERAND
def
__radd__
(
self
,
other
):
return
add
(
other
,
self
)
...
...
@@ -694,7 +693,7 @@ class upgrade_to_float(object):
uint32
:
float64
,
uint64
:
float64
}
return
get_scalar_type
(
Scalar
.
upcast
(
*
[
conv
.
get
(
type
,
type
)
for
type
in
types
])),
for
type
in
types
])),
class
same_out
(
object
):
...
...
@@ -942,9 +941,9 @@ class UnaryScalarOp(ScalarOp):
(
x
,)
=
inputs
(
z
,)
=
outputs
if
(
not
theano
.
config
.
lib
.
amdlibm
or
# We compare the dtype AND the broadcast flag
# as this function do not broadcast
node
.
inputs
[
0
]
.
type
!=
node
.
outputs
[
0
]
.
type
):
# We compare the dtype AND the broadcast flag
# as this function do not broadcast
node
.
inputs
[
0
]
.
type
!=
node
.
outputs
[
0
]
.
type
):
raise
theano
.
gof
.
utils
.
MethodNotDefined
()
dtype
=
node
.
inputs
[
0
]
.
type
.
dtype_specs
()[
1
]
...
...
@@ -1176,7 +1175,7 @@ class InRange(LogicalComparison):
cmp1
=
'>='
# backport
#cmp1 = '>' if self.openlow else '>='
#
cmp1 = '>' if self.openlow else '>='
if
self
.
openhi
:
cmp2
=
'<'
...
...
@@ -1184,7 +1183,7 @@ class InRange(LogicalComparison):
cmp2
=
'<='
# backport
#cmp2 = '<' if self.openhi else '<='
#
cmp2 = '<' if self.openhi else '<='
return
(
"
%(z)
s =
%(x)
s
%(cmp1)
s
%(low)
s &&"
"
%(x)
s
%(cmp2)
s
%(hi)
s;"
%
locals
())
...
...
@@ -1473,7 +1472,7 @@ class Mul(ScalarOp):
# output is complex. The rest of this function make this supposition.
output_type
=
self
.
output_types
([
i
.
type
for
i
in
inputs
])[
0
]
if
output_type
in
complex_types
:
if
not
gz
.
type
in
complex_types
:
if
gz
.
type
not
in
complex_types
:
raise
TypeError
(
'Mul with output_type '
+
str
(
output_type
)
+
' expected gz type to be complex, got gz with type '
+
...
...
@@ -1600,7 +1599,7 @@ class TrueDiv(BinaryScalarOp):
node
.
inputs
[
1
]
.
type
in
complex_types
])
==
1
:
raise
NotImplementedError
(
'type not supported'
,
type
)
if
(
node
.
inputs
[
0
]
.
type
in
discrete_types
and
node
.
inputs
[
1
]
.
type
in
discrete_types
):
node
.
inputs
[
1
]
.
type
in
discrete_types
):
return
"
%(z)
s = ((double)
%(x)
s) /
%(y)
s;"
%
locals
()
return
"
%(z)
s =
%(x)
s /
%(y)
s;"
%
locals
()
...
...
@@ -1710,7 +1709,7 @@ floor_div = int_div
def
mod_check
(
x
,
y
):
if
(
as_scalar
(
x
)
.
type
in
complex_types
or
as_scalar
(
y
)
.
type
in
complex_types
):
as_scalar
(
y
)
.
type
in
complex_types
):
# Currently forbidden.
raise
Mod
.
complex_error
else
:
...
...
@@ -1808,7 +1807,7 @@ class Pow(BinaryScalarOp):
(
x
,
y
)
=
inputs
(
z
,)
=
outputs
if
(
node
.
inputs
[
0
]
.
type
in
complex_types
or
node
.
inputs
[
1
]
.
type
in
complex_types
):
node
.
inputs
[
1
]
.
type
in
complex_types
):
raise
NotImplementedError
(
'type not supported'
,
type
)
return
"
%(z)
s = pow(
%(x)
s,
%(y)
s);"
%
locals
()
...
...
@@ -1838,10 +1837,10 @@ class Pow(BinaryScalarOp):
# We compare the dtype AND the broadcast flag
# as this function do not broadcast
if
(
node
.
inputs
[
0
]
.
type
==
node
.
outputs
[
0
]
.
type
and
node
.
inputs
[
1
]
.
type
==
node
.
outputs
[
0
]
.
type
and
# amdlibm 3.0 do not have a float64 version of this SIMD function
node
.
inputs
[
0
]
.
dtype
==
'float32'
and
node
.
inputs
[
1
]
.
dtype
==
'float32'
):
node
.
inputs
[
1
]
.
type
==
node
.
outputs
[
0
]
.
type
and
# amdlibm 3.0 do not have a float64 version of this SIMD function
node
.
inputs
[
0
]
.
dtype
==
'float32'
and
node
.
inputs
[
1
]
.
dtype
==
'float32'
):
dtype
=
'float'
fct
=
"amd_vrsa_powf"
return
"""
...
...
@@ -2014,19 +2013,19 @@ convert_to_complex64 = Cast(complex64, name='convert_to_complex64')
convert_to_complex128
=
Cast
(
complex128
,
name
=
'convert_to_complex128'
)
_cast_mapping
=
{
'int8'
:
convert_to_int8
,
'int16'
:
convert_to_int16
,
'int32'
:
convert_to_int32
,
'int64'
:
convert_to_int64
,
'uint8'
:
convert_to_uint8
,
'uint16'
:
convert_to_uint16
,
'uint32'
:
convert_to_uint32
,
'uint64'
:
convert_to_uint64
,
'float16'
:
convert_to_float16
,
'float32'
:
convert_to_float32
,
'float64'
:
convert_to_float64
,
'complex64'
:
convert_to_complex64
,
'complex128'
:
convert_to_complex128
}
'int8'
:
convert_to_int8
,
'int16'
:
convert_to_int16
,
'int32'
:
convert_to_int32
,
'int64'
:
convert_to_int64
,
'uint8'
:
convert_to_uint8
,
'uint16'
:
convert_to_uint16
,
'uint32'
:
convert_to_uint32
,
'uint64'
:
convert_to_uint64
,
'float16'
:
convert_to_float16
,
'float32'
:
convert_to_float32
,
'float64'
:
convert_to_float64
,
'complex64'
:
convert_to_complex64
,
'complex128'
:
convert_to_complex128
}
def
cast
(
x
,
dtype
):
...
...
@@ -2201,7 +2200,7 @@ class RoundHalfToEven(UnaryScalarOp):
(
x
,)
=
inputs
(
z
,)
=
outputs
typ
=
node
.
outputs
[
0
]
.
type
.
dtype
if
not
typ
in
[
'float32'
,
'float64'
]:
if
typ
not
in
[
'float32'
,
'float64'
]:
Exception
(
"The output should be float32 or float64"
)
return
dedent
(
"""
...
...
@@ -2946,7 +2945,7 @@ class ArcTan2(BinaryScalarOp):
(
y
,
x
)
=
inputs
(
z
,)
=
outputs
if
(
node
.
inputs
[
0
]
.
type
in
complex_types
or
node
.
inputs
[
1
]
.
type
in
complex_types
):
node
.
inputs
[
1
]
.
type
in
complex_types
):
raise
NotImplementedError
(
'type not supported'
,
type
)
return
"
%(z)
s = atan2(
%(y)
s,
%(x)
s);"
%
locals
()
arctan2
=
ArcTan2
(
upgrade_to_float
,
name
=
'arctan2'
)
...
...
@@ -3309,7 +3308,7 @@ class Composite(ScalarOp):
"All orphans in the fgraph to Composite must"
" be Constant instances."
)
elif
(
any
(
i
.
dtype
==
'float16'
for
i
in
var
.
owner
.
inputs
)
or
any
(
o
.
dtype
==
'float16'
for
o
in
var
.
owner
.
outputs
)):
any
(
o
.
dtype
==
'float16'
for
o
in
var
.
owner
.
outputs
)):
# flag for elemwise ops to check.
self
.
inner_float16
=
True
...
...
@@ -3325,13 +3324,13 @@ class Composite(ScalarOp):
name
=
"V
%%(id)
s_tmp
%
i"
%
i
subd
[
output
]
=
name
_c_code
+=
"
%
s
%
s;
\n
"
%
(
output
.
type
.
dtype_specs
()[
1
],
name
)
s
=
node
.
op
.
c_code
(
node
,
self
.
nodenames
[
j
]
,
[
subd
[
input
]
for
input
in
node
.
inputs
],
[
subd
[
output
]
for
output
in
node
.
out
puts
],
dict
(
fail
=
"
%(fail)
s"
,
id
=
"
%%(id)
s_
%
i"
%
j
))
output
.
type
.
dtype_specs
()[
1
],
name
)
s
=
node
.
op
.
c_code
(
node
,
self
.
nodenames
[
j
],
[
subd
[
input
]
for
input
in
node
.
in
puts
],
[
subd
[
output
]
for
output
in
node
.
outputs
]
,
dict
(
fail
=
"
%(fail)
s"
,
id
=
"
%%(id)
s_
%
i"
%
j
))
_c_code
+=
s
_c_code
+=
"
\n
"
_c_code
+=
"}
\n
"
...
...
@@ -3454,7 +3453,7 @@ class Composite(ScalarOp):
def
make_node
(
self
,
*
inputs
):
if
(
tuple
([
i
.
type
for
i
in
self
.
inputs
])
==
tuple
([
i
.
type
for
i
in
inputs
])):
tuple
([
i
.
type
for
i
in
inputs
])):
return
super
(
Composite
,
self
)
.
make_node
(
*
inputs
)
else
:
# Make a new op with the right input type.
...
...
@@ -3489,7 +3488,7 @@ class Composite(ScalarOp):
izip
((
"o
%
i"
%
i
for
i
in
xrange
(
len
(
onames
))),
onames
)),
**
sub
)
d
[
'nodename'
]
=
nodename
if
not
'id'
in
sub
:
if
'id'
not
in
sub
:
# The use of a dummy id is safe as the code is in a separate block.
# It won't generate conflicting variable name.
d
[
'id'
]
=
'_DUMMY_ID_'
...
...
@@ -3521,8 +3520,8 @@ class Composite(ScalarOp):
for
subnode
,
subnodename
in
zip
(
self
.
fgraph
.
toposort
(),
self
.
nodenames
):
try
:
subnode_support_code
=
subnode
.
op
.
c_support_code_apply
(
subnode
,
subnodename
%
dict
(
nodename
=
name
))
subnode
,
subnodename
%
dict
(
nodename
=
name
))
if
subnode_support_code
:
rval
.
append
(
subnode_support_code
)
except
gof
.
utils
.
MethodNotDefined
:
...
...
@@ -3536,9 +3535,9 @@ class Composite(ScalarOp):
def
__eq__
(
self
,
other
):
if
self
is
other
:
return
True
if
(
type
(
self
)
!=
type
(
other
)
or
self
.
nin
!=
other
.
nin
or
self
.
nout
!=
other
.
nout
):
if
(
type
(
self
)
!=
type
(
other
)
or
self
.
nin
!=
other
.
nin
or
self
.
nout
!=
other
.
nout
):
return
False
# see __hash__ for comment on why there is no mention of fgraph
# or module cache key here.
...
...
@@ -3546,9 +3545,9 @@ class Composite(ScalarOp):
def
__hash__
(
self
):
rval
=
hash
((
type
(
self
),
self
.
nin
,
self
.
nout
,
self
.
_c_code
))
self
.
nin
,
self
.
nout
,
self
.
_c_code
))
# Note that in general, the configparser settings at the time
# of code generation (__init__) affect the semantics of this Op.
# This function assumes that all relevant info about the configparser
...
...
theano/tests/test_flake8.py
浏览文件 @
2b73732a
...
...
@@ -116,7 +116,6 @@ whitelist_flake8 = [
"tensor/nnet/tests/test_sigm.py"
,
"scalar/basic_sympy.py"
,
"scalar/__init__.py"
,
"scalar/basic.py"
,
"scalar/tests/test_basic.py"
,
"sandbox/test_theano_object.py"
,
"sandbox/test_scan.py"
,
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论