Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
2b73732a
提交
2b73732a
authored
7月 10, 2015
作者:
Iban Harlouchet
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
flake8 of theano/scalar/basic.py
上级
62c810f3
隐藏空白字符变更
内嵌
并排
正在显示
2 个修改的文件
包含
76 行增加
和
78 行删除
+76
-78
basic.py
theano/scalar/basic.py
+76
-77
test_flake8.py
theano/tests/test_flake8.py
+0
-1
没有找到文件。
theano/scalar/basic.py
浏览文件 @
2b73732a
...
@@ -242,22 +242,21 @@ class Scalar(Type):
...
@@ -242,22 +242,21 @@ class Scalar(Type):
print(dtype, np.zeros(1, dtype=dtype).dtype.num)
print(dtype, np.zeros(1, dtype=dtype).dtype.num)
"""
"""
return
{
# dtype: (py_type, c_type, cls_name)
return
{
# dtype: (py_type, c_type, cls_name)
'float16'
:
(
numpy
.
float16
,
'npy_float16'
,
'Float16'
),
'float16'
:
(
numpy
.
float16
,
'npy_float16'
,
'Float16'
),
'float32'
:
(
numpy
.
float32
,
'npy_float32'
,
'Float32'
),
'float32'
:
(
numpy
.
float32
,
'npy_float32'
,
'Float32'
),
'float64'
:
(
numpy
.
float64
,
'npy_float64'
,
'Float64'
),
'float64'
:
(
numpy
.
float64
,
'npy_float64'
,
'Float64'
),
'complex128'
:
(
numpy
.
complex128
,
'theano_complex128'
,
'complex128'
:
(
numpy
.
complex128
,
'theano_complex128'
,
'Complex128'
),
'Complex128'
),
'complex64'
:
(
numpy
.
complex64
,
'theano_complex64'
,
'complex64'
:
(
numpy
.
complex64
,
'theano_complex64'
,
'Complex64'
),
'Complex64'
),
'uint8'
:
(
numpy
.
uint8
,
'npy_uint8'
,
'UInt8'
),
'uint8'
:
(
numpy
.
uint8
,
'npy_uint8'
,
'UInt8'
),
'int8'
:
(
numpy
.
int8
,
'npy_int8'
,
'Int8'
),
'int8'
:
(
numpy
.
int8
,
'npy_int8'
,
'Int8'
),
'uint16'
:
(
numpy
.
uint16
,
'npy_uint16'
,
'UInt16'
),
'uint16'
:
(
numpy
.
uint16
,
'npy_uint16'
,
'UInt16'
),
'int16'
:
(
numpy
.
int16
,
'npy_int16'
,
'Int16'
),
'int16'
:
(
numpy
.
int16
,
'npy_int16'
,
'Int16'
),
'uint32'
:
(
numpy
.
uint32
,
'npy_uint32'
,
'UInt32'
),
'uint32'
:
(
numpy
.
uint32
,
'npy_uint32'
,
'UInt32'
),
'int32'
:
(
numpy
.
int32
,
'npy_int32'
,
'Int32'
),
'int32'
:
(
numpy
.
int32
,
'npy_int32'
,
'Int32'
),
'uint64'
:
(
numpy
.
uint64
,
'npy_uint64'
,
'UInt64'
),
'uint64'
:
(
numpy
.
uint64
,
'npy_uint64'
,
'UInt64'
),
'int64'
:
(
numpy
.
int64
,
'npy_int64'
,
'Int64'
)
'int64'
:
(
numpy
.
int64
,
'npy_int64'
,
'Int64'
)
}[
self
.
dtype
]
}[
self
.
dtype
]
except
KeyError
:
except
KeyError
:
raise
TypeError
(
"Unsupported dtype for
%
s:
%
s"
%
(
raise
TypeError
(
"Unsupported dtype for
%
s:
%
s"
%
(
self
.
__class__
.
__name__
,
self
.
dtype
))
self
.
__class__
.
__name__
,
self
.
dtype
))
...
@@ -348,7 +347,7 @@ class Scalar(Type):
...
@@ -348,7 +347,7 @@ class Scalar(Type):
# 'npy_intX', some C code may not compile, e.g. when assigning
# 'npy_intX', some C code may not compile, e.g. when assigning
# the value 0 (cast to 'int' in C) to a theano_complex64.
# the value 0 (cast to 'int' in C) to a theano_complex64.
if
(
numpy
.
dtype
(
'intc'
)
.
num
not
in
if
(
numpy
.
dtype
(
'intc'
)
.
num
not
in
[
numpy
.
dtype
(
d
[
4
:])
.
num
for
d
in
real_types
]):
[
numpy
.
dtype
(
d
[
4
:])
.
num
for
d
in
real_types
]):
# In that case we add the 'int' type to the real types.
# In that case we add the 'int' type to the real types.
real_types
.
append
(
'int'
)
real_types
.
append
(
'int'
)
...
@@ -424,9 +423,9 @@ class Scalar(Type):
...
@@ -424,9 +423,9 @@ class Scalar(Type):
operator_eq
=
''
.
join
(
operator_eq_real
(
ctype
,
rtype
)
operator_eq
=
''
.
join
(
operator_eq_real
(
ctype
,
rtype
)
for
ctype
in
cplx_types
for
ctype
in
cplx_types
for
rtype
in
real_types
)
\
for
rtype
in
real_types
)
\
+
''
.
join
(
operator_eq_cplx
(
ctype1
,
ctype2
)
+
''
.
join
(
operator_eq_cplx
(
ctype1
,
ctype2
)
for
ctype1
in
cplx_types
for
ctype1
in
cplx_types
for
ctype2
in
cplx_types
)
for
ctype2
in
cplx_types
)
# We are not using C++ generic templating here, because this would
# We are not using C++ generic templating here, because this would
# generate two different functions for adding a complex64 and a
# generate two different functions for adding a complex64 and a
...
@@ -474,11 +473,11 @@ class Scalar(Type):
...
@@ -474,11 +473,11 @@ class Scalar(Type):
for
rtype
in
real_types
)
for
rtype
in
real_types
)
return
template
%
dict
(
nbits
=
64
,
half_nbits
=
32
)
\
return
template
%
dict
(
nbits
=
64
,
half_nbits
=
32
)
\
+
template
%
dict
(
nbits
=
128
,
half_nbits
=
64
)
\
+
template
%
dict
(
nbits
=
128
,
half_nbits
=
64
)
\
+
operator_eq
\
+
operator_eq
\
+
operator_plus
\
+
operator_plus
\
+
operator_minus
\
+
operator_minus
\
+
operator_mul
+
operator_mul
else
:
else
:
return
""
return
""
...
@@ -544,9 +543,9 @@ class _scalar_py_operators:
...
@@ -544,9 +543,9 @@ class _scalar_py_operators:
return
neg
(
self
)
return
neg
(
self
)
# CASTS
# CASTS
#def __int__(self): return AsInt(self).out
#
def __int__(self): return AsInt(self).out
#def __float__(self): return AsDouble(self).out
#
def __float__(self): return AsDouble(self).out
#def __complex__(self): return AsComplex(self).out
#
def __complex__(self): return AsComplex(self).out
# BITWISE
# BITWISE
def
__invert__
(
self
):
def
__invert__
(
self
):
...
@@ -583,7 +582,7 @@ class _scalar_py_operators:
...
@@ -583,7 +582,7 @@ class _scalar_py_operators:
def
__ge__
(
self
,
other
):
def
__ge__
(
self
,
other
):
return
ge
(
self
,
other
)
return
ge
(
self
,
other
)
#ARITHMETIC - NORMAL
#
ARITHMETIC - NORMAL
def
__add__
(
self
,
other
):
def
__add__
(
self
,
other
):
return
add
(
self
,
other
)
return
add
(
self
,
other
)
...
@@ -609,7 +608,7 @@ class _scalar_py_operators:
...
@@ -609,7 +608,7 @@ class _scalar_py_operators:
def
__pow__
(
self
,
other
):
def
__pow__
(
self
,
other
):
return
pow
(
self
,
other
)
return
pow
(
self
,
other
)
#ARITHMETIC - RIGHT-OPERAND
#
ARITHMETIC - RIGHT-OPERAND
def
__radd__
(
self
,
other
):
def
__radd__
(
self
,
other
):
return
add
(
other
,
self
)
return
add
(
other
,
self
)
...
@@ -694,7 +693,7 @@ class upgrade_to_float(object):
...
@@ -694,7 +693,7 @@ class upgrade_to_float(object):
uint32
:
float64
,
uint32
:
float64
,
uint64
:
float64
}
uint64
:
float64
}
return
get_scalar_type
(
Scalar
.
upcast
(
*
[
conv
.
get
(
type
,
type
)
return
get_scalar_type
(
Scalar
.
upcast
(
*
[
conv
.
get
(
type
,
type
)
for
type
in
types
])),
for
type
in
types
])),
class
same_out
(
object
):
class
same_out
(
object
):
...
@@ -942,9 +941,9 @@ class UnaryScalarOp(ScalarOp):
...
@@ -942,9 +941,9 @@ class UnaryScalarOp(ScalarOp):
(
x
,)
=
inputs
(
x
,)
=
inputs
(
z
,)
=
outputs
(
z
,)
=
outputs
if
(
not
theano
.
config
.
lib
.
amdlibm
or
if
(
not
theano
.
config
.
lib
.
amdlibm
or
# We compare the dtype AND the broadcast flag
# We compare the dtype AND the broadcast flag
# as this function do not broadcast
# as this function do not broadcast
node
.
inputs
[
0
]
.
type
!=
node
.
outputs
[
0
]
.
type
):
node
.
inputs
[
0
]
.
type
!=
node
.
outputs
[
0
]
.
type
):
raise
theano
.
gof
.
utils
.
MethodNotDefined
()
raise
theano
.
gof
.
utils
.
MethodNotDefined
()
dtype
=
node
.
inputs
[
0
]
.
type
.
dtype_specs
()[
1
]
dtype
=
node
.
inputs
[
0
]
.
type
.
dtype_specs
()[
1
]
...
@@ -1176,7 +1175,7 @@ class InRange(LogicalComparison):
...
@@ -1176,7 +1175,7 @@ class InRange(LogicalComparison):
cmp1
=
'>='
cmp1
=
'>='
# backport
# backport
#cmp1 = '>' if self.openlow else '>='
#
cmp1 = '>' if self.openlow else '>='
if
self
.
openhi
:
if
self
.
openhi
:
cmp2
=
'<'
cmp2
=
'<'
...
@@ -1184,7 +1183,7 @@ class InRange(LogicalComparison):
...
@@ -1184,7 +1183,7 @@ class InRange(LogicalComparison):
cmp2
=
'<='
cmp2
=
'<='
# backport
# backport
#cmp2 = '<' if self.openhi else '<='
#
cmp2 = '<' if self.openhi else '<='
return
(
"
%(z)
s =
%(x)
s
%(cmp1)
s
%(low)
s &&"
return
(
"
%(z)
s =
%(x)
s
%(cmp1)
s
%(low)
s &&"
"
%(x)
s
%(cmp2)
s
%(hi)
s;"
%
locals
())
"
%(x)
s
%(cmp2)
s
%(hi)
s;"
%
locals
())
...
@@ -1473,7 +1472,7 @@ class Mul(ScalarOp):
...
@@ -1473,7 +1472,7 @@ class Mul(ScalarOp):
# output is complex. The rest of this function make this supposition.
# output is complex. The rest of this function make this supposition.
output_type
=
self
.
output_types
([
i
.
type
for
i
in
inputs
])[
0
]
output_type
=
self
.
output_types
([
i
.
type
for
i
in
inputs
])[
0
]
if
output_type
in
complex_types
:
if
output_type
in
complex_types
:
if
not
gz
.
type
in
complex_types
:
if
gz
.
type
not
in
complex_types
:
raise
TypeError
(
raise
TypeError
(
'Mul with output_type '
+
str
(
output_type
)
+
'Mul with output_type '
+
str
(
output_type
)
+
' expected gz type to be complex, got gz with type '
+
' expected gz type to be complex, got gz with type '
+
...
@@ -1600,7 +1599,7 @@ class TrueDiv(BinaryScalarOp):
...
@@ -1600,7 +1599,7 @@ class TrueDiv(BinaryScalarOp):
node
.
inputs
[
1
]
.
type
in
complex_types
])
==
1
:
node
.
inputs
[
1
]
.
type
in
complex_types
])
==
1
:
raise
NotImplementedError
(
'type not supported'
,
type
)
raise
NotImplementedError
(
'type not supported'
,
type
)
if
(
node
.
inputs
[
0
]
.
type
in
discrete_types
and
if
(
node
.
inputs
[
0
]
.
type
in
discrete_types
and
node
.
inputs
[
1
]
.
type
in
discrete_types
):
node
.
inputs
[
1
]
.
type
in
discrete_types
):
return
"
%(z)
s = ((double)
%(x)
s) /
%(y)
s;"
%
locals
()
return
"
%(z)
s = ((double)
%(x)
s) /
%(y)
s;"
%
locals
()
return
"
%(z)
s =
%(x)
s /
%(y)
s;"
%
locals
()
return
"
%(z)
s =
%(x)
s /
%(y)
s;"
%
locals
()
...
@@ -1710,7 +1709,7 @@ floor_div = int_div
...
@@ -1710,7 +1709,7 @@ floor_div = int_div
def
mod_check
(
x
,
y
):
def
mod_check
(
x
,
y
):
if
(
as_scalar
(
x
)
.
type
in
complex_types
or
if
(
as_scalar
(
x
)
.
type
in
complex_types
or
as_scalar
(
y
)
.
type
in
complex_types
):
as_scalar
(
y
)
.
type
in
complex_types
):
# Currently forbidden.
# Currently forbidden.
raise
Mod
.
complex_error
raise
Mod
.
complex_error
else
:
else
:
...
@@ -1808,7 +1807,7 @@ class Pow(BinaryScalarOp):
...
@@ -1808,7 +1807,7 @@ class Pow(BinaryScalarOp):
(
x
,
y
)
=
inputs
(
x
,
y
)
=
inputs
(
z
,)
=
outputs
(
z
,)
=
outputs
if
(
node
.
inputs
[
0
]
.
type
in
complex_types
or
if
(
node
.
inputs
[
0
]
.
type
in
complex_types
or
node
.
inputs
[
1
]
.
type
in
complex_types
):
node
.
inputs
[
1
]
.
type
in
complex_types
):
raise
NotImplementedError
(
'type not supported'
,
type
)
raise
NotImplementedError
(
'type not supported'
,
type
)
return
"
%(z)
s = pow(
%(x)
s,
%(y)
s);"
%
locals
()
return
"
%(z)
s = pow(
%(x)
s,
%(y)
s);"
%
locals
()
...
@@ -1838,10 +1837,10 @@ class Pow(BinaryScalarOp):
...
@@ -1838,10 +1837,10 @@ class Pow(BinaryScalarOp):
# We compare the dtype AND the broadcast flag
# We compare the dtype AND the broadcast flag
# as this function do not broadcast
# as this function do not broadcast
if
(
node
.
inputs
[
0
]
.
type
==
node
.
outputs
[
0
]
.
type
and
if
(
node
.
inputs
[
0
]
.
type
==
node
.
outputs
[
0
]
.
type
and
node
.
inputs
[
1
]
.
type
==
node
.
outputs
[
0
]
.
type
and
node
.
inputs
[
1
]
.
type
==
node
.
outputs
[
0
]
.
type
and
# amdlibm 3.0 do not have a float64 version of this SIMD function
# amdlibm 3.0 do not have a float64 version of this SIMD function
node
.
inputs
[
0
]
.
dtype
==
'float32'
and
node
.
inputs
[
0
]
.
dtype
==
'float32'
and
node
.
inputs
[
1
]
.
dtype
==
'float32'
):
node
.
inputs
[
1
]
.
dtype
==
'float32'
):
dtype
=
'float'
dtype
=
'float'
fct
=
"amd_vrsa_powf"
fct
=
"amd_vrsa_powf"
return
"""
return
"""
...
@@ -2014,19 +2013,19 @@ convert_to_complex64 = Cast(complex64, name='convert_to_complex64')
...
@@ -2014,19 +2013,19 @@ convert_to_complex64 = Cast(complex64, name='convert_to_complex64')
convert_to_complex128
=
Cast
(
complex128
,
name
=
'convert_to_complex128'
)
convert_to_complex128
=
Cast
(
complex128
,
name
=
'convert_to_complex128'
)
_cast_mapping
=
{
_cast_mapping
=
{
'int8'
:
convert_to_int8
,
'int8'
:
convert_to_int8
,
'int16'
:
convert_to_int16
,
'int16'
:
convert_to_int16
,
'int32'
:
convert_to_int32
,
'int32'
:
convert_to_int32
,
'int64'
:
convert_to_int64
,
'int64'
:
convert_to_int64
,
'uint8'
:
convert_to_uint8
,
'uint8'
:
convert_to_uint8
,
'uint16'
:
convert_to_uint16
,
'uint16'
:
convert_to_uint16
,
'uint32'
:
convert_to_uint32
,
'uint32'
:
convert_to_uint32
,
'uint64'
:
convert_to_uint64
,
'uint64'
:
convert_to_uint64
,
'float16'
:
convert_to_float16
,
'float16'
:
convert_to_float16
,
'float32'
:
convert_to_float32
,
'float32'
:
convert_to_float32
,
'float64'
:
convert_to_float64
,
'float64'
:
convert_to_float64
,
'complex64'
:
convert_to_complex64
,
'complex64'
:
convert_to_complex64
,
'complex128'
:
convert_to_complex128
}
'complex128'
:
convert_to_complex128
}
def
cast
(
x
,
dtype
):
def
cast
(
x
,
dtype
):
...
@@ -2201,7 +2200,7 @@ class RoundHalfToEven(UnaryScalarOp):
...
@@ -2201,7 +2200,7 @@ class RoundHalfToEven(UnaryScalarOp):
(
x
,)
=
inputs
(
x
,)
=
inputs
(
z
,)
=
outputs
(
z
,)
=
outputs
typ
=
node
.
outputs
[
0
]
.
type
.
dtype
typ
=
node
.
outputs
[
0
]
.
type
.
dtype
if
not
typ
in
[
'float32'
,
'float64'
]:
if
typ
not
in
[
'float32'
,
'float64'
]:
Exception
(
"The output should be float32 or float64"
)
Exception
(
"The output should be float32 or float64"
)
return
dedent
(
"""
return
dedent
(
"""
...
@@ -2946,7 +2945,7 @@ class ArcTan2(BinaryScalarOp):
...
@@ -2946,7 +2945,7 @@ class ArcTan2(BinaryScalarOp):
(
y
,
x
)
=
inputs
(
y
,
x
)
=
inputs
(
z
,)
=
outputs
(
z
,)
=
outputs
if
(
node
.
inputs
[
0
]
.
type
in
complex_types
or
if
(
node
.
inputs
[
0
]
.
type
in
complex_types
or
node
.
inputs
[
1
]
.
type
in
complex_types
):
node
.
inputs
[
1
]
.
type
in
complex_types
):
raise
NotImplementedError
(
'type not supported'
,
type
)
raise
NotImplementedError
(
'type not supported'
,
type
)
return
"
%(z)
s = atan2(
%(y)
s,
%(x)
s);"
%
locals
()
return
"
%(z)
s = atan2(
%(y)
s,
%(x)
s);"
%
locals
()
arctan2
=
ArcTan2
(
upgrade_to_float
,
name
=
'arctan2'
)
arctan2
=
ArcTan2
(
upgrade_to_float
,
name
=
'arctan2'
)
...
@@ -3309,7 +3308,7 @@ class Composite(ScalarOp):
...
@@ -3309,7 +3308,7 @@ class Composite(ScalarOp):
"All orphans in the fgraph to Composite must"
"All orphans in the fgraph to Composite must"
" be Constant instances."
)
" be Constant instances."
)
elif
(
any
(
i
.
dtype
==
'float16'
for
i
in
var
.
owner
.
inputs
)
or
elif
(
any
(
i
.
dtype
==
'float16'
for
i
in
var
.
owner
.
inputs
)
or
any
(
o
.
dtype
==
'float16'
for
o
in
var
.
owner
.
outputs
)):
any
(
o
.
dtype
==
'float16'
for
o
in
var
.
owner
.
outputs
)):
# flag for elemwise ops to check.
# flag for elemwise ops to check.
self
.
inner_float16
=
True
self
.
inner_float16
=
True
...
@@ -3325,13 +3324,13 @@ class Composite(ScalarOp):
...
@@ -3325,13 +3324,13 @@ class Composite(ScalarOp):
name
=
"V
%%(id)
s_tmp
%
i"
%
i
name
=
"V
%%(id)
s_tmp
%
i"
%
i
subd
[
output
]
=
name
subd
[
output
]
=
name
_c_code
+=
"
%
s
%
s;
\n
"
%
(
_c_code
+=
"
%
s
%
s;
\n
"
%
(
output
.
type
.
dtype_specs
()[
1
],
name
)
output
.
type
.
dtype_specs
()[
1
],
name
)
s
=
node
.
op
.
c_code
(
node
,
s
=
node
.
op
.
c_code
(
self
.
nodenames
[
j
]
,
node
,
[
subd
[
input
]
for
input
in
node
.
inputs
],
self
.
nodenames
[
j
],
[
subd
[
output
]
for
output
in
node
.
out
puts
],
[
subd
[
input
]
for
input
in
node
.
in
puts
],
dict
(
fail
=
"
%(fail)
s"
,
[
subd
[
output
]
for
output
in
node
.
outputs
]
,
id
=
"
%%(id)
s_
%
i"
%
j
))
dict
(
fail
=
"
%(fail)
s"
,
id
=
"
%%(id)
s_
%
i"
%
j
))
_c_code
+=
s
_c_code
+=
s
_c_code
+=
"
\n
"
_c_code
+=
"
\n
"
_c_code
+=
"}
\n
"
_c_code
+=
"}
\n
"
...
@@ -3454,7 +3453,7 @@ class Composite(ScalarOp):
...
@@ -3454,7 +3453,7 @@ class Composite(ScalarOp):
def
make_node
(
self
,
*
inputs
):
def
make_node
(
self
,
*
inputs
):
if
(
tuple
([
i
.
type
for
i
in
self
.
inputs
])
==
if
(
tuple
([
i
.
type
for
i
in
self
.
inputs
])
==
tuple
([
i
.
type
for
i
in
inputs
])):
tuple
([
i
.
type
for
i
in
inputs
])):
return
super
(
Composite
,
self
)
.
make_node
(
*
inputs
)
return
super
(
Composite
,
self
)
.
make_node
(
*
inputs
)
else
:
else
:
# Make a new op with the right input type.
# Make a new op with the right input type.
...
@@ -3489,7 +3488,7 @@ class Composite(ScalarOp):
...
@@ -3489,7 +3488,7 @@ class Composite(ScalarOp):
izip
((
"o
%
i"
%
i
for
i
in
xrange
(
len
(
onames
))),
izip
((
"o
%
i"
%
i
for
i
in
xrange
(
len
(
onames
))),
onames
)),
**
sub
)
onames
)),
**
sub
)
d
[
'nodename'
]
=
nodename
d
[
'nodename'
]
=
nodename
if
not
'id'
in
sub
:
if
'id'
not
in
sub
:
# The use of a dummy id is safe as the code is in a separate block.
# The use of a dummy id is safe as the code is in a separate block.
# It won't generate conflicting variable name.
# It won't generate conflicting variable name.
d
[
'id'
]
=
'_DUMMY_ID_'
d
[
'id'
]
=
'_DUMMY_ID_'
...
@@ -3521,8 +3520,8 @@ class Composite(ScalarOp):
...
@@ -3521,8 +3520,8 @@ class Composite(ScalarOp):
for
subnode
,
subnodename
in
zip
(
self
.
fgraph
.
toposort
(),
self
.
nodenames
):
for
subnode
,
subnodename
in
zip
(
self
.
fgraph
.
toposort
(),
self
.
nodenames
):
try
:
try
:
subnode_support_code
=
subnode
.
op
.
c_support_code_apply
(
subnode_support_code
=
subnode
.
op
.
c_support_code_apply
(
subnode
,
subnode
,
subnodename
%
dict
(
nodename
=
name
))
subnodename
%
dict
(
nodename
=
name
))
if
subnode_support_code
:
if
subnode_support_code
:
rval
.
append
(
subnode_support_code
)
rval
.
append
(
subnode_support_code
)
except
gof
.
utils
.
MethodNotDefined
:
except
gof
.
utils
.
MethodNotDefined
:
...
@@ -3536,9 +3535,9 @@ class Composite(ScalarOp):
...
@@ -3536,9 +3535,9 @@ class Composite(ScalarOp):
def
__eq__
(
self
,
other
):
def
__eq__
(
self
,
other
):
if
self
is
other
:
if
self
is
other
:
return
True
return
True
if
(
type
(
self
)
!=
type
(
other
)
if
(
type
(
self
)
!=
type
(
other
)
or
or
self
.
nin
!=
other
.
nin
self
.
nin
!=
other
.
nin
or
or
self
.
nout
!=
other
.
nout
):
self
.
nout
!=
other
.
nout
):
return
False
return
False
# see __hash__ for comment on why there is no mention of fgraph
# see __hash__ for comment on why there is no mention of fgraph
# or module cache key here.
# or module cache key here.
...
@@ -3546,9 +3545,9 @@ class Composite(ScalarOp):
...
@@ -3546,9 +3545,9 @@ class Composite(ScalarOp):
def
__hash__
(
self
):
def
__hash__
(
self
):
rval
=
hash
((
type
(
self
),
rval
=
hash
((
type
(
self
),
self
.
nin
,
self
.
nin
,
self
.
nout
,
self
.
nout
,
self
.
_c_code
))
self
.
_c_code
))
# Note that in general, the configparser settings at the time
# Note that in general, the configparser settings at the time
# of code generation (__init__) affect the semantics of this Op.
# of code generation (__init__) affect the semantics of this Op.
# This function assumes that all relevant info about the configparser
# This function assumes that all relevant info about the configparser
...
...
theano/tests/test_flake8.py
浏览文件 @
2b73732a
...
@@ -116,7 +116,6 @@ whitelist_flake8 = [
...
@@ -116,7 +116,6 @@ whitelist_flake8 = [
"tensor/nnet/tests/test_sigm.py"
,
"tensor/nnet/tests/test_sigm.py"
,
"scalar/basic_sympy.py"
,
"scalar/basic_sympy.py"
,
"scalar/__init__.py"
,
"scalar/__init__.py"
,
"scalar/basic.py"
,
"scalar/tests/test_basic.py"
,
"scalar/tests/test_basic.py"
,
"sandbox/test_theano_object.py"
,
"sandbox/test_theano_object.py"
,
"sandbox/test_scan.py"
,
"sandbox/test_scan.py"
,
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论