Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
47fc8d35
提交
47fc8d35
authored
12月 29, 2021
作者:
Brandon T. Willard
提交者:
Brandon T. Willard
1月 13, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Replace not `... == ...` with `... != ...`
上级
fbc28965
隐藏空白字符变更
内嵌
并排
正在显示
22 个修改的文件
包含
33 行增加
和
33 行删除
+33
-33
types.py
aesara/compile/function/types.py
+1
-1
io.py
aesara/compile/io.py
+1
-1
configdefaults.py
aesara/configdefaults.py
+1
-1
d3viz.py
aesara/d3viz/d3viz.py
+1
-1
dnn.py
aesara/gpuarray/dnn.py
+2
-2
pool.py
aesara/gpuarray/pool.py
+1
-1
basic.py
aesara/graph/basic.py
+1
-1
fg.py
aesara/graph/fg.py
+2
-2
opt.py
aesara/graph/opt.py
+1
-1
basic.py
aesara/link/basic.py
+1
-1
basic.py
aesara/scan/basic.py
+2
-2
basic_opt.py
aesara/tensor/basic_opt.py
+2
-2
elemwise.py
aesara/tensor/elemwise.py
+1
-1
math_opt.py
aesara/tensor/math_opt.py
+3
-3
abstract_conv.py
aesara/tensor/nnet/abstract_conv.py
+1
-1
basic.py
aesara/tensor/nnet/basic.py
+2
-2
conv.py
aesara/tensor/nnet/conv.py
+3
-3
opt.py
aesara/tensor/nnet/opt.py
+1
-1
shape.py
aesara/tensor/shape.py
+1
-1
type.py
aesara/typed_list/type.py
+1
-1
test_features.py
tests/graph/test_features.py
+2
-2
test_math_opt.py
tests/tensor/test_math_opt.py
+2
-2
没有找到文件。
aesara/compile/function/types.py
浏览文件 @
47fc8d35
...
@@ -1128,7 +1128,7 @@ def _pickle_Function(f):
...
@@ -1128,7 +1128,7 @@ def _pickle_Function(f):
# HACK to detect aliased storage.
# HACK to detect aliased storage.
# This is here because aliased relationships are not [currently]
# This is here because aliased relationships are not [currently]
# preserved across the pickle operation
# preserved across the pickle operation
if
not
(
f
.
pickle_aliased_memory_strategy
==
"ignore"
)
:
if
f
.
pickle_aliased_memory_strategy
!=
"ignore"
:
all_data
=
input_storage
+
inputs_data
all_data
=
input_storage
+
inputs_data
for
i
,
d_i
in
enumerate
(
all_data
):
for
i
,
d_i
in
enumerate
(
all_data
):
for
j
,
d_j
in
enumerate
(
all_data
):
for
j
,
d_j
in
enumerate
(
all_data
):
...
...
aesara/compile/io.py
浏览文件 @
47fc8d35
...
@@ -79,7 +79,7 @@ class SymbolicInput:
...
@@ -79,7 +79,7 @@ class SymbolicInput:
raise
TypeError
(
f
"name must be a string! (got: {self.name})"
)
raise
TypeError
(
f
"name must be a string! (got: {self.name})"
)
self
.
update
=
update
self
.
update
=
update
if
update
is
not
None
:
if
update
is
not
None
:
if
not
variable
.
type
=
=
update
.
type
:
if
variable
.
type
!
=
update
.
type
:
raise
TypeError
(
raise
TypeError
(
f
"Variable '{variable}' has type {variable.type} but an update of "
f
"Variable '{variable}' has type {variable.type} but an update of "
f
"type {update.type}. The type of the update should be "
f
"type {update.type}. The type of the update should be "
...
...
aesara/configdefaults.py
浏览文件 @
47fc8d35
...
@@ -1240,7 +1240,7 @@ def add_multiprocessing_configvars():
...
@@ -1240,7 +1240,7 @@ def add_multiprocessing_configvars():
f
"The environment variable OMP_NUM_THREADS should be a number, got '{var}'."
f
"The environment variable OMP_NUM_THREADS should be a number, got '{var}'."
)
)
else
:
else
:
default_openmp
=
not
int
(
var
)
=
=
1
default_openmp
=
int
(
var
)
!
=
1
else
:
else
:
# Check the number of cores availables.
# Check the number of cores availables.
count
=
os
.
cpu_count
()
count
=
os
.
cpu_count
()
...
...
aesara/d3viz/d3viz.py
浏览文件 @
47fc8d35
...
@@ -82,7 +82,7 @@ def d3viz(fct, outfile, copy_deps=True, *args, **kwargs):
...
@@ -82,7 +82,7 @@ def d3viz(fct, outfile, copy_deps=True, *args, **kwargs):
# Create output directory if not existing
# Create output directory if not existing
outdir
=
os
.
path
.
dirname
(
outfile
)
outdir
=
os
.
path
.
dirname
(
outfile
)
if
not
outdir
=
=
""
and
not
os
.
path
.
exists
(
outdir
):
if
outdir
!
=
""
and
not
os
.
path
.
exists
(
outdir
):
os
.
makedirs
(
outdir
)
os
.
makedirs
(
outdir
)
# Read template HTML file
# Read template HTML file
...
...
aesara/gpuarray/dnn.py
浏览文件 @
47fc8d35
...
@@ -248,7 +248,7 @@ def dnn_available(context_name):
...
@@ -248,7 +248,7 @@ def dnn_available(context_name):
ctx
=
get_context
(
context_name
)
ctx
=
get_context
(
context_name
)
if
not
ctx
.
kind
=
=
b
"cuda"
:
if
ctx
.
kind
!
=
b
"cuda"
:
dnn_available
.
msg
=
"Not on a CUDA device."
dnn_available
.
msg
=
"Not on a CUDA device."
return
False
return
False
...
@@ -719,7 +719,7 @@ def ensure_dt(val, default, name, dtype):
...
@@ -719,7 +719,7 @@ def ensure_dt(val, default, name, dtype):
val
=
as_scalar
(
val
)
val
=
as_scalar
(
val
)
if
not
isinstance
(
val
.
type
,
aesara
.
scalar
.
Scalar
):
if
not
isinstance
(
val
.
type
,
aesara
.
scalar
.
Scalar
):
raise
TypeError
(
f
"{name}: expected a scalar value"
)
raise
TypeError
(
f
"{name}: expected a scalar value"
)
if
not
val
.
type
.
dtype
=
=
dtype
:
if
val
.
type
.
dtype
!
=
dtype
:
val
=
val
.
astype
(
dtype
)
val
=
val
.
astype
(
dtype
)
return
val
return
val
...
...
aesara/gpuarray/pool.py
浏览文件 @
47fc8d35
...
@@ -258,7 +258,7 @@ class GpuAveragePoolGrad(CGpuKernelBase):
...
@@ -258,7 +258,7 @@ class GpuAveragePoolGrad(CGpuKernelBase):
if
pad
is
None
:
if
pad
is
None
:
pad
=
(
0
,)
*
nd
pad
=
(
0
,)
*
nd
elif
isinstance
(
pad
,
(
tuple
,
list
)):
elif
isinstance
(
pad
,
(
tuple
,
list
)):
if
max
(
pad
)
!=
0
and
not
self
.
mode
=
=
"average_exc_pad"
:
if
max
(
pad
)
!=
0
and
self
.
mode
!
=
"average_exc_pad"
:
raise
ValueError
(
"Padding must be zero for average_exc_pad"
)
raise
ValueError
(
"Padding must be zero for average_exc_pad"
)
ws
=
as_tensor_variable
(
ws
)
ws
=
as_tensor_variable
(
ws
)
stride
=
as_tensor_variable
(
stride
)
stride
=
as_tensor_variable
(
stride
)
...
...
aesara/graph/basic.py
浏览文件 @
47fc8d35
...
@@ -249,7 +249,7 @@ class Apply(Node):
...
@@ -249,7 +249,7 @@ class Apply(Node):
remake_node
=
False
remake_node
=
False
new_inputs
=
inputs
[:]
new_inputs
=
inputs
[:]
for
i
,
(
curr
,
new
)
in
enumerate
(
zip
(
self
.
inputs
,
new_inputs
)):
for
i
,
(
curr
,
new
)
in
enumerate
(
zip
(
self
.
inputs
,
new_inputs
)):
if
not
curr
.
type
=
=
new
.
type
:
if
curr
.
type
!
=
new
.
type
:
if
strict
:
if
strict
:
# If compatible, casts new into curr.type
# If compatible, casts new into curr.type
new_inputs
[
i
]
=
curr
.
type
.
filter_variable
(
new
)
new_inputs
[
i
]
=
curr
.
type
.
filter_variable
(
new
)
...
...
aesara/graph/fg.py
浏览文件 @
47fc8d35
...
@@ -450,7 +450,7 @@ class FunctionGraph(MetaObject):
...
@@ -450,7 +450,7 @@ class FunctionGraph(MetaObject):
# TODO: ERROR HANDLING FOR LISTENERS (should it complete the change or revert it?)
# TODO: ERROR HANDLING FOR LISTENERS (should it complete the change or revert it?)
if
node
==
"output"
:
if
node
==
"output"
:
r
=
self
.
outputs
[
i
]
r
=
self
.
outputs
[
i
]
if
not
r
.
type
=
=
new_var
.
type
:
if
r
.
type
!
=
new_var
.
type
:
raise
TypeError
(
raise
TypeError
(
f
"The type of the replacement ({new_var.type}) must be the"
f
"The type of the replacement ({new_var.type}) must be the"
f
" same as the type of the original Variable ({r.type})."
f
" same as the type of the original Variable ({r.type})."
...
@@ -458,7 +458,7 @@ class FunctionGraph(MetaObject):
...
@@ -458,7 +458,7 @@ class FunctionGraph(MetaObject):
self
.
outputs
[
i
]
=
new_var
self
.
outputs
[
i
]
=
new_var
else
:
else
:
r
=
node
.
inputs
[
i
]
r
=
node
.
inputs
[
i
]
if
not
r
.
type
=
=
new_var
.
type
:
if
r
.
type
!
=
new_var
.
type
:
raise
TypeError
(
raise
TypeError
(
f
"The type of the replacement ({new_var.type}) must be the"
f
"The type of the replacement ({new_var.type}) must be the"
f
" same as the type of the original Variable ({r.type})."
f
" same as the type of the original Variable ({r.type})."
...
...
aesara/graph/opt.py
浏览文件 @
47fc8d35
...
@@ -2832,7 +2832,7 @@ def _check_chain(r, chain):
...
@@ -2832,7 +2832,7 @@ def _check_chain(r, chain):
elif
r
.
owner
is
None
:
elif
r
.
owner
is
None
:
return
False
return
False
elif
isinstance
(
elem
,
Op
):
elif
isinstance
(
elem
,
Op
):
if
not
r
.
owner
.
op
=
=
elem
:
if
r
.
owner
.
op
!
=
elem
:
return
False
return
False
else
:
else
:
try
:
try
:
...
...
aesara/link/basic.py
浏览文件 @
47fc8d35
...
@@ -565,7 +565,7 @@ class WrapLinker(Linker):
...
@@ -565,7 +565,7 @@ class WrapLinker(Linker):
order_list0
=
order_lists
[
0
]
order_list0
=
order_lists
[
0
]
for
order_list
in
order_lists
[
1
:]:
for
order_list
in
order_lists
[
1
:]:
if
not
order_list0
=
=
order_list
:
if
order_list0
!
=
order_list
:
raise
Exception
(
raise
Exception
(
"All linkers to WrapLinker should execute operations in the same order."
"All linkers to WrapLinker should execute operations in the same order."
)
)
...
...
aesara/scan/basic.py
浏览文件 @
47fc8d35
...
@@ -761,7 +761,7 @@ def scan(
...
@@ -761,7 +761,7 @@ def scan(
if
isinstance
(
inner_out
.
type
,
TensorType
)
and
return_steps
.
get
(
pos
,
0
)
!=
1
:
if
isinstance
(
inner_out
.
type
,
TensorType
)
and
return_steps
.
get
(
pos
,
0
)
!=
1
:
outputs
[
pos
]
=
at
.
unbroadcast
(
shape_padleft
(
inner_out
),
0
)
outputs
[
pos
]
=
at
.
unbroadcast
(
shape_padleft
(
inner_out
),
0
)
if
return_list
is
not
True
and
len
(
outputs
)
==
1
:
if
not
return_list
and
len
(
outputs
)
==
1
:
outputs
=
outputs
[
0
]
outputs
=
outputs
[
0
]
return
(
outputs
,
updates
)
return
(
outputs
,
updates
)
...
@@ -1136,7 +1136,7 @@ def scan(
...
@@ -1136,7 +1136,7 @@ def scan(
# refers to update rule of index -1 - `pos`.
# refers to update rule of index -1 - `pos`.
update_map
[
sit_sot_shared
[
abs
(
pos
)
-
1
]]
=
_scan_out_list
[
idx
][
-
1
]
update_map
[
sit_sot_shared
[
abs
(
pos
)
-
1
]]
=
_scan_out_list
[
idx
][
-
1
]
scan_out_list
=
[
x
for
x
in
scan_out_list
if
x
is
not
None
]
scan_out_list
=
[
x
for
x
in
scan_out_list
if
x
is
not
None
]
if
return_list
is
not
True
and
len
(
scan_out_list
)
==
1
:
if
not
return_list
and
len
(
scan_out_list
)
==
1
:
scan_out_list
=
scan_out_list
[
0
]
scan_out_list
=
scan_out_list
[
0
]
elif
len
(
scan_out_list
)
==
0
:
elif
len
(
scan_out_list
)
==
0
:
scan_out_list
=
None
scan_out_list
=
None
...
...
aesara/tensor/basic_opt.py
浏览文件 @
47fc8d35
...
@@ -1390,7 +1390,7 @@ class ShapeFeature(features.Feature):
...
@@ -1390,7 +1390,7 @@ class ShapeFeature(features.Feature):
return
False
return
False
opx
=
dx
.
owner
.
op
opx
=
dx
.
owner
.
op
opy
=
dy
.
owner
.
op
opy
=
dy
.
owner
.
op
if
not
(
opx
.
i
==
opy
.
i
)
:
if
opx
.
i
!=
opy
.
i
:
return
False
return
False
# FB I'm not sure if this handle correctly constants.
# FB I'm not sure if this handle correctly constants.
if
dx
.
owner
.
inputs
[
0
]
==
dy
.
owner
.
inputs
[
0
]:
if
dx
.
owner
.
inputs
[
0
]
==
dy
.
owner
.
inputs
[
0
]:
...
@@ -1633,7 +1633,7 @@ def local_fill_sink(fgraph, node):
...
@@ -1633,7 +1633,7 @@ def local_fill_sink(fgraph, node):
if
(
if
(
hasattr
(
client
,
"op"
)
hasattr
(
client
,
"op"
)
and
isinstance
(
client
.
op
,
Elemwise
)
and
isinstance
(
client
.
op
,
Elemwise
)
and
not
client
.
op
=
=
fill
and
client
.
op
!
=
fill
):
):
client_inputs
=
client
.
inputs
[:]
client_inputs
=
client
.
inputs
[:]
client_inputs
[
cl_idx
]
=
c
client_inputs
[
cl_idx
]
=
c
...
...
aesara/tensor/elemwise.py
浏览文件 @
47fc8d35
...
@@ -187,7 +187,7 @@ class DimShuffle(ExternalCOp):
...
@@ -187,7 +187,7 @@ class DimShuffle(ExternalCOp):
def
make_node
(
self
,
_input
):
def
make_node
(
self
,
_input
):
input
=
as_tensor_variable
(
_input
)
input
=
as_tensor_variable
(
_input
)
ib
=
tuple
(
input
.
type
.
broadcastable
)
ib
=
tuple
(
input
.
type
.
broadcastable
)
if
not
ib
=
=
self
.
input_broadcastable
:
if
ib
!
=
self
.
input_broadcastable
:
if
len
(
ib
)
!=
len
(
self
.
input_broadcastable
):
if
len
(
ib
)
!=
len
(
self
.
input_broadcastable
):
raise
TypeError
(
raise
TypeError
(
"The number of dimensions of the "
"The number of dimensions of the "
...
...
aesara/tensor/math_opt.py
浏览文件 @
47fc8d35
...
@@ -1021,7 +1021,7 @@ class AlgebraicCanonizer(LocalOptimizer):
...
@@ -1021,7 +1021,7 @@ class AlgebraicCanonizer(LocalOptimizer):
assert
(
new
.
type
==
out
.
type
)
==
(
not
(
new
.
type
!=
out
.
type
))
assert
(
new
.
type
==
out
.
type
)
==
(
not
(
new
.
type
!=
out
.
type
))
if
n
ot
(
new
.
type
==
out
.
type
)
:
if
n
ew
.
type
!=
out
.
type
:
new
=
fill_chain
(
new
,
node
.
inputs
)[
0
]
new
=
fill_chain
(
new
,
node
.
inputs
)[
0
]
if
new
.
type
==
out
.
type
:
if
new
.
type
==
out
.
type
:
...
@@ -1138,7 +1138,7 @@ def local_sum_prod_mul_by_scalar(fgraph, node):
...
@@ -1138,7 +1138,7 @@ def local_sum_prod_mul_by_scalar(fgraph, node):
new_op_input_nb_elements
=
new_op_input
.
size
new_op_input_nb_elements
=
new_op_input
.
size
new_op_output
=
node
.
op
(
new_op_input
)
new_op_output
=
node
.
op
(
new_op_input
)
if
not
len
(
non_scalars
)
=
=
0
:
if
len
(
non_scalars
)
!
=
0
:
# Copy over stacktrace from previous output to new mul op,
# Copy over stacktrace from previous output to new mul op,
# for same reason as above.
# for same reason as above.
copy_stack_trace
(
node
.
outputs
,
new_op_output
)
copy_stack_trace
(
node
.
outputs
,
new_op_output
)
...
@@ -2519,7 +2519,7 @@ def local_greedy_distributor(fgraph, node):
...
@@ -2519,7 +2519,7 @@ def local_greedy_distributor(fgraph, node):
rval
=
local_mul_canonizer
.
merge_num_denum
(
new_num
,
new_denum
)
rval
=
local_mul_canonizer
.
merge_num_denum
(
new_num
,
new_denum
)
if
not
(
rval
.
type
==
out
.
type
)
:
if
rval
.
type
!=
out
.
type
:
# WHY DOES THIS HAPPEN?
# WHY DOES THIS HAPPEN?
return
False
return
False
...
...
aesara/tensor/nnet/abstract_conv.py
浏览文件 @
47fc8d35
...
@@ -3300,7 +3300,7 @@ class AbstractConv_gradInputs(BaseAbstractConv):
...
@@ -3300,7 +3300,7 @@ class AbstractConv_gradInputs(BaseAbstractConv):
expected_topgrad_shape
=
get_conv_output_shape
(
expected_topgrad_shape
=
get_conv_output_shape
(
imshp
,
kern
.
shape
,
self
.
border_mode
,
self
.
subsample
,
self
.
filter_dilation
imshp
,
kern
.
shape
,
self
.
border_mode
,
self
.
subsample
,
self
.
filter_dilation
)
)
if
not
tuple
(
expected_topgrad_shape
)
=
=
tuple
(
topgrad
.
shape
):
if
tuple
(
expected_topgrad_shape
)
!
=
tuple
(
topgrad
.
shape
):
raise
ValueError
(
raise
ValueError
(
"invalid input_shape for gradInputs: the given input_shape "
"invalid input_shape for gradInputs: the given input_shape "
"would produce an output of shape {}, but the given topgrad "
"would produce an output of shape {}, but the given topgrad "
...
...
aesara/tensor/nnet/basic.py
浏览文件 @
47fc8d35
...
@@ -2527,7 +2527,7 @@ class Prepend_scalar_constant_to_each_row(Op):
...
@@ -2527,7 +2527,7 @@ class Prepend_scalar_constant_to_each_row(Op):
def
make_node
(
self
,
mat
):
def
make_node
(
self
,
mat
):
# check type of input
# check type of input
x
=
at
.
as_tensor_variable
(
mat
)
x
=
at
.
as_tensor_variable
(
mat
)
if
not
mat
.
type
.
broadcastable
=
=
(
False
,
False
):
if
mat
.
type
.
broadcastable
!
=
(
False
,
False
):
raise
TypeError
(
"Expected a matrix as input"
)
raise
TypeError
(
"Expected a matrix as input"
)
y
=
at
.
as_tensor_variable
(
self
.
val
)
y
=
at
.
as_tensor_variable
(
self
.
val
)
assert
y
.
ndim
==
0
assert
y
.
ndim
==
0
...
@@ -2574,7 +2574,7 @@ class Prepend_scalar_to_each_row(Op):
...
@@ -2574,7 +2574,7 @@ class Prepend_scalar_to_each_row(Op):
x
=
at
.
as_tensor_variable
(
mat
)
x
=
at
.
as_tensor_variable
(
mat
)
if
isinstance
(
val
,
float
):
if
isinstance
(
val
,
float
):
val
=
aes
.
constant
(
val
)
val
=
aes
.
constant
(
val
)
if
not
mat
.
type
.
broadcastable
=
=
(
False
,
False
):
if
mat
.
type
.
broadcastable
!
=
(
False
,
False
):
raise
TypeError
(
"Expected a matrix as input"
)
raise
TypeError
(
"Expected a matrix as input"
)
y
=
at
.
as_tensor_variable
(
val
)
y
=
at
.
as_tensor_variable
(
val
)
assert
y
.
ndim
==
0
assert
y
.
ndim
==
0
...
...
aesara/tensor/nnet/conv.py
浏览文件 @
47fc8d35
...
@@ -1085,7 +1085,7 @@ class ConvOp(OpenMPOp):
...
@@ -1085,7 +1085,7 @@ class ConvOp(OpenMPOp):
# Determine gradient on inputs ########
# Determine gradient on inputs ########
mode
=
"valid"
mode
=
"valid"
if
not
self
.
out_mode
=
=
"full"
:
if
self
.
out_mode
!
=
"full"
:
mode
=
"full"
mode
=
"full"
filters
=
kerns
.
dimshuffle
((
1
,
0
,
2
,
3
))
filters
=
kerns
.
dimshuffle
((
1
,
0
,
2
,
3
))
...
@@ -1416,7 +1416,7 @@ if(%(value)s != %(expected)s){
...
@@ -1416,7 +1416,7 @@ if(%(value)s != %(expected)s){
d
[
"self_imshp_logical_stride_c"
]
=
int
(
d
[
"self_imshp_logical_stride_c"
]
=
int
(
np
.
ceil
(
self
.
imshp_logical
[
2
]
/
float
(
self
.
imshp
[
2
]))
np
.
ceil
(
self
.
imshp_logical
[
2
]
/
float
(
self
.
imshp
[
2
]))
)
)
if
not
self
.
imshp
[
0
]
=
=
1
:
if
self
.
imshp
[
0
]
!
=
1
:
d
[
"affectation"
]
=
"+="
d
[
"affectation"
]
=
"+="
d
[
"all_shape"
]
=
"1"
d
[
"all_shape"
]
=
"1"
d
[
"dim_zz_const"
]
=
"const"
d
[
"dim_zz_const"
]
=
"const"
...
@@ -1473,7 +1473,7 @@ if(kerns_dim[1] != img2d_dim[1]){
...
@@ -1473,7 +1473,7 @@ if(kerns_dim[1] != img2d_dim[1]){
f
"Type {node.inputs[0].type.dtype} not implemented"
f
"Type {node.inputs[0].type.dtype} not implemented"
)
)
d
[
"gemm"
]
=
"dgemm_"
d
[
"gemm"
]
=
"dgemm_"
if
not
d
[
"type"
]
=
=
"double"
:
if
d
[
"type"
]
!
=
"double"
:
d
[
"gemm"
]
=
"sgemm_"
d
[
"gemm"
]
=
"sgemm_"
if
self
.
imshp
!=
self
.
imshp_logical
or
self
.
kshp
!=
self
.
kshp_logical
:
if
self
.
imshp
!=
self
.
imshp_logical
or
self
.
kshp
!=
self
.
kshp_logical
:
...
...
aesara/tensor/nnet/opt.py
浏览文件 @
47fc8d35
...
@@ -448,7 +448,7 @@ def local_conv2d_gradinputs_cpu(fgraph, node):
...
@@ -448,7 +448,7 @@ def local_conv2d_gradinputs_cpu(fgraph, node):
return
None
return
None
mode
=
"valid"
mode
=
"valid"
if
no
t
node
.
op
.
border_mode
=
=
"full"
:
if
no
de
.
op
.
border_mode
!
=
"full"
:
mode
=
"full"
mode
=
"full"
filters
=
kern
.
dimshuffle
((
1
,
0
,
2
,
3
))
filters
=
kern
.
dimshuffle
((
1
,
0
,
2
,
3
))
filters
=
filters
[:,
:,
::
-
1
,
::
-
1
]
filters
=
filters
[:,
:,
::
-
1
,
::
-
1
]
...
...
aesara/tensor/shape.py
浏览文件 @
47fc8d35
...
@@ -413,7 +413,7 @@ class SpecifyShape(COp):
...
@@ -413,7 +413,7 @@ class SpecifyShape(COp):
raise
AssertionError
(
raise
AssertionError
(
f
"SpecifyShape: Got {x.ndim} dimensions (shape {x.shape}), expected {ndim} dimensions with shape {tuple(shape)}."
f
"SpecifyShape: Got {x.ndim} dimensions (shape {x.shape}), expected {ndim} dimensions with shape {tuple(shape)}."
)
)
if
not
np
.
all
(
x
.
shape
==
shape
):
if
x
.
shape
!=
tuple
(
shape
):
raise
AssertionError
(
raise
AssertionError
(
f
"SpecifyShape: Got shape {x.shape}, expected {tuple(shape)}."
f
"SpecifyShape: Got shape {x.shape}, expected {tuple(shape)}."
)
)
...
...
aesara/typed_list/type.py
浏览文件 @
47fc8d35
...
@@ -75,7 +75,7 @@ class TypedListType(CType):
...
@@ -75,7 +75,7 @@ class TypedListType(CType):
return
0
return
0
def
values_eq
(
self
,
a
,
b
):
def
values_eq
(
self
,
a
,
b
):
if
not
len
(
a
)
=
=
len
(
b
):
if
len
(
a
)
!
=
len
(
b
):
return
False
return
False
for
x
in
range
(
len
(
a
)):
for
x
in
range
(
len
(
a
)):
...
...
tests/graph/test_features.py
浏览文件 @
47fc8d35
...
@@ -74,7 +74,7 @@ class TestNodeFinder:
...
@@ -74,7 +74,7 @@ class TestNodeFinder:
assert
hasattr
(
g
,
"get_nodes"
)
assert
hasattr
(
g
,
"get_nodes"
)
for
type
,
num
in
((
add
,
3
),
(
sigmoid
,
3
),
(
dot
,
2
)):
for
type
,
num
in
((
add
,
3
),
(
sigmoid
,
3
),
(
dot
,
2
)):
if
not
len
([
t
for
t
in
g
.
get_nodes
(
type
)])
=
=
num
:
if
len
([
t
for
t
in
g
.
get_nodes
(
type
)])
!
=
num
:
raise
Exception
(
"Expected:
%
i times
%
s"
%
(
num
,
type
))
raise
Exception
(
"Expected:
%
i times
%
s"
%
(
num
,
type
))
new_e0
=
add
(
y
,
z
)
new_e0
=
add
(
y
,
z
)
assert
e0
.
owner
in
g
.
get_nodes
(
dot
)
assert
e0
.
owner
in
g
.
get_nodes
(
dot
)
...
@@ -83,7 +83,7 @@ class TestNodeFinder:
...
@@ -83,7 +83,7 @@ class TestNodeFinder:
assert
e0
.
owner
not
in
g
.
get_nodes
(
dot
)
assert
e0
.
owner
not
in
g
.
get_nodes
(
dot
)
assert
new_e0
.
owner
in
g
.
get_nodes
(
add
)
assert
new_e0
.
owner
in
g
.
get_nodes
(
add
)
for
type
,
num
in
((
add
,
4
),
(
sigmoid
,
3
),
(
dot
,
1
)):
for
type
,
num
in
((
add
,
4
),
(
sigmoid
,
3
),
(
dot
,
1
)):
if
not
len
([
t
for
t
in
g
.
get_nodes
(
type
)])
=
=
num
:
if
len
([
t
for
t
in
g
.
get_nodes
(
type
)])
!
=
num
:
raise
Exception
(
"Expected:
%
i times
%
s"
%
(
num
,
type
))
raise
Exception
(
"Expected:
%
i times
%
s"
%
(
num
,
type
))
...
...
tests/tensor/test_math_opt.py
浏览文件 @
47fc8d35
...
@@ -1740,7 +1740,7 @@ class TestFusion:
...
@@ -1740,7 +1740,7 @@ class TestFusion:
topo
=
f
.
maker
.
fgraph
.
toposort
()
topo
=
f
.
maker
.
fgraph
.
toposort
()
topo_
=
[
n
for
n
in
topo
if
not
isinstance
(
n
.
op
,
self
.
topo_exclude
)]
topo_
=
[
n
for
n
in
topo
if
not
isinstance
(
n
.
op
,
self
.
topo_exclude
)]
if
assert_len_topo
:
if
assert_len_topo
:
if
not
len
(
topo_
)
=
=
nb_elemwise
:
if
len
(
topo_
)
!
=
nb_elemwise
:
fail3
.
append
((
id
,
topo_
,
nb_elemwise
))
fail3
.
append
((
id
,
topo_
,
nb_elemwise
))
if
nb_elemwise
==
1
:
if
nb_elemwise
==
1
:
# if no variable appears multiple times in the
# if no variable appears multiple times in the
...
@@ -1753,7 +1753,7 @@ class TestFusion:
...
@@ -1753,7 +1753,7 @@ class TestFusion:
)
)
assert
expected_len_sym_inputs
==
len
(
sym_inputs
)
assert
expected_len_sym_inputs
==
len
(
sym_inputs
)
if
not
out_dtype
=
=
out
.
dtype
:
if
out_dtype
!
=
out
.
dtype
:
fail4
.
append
((
id
,
out_dtype
,
out
.
dtype
))
fail4
.
append
((
id
,
out_dtype
,
out
.
dtype
))
assert
len
(
fail1
+
fail2
+
fail3
+
fail4
)
==
0
assert
len
(
fail1
+
fail2
+
fail3
+
fail4
)
==
0
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论