提交 8dd3215e authored 作者: amrithasuresh's avatar amrithasuresh

Updated numpy as np

上级 07996d4e
...@@ -54,7 +54,7 @@ import logging ...@@ -54,7 +54,7 @@ import logging
import copy import copy
from sys import maxsize from sys import maxsize
from collections import OrderedDict from collections import OrderedDict
import numpy import numpy as np
import theano import theano
from theano import tensor, scalar from theano import tensor, scalar
...@@ -636,7 +636,7 @@ class PushOutSeqScan(gof.Optimizer): ...@@ -636,7 +636,7 @@ class PushOutSeqScan(gof.Optimizer):
if out in op.inner_mitsot_outs(ls): if out in op.inner_mitsot_outs(ls):
odx = op.inner_mitsot_outs(ls).index(out) odx = op.inner_mitsot_outs(ls).index(out)
inp = op.outer_mitsot(node)[odx] inp = op.outer_mitsot(node)[odx]
st = abs(numpy.min(op.mitsot_taps())) st = abs(np.min(op.mitsot_taps()))
y = tensor.set_subtensor(inp[st:], _y) y = tensor.set_subtensor(inp[st:], _y)
elif out in op.inner_sitsot_outs(ls): elif out in op.inner_sitsot_outs(ls):
odx = op.inner_sitsot_outs(ls).index(out) odx = op.inner_sitsot_outs(ls).index(out)
...@@ -1373,7 +1373,7 @@ class ScanSaveMem(gof.Optimizer): ...@@ -1373,7 +1373,7 @@ class ScanSaveMem(gof.Optimizer):
# TODO: Simplify the number of steps needed. # TODO: Simplify the number of steps needed.
# FB: This need good testing, left to later. # FB: This need good testing, left to later.
# call get_scalar_constant_value()? it can # call get_scalar_constant_value()? it can
# return python/numpy scalar or numpy.ndarray # return python/numpy scalar or np.ndarray
# currently. # currently.
# pval = pre_greedy_local_optimizer(list_opt_slice, # pval = pre_greedy_local_optimizer(list_opt_slice,
# pval) # pval)
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论