提交 640c12f3 authored 作者: Brandon T. Willard's avatar Brandon T. Willard 提交者: Brandon T. Willard

Remove fast_compile_gpu tag

上级 5415f12d
......@@ -71,14 +71,8 @@ OPT_NONE = OptimizationQuery(include=[], exclude=exclude)
OPT_MERGE = OptimizationQuery(include=["merge"], exclude=exclude)
OPT_FAST_RUN = OptimizationQuery(include=["fast_run"], exclude=exclude)
OPT_FAST_RUN_STABLE = OPT_FAST_RUN.requiring("stable")
# We need fast_compile_gpu here. As on the GPU, we don't have all
# operation that exist in fast_compile, but have some that get
# introduced in fast_run, we want those optimization to also run in
# fast_compile+gpu. We can't tag them just as 'gpu', as this would
# exclude them if we exclude 'gpu'.
OPT_FAST_COMPILE = OptimizationQuery(
include=["fast_compile", "fast_compile_gpu"], exclude=exclude
)
OPT_FAST_COMPILE = OptimizationQuery(include=["fast_compile"], exclude=exclude)
OPT_STABILIZE = OptimizationQuery(include=["fast_run"], exclude=exclude)
OPT_STABILIZE.position_cutoff = 1.5000001
OPT_NONE.name = "OPT_NONE"
......@@ -252,9 +246,7 @@ optdb.register(
) # 'fast_run', 'fast_compile')
# misc special cases for speed
optdb.register(
"specialize", EquilibriumDB(), "fast_run", "fast_compile_gpu", position=2
)
optdb.register("specialize", EquilibriumDB(), "fast_run", "fast_compile", position=2)
# misc special cases for speed that break canonicalization
optdb.register("uncanonicalize", EquilibriumDB(), "fast_run", position=3)
......
......@@ -1156,7 +1156,7 @@ def logsoftmax(c, axis=UNSET_AXIS):
return LogSoftmax(axis=axis)(c)
@register_specialize("fast_compile_gpu")
@register_specialize("fast_compile")
@local_optimizer([softmax_legacy])
def local_softmax_with_bias(fgraph, node):
"""
......@@ -1852,8 +1852,8 @@ class CrossentropyCategorical1Hot(Op):
crossentropy_categorical_1hot = CrossentropyCategorical1Hot()
@register_stabilize("fast_compile_gpu")
@register_specialize("fast_compile_gpu")
@register_stabilize("fast_compile")
@register_specialize("fast_compile")
@optimizer
def crossentropy_to_crossentropy_with_softmax_with_bias(fgraph):
"""
......@@ -1953,13 +1953,13 @@ optdb.register(
crossentropy_to_crossentropy_with_softmax,
"fast_run",
"xent",
"fast_compile_gpu",
"fast_compile",
position=2.01,
)
@register_specialize(
"fast_compile_gpu", "local_crossentropy_to_crossentropy_with_softmax_grad"
"fast_compile", "local_crossentropy_to_crossentropy_with_softmax_grad"
) # old name
@local_optimizer([softmax_grad_legacy])
def local_softmax_grad_to_crossentropy_with_softmax_grad(fgraph, node):
......@@ -1977,7 +1977,7 @@ def local_softmax_grad_to_crossentropy_with_softmax_grad(fgraph, node):
return [dx]
@register_specialize("fast_compile_gpu")
@register_specialize("fast_compile")
@local_optimizer([MaxAndArgmax])
def local_argmax_pushdown(fgraph, node):
if (
......@@ -2066,7 +2066,7 @@ def _is_const(z, val, approx=False):
return np.all(maybe == val)
@register_specialize("fast_compile_gpu")
@register_specialize("fast_compile")
@local_optimizer([AdvancedSubtensor, log])
def local_advanced_indexing_crossentropy_onehot(fgraph, node):
log_op = None
......@@ -2114,7 +2114,7 @@ def local_advanced_indexing_crossentropy_onehot(fgraph, node):
return [ret]
@register_specialize("fast_compile_gpu")
@register_specialize("fast_compile")
@local_optimizer([softmax_grad_legacy])
def local_advanced_indexing_crossentropy_onehot_grad(fgraph, node):
if not (node.op == softmax_grad_legacy and node.inputs[1].ndim == 2):
......@@ -2329,7 +2329,7 @@ def local_advanced_indexing_crossentropy_onehot_grad(fgraph, node):
return
@register_specialize("fast_compile_gpu")
@register_specialize("fast_compile")
@local_optimizer([softmax_with_bias])
def graph_merge_softmax_with_crossentropy_softmax(fgraph, node):
if node.op == softmax_with_bias:
......
......@@ -704,7 +704,7 @@ def local_subtensor_inc_subtensor(fgraph, node):
@register_specialize
@register_canonicalize("fast_compile_gpu")
@register_canonicalize("fast_compile")
@register_useless
@local_optimizer([Subtensor, AdvancedSubtensor1])
def local_subtensor_make_vector(fgraph, node):
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论