提交 3ae40423 authored 作者: Frederic's avatar Frederic

small doc fix.

上级 370b2ff1
...@@ -40,7 +40,7 @@ following methods: ...@@ -40,7 +40,7 @@ following methods:
variables of the computation must be put. More specifically: variables of the computation must be put. More specifically:
- ``node``: This is a reference to an Apply node which was previously - ``node``: This is a reference to an Apply node which was previously
obtained via ``mul``'s ``make_node`` method. It is typically not obtained via the ``Op``'s ``make_node`` method. It is typically not
used in simple Ops, but it contains symbolic information that used in simple Ops, but it contains symbolic information that
could be required for complex Ops. could be required for complex Ops.
...@@ -94,18 +94,14 @@ following methods: ...@@ -94,18 +94,14 @@ following methods:
lifetime of self. Op instances should be immutable in this lifetime of self. Op instances should be immutable in this
sense. sense.
.. function:: __ne__(other)
*Default:* ``(not (self==other))``
.. function:: grad(inputs, output_gradients) .. function:: grad(inputs, output_gradients)
Optional. Optional (but needed if you want to have it work with {tensor,sparse}.grad())
If the Op you are defining is differentiable, you can define its If the Op you are defining is differentiable, you can define its
gradient symbolically in this method. gradient symbolically in this method.
Both the ``inputs`` and ``output_gradients`` will be Both the ``inputs`` and ``output_gradients`` will be list of Theano
Variables. This method must return a list containing one Variable Variables. This method must return a list containing one Variable
(or ``None``) for each input. Each returned Variable represents the (or ``None``) for each input. Each returned Variable represents the
gradient with respect to that input given the symbolic gradients gradient with respect to that input given the symbolic gradients
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论