提交 8e8b9f99 authored 作者: Olivier Delalleau's avatar Olivier Delalleau

Typo fixes

上级 d4dfbf2a
...@@ -65,8 +65,8 @@ New Features ...@@ -65,8 +65,8 @@ New Features
(Frederic B., Simon McGregor) (Frederic B., Simon McGregor)
* MRG random now raises an error with a clear message when the passed shape * MRG random now raises an error with a clear message when the passed shape
contains dimensions with bad value like 0. (Frédéric B. reported by Ian G.) contains dimensions with bad value like 0. (Frédéric B. reported by Ian G.)
* "CudaNdarray[*] = ndarray" work in more case (Frederic B.) * "CudaNdarray[*] = ndarray" works in more cases (Frederic B.)
* "CudaNdarray[*] += ndarray" work in more case (Frederic B.) * "CudaNdarray[*] += ndarray" works in more cases (Frederic B.)
* We add dimensions to CudaNdarray to automatically broadcast more frequently. * We add dimensions to CudaNdarray to automatically broadcast more frequently.
(Frederic B.) (Frederic B.)
* theano.tensor.argsort that wrap numpy.argsort (Hani Almousli). * theano.tensor.argsort that wrap numpy.argsort (Hani Almousli).
......
...@@ -1939,10 +1939,10 @@ class GpuAdvancedIncSubtensor1(tensor.AdvancedIncSubtensor1, GpuOp): ...@@ -1939,10 +1939,10 @@ class GpuAdvancedIncSubtensor1(tensor.AdvancedIncSubtensor1, GpuOp):
return Apply(self, [x_, y_, ilist_], [x_.type()]) return Apply(self, [x_, y_, ilist_], [x_.type()])
# CudaNdarray_Subscript() don't support Advanced slicing. # CudaNdarray_Subscript() doesn't support Advanced slicing.
# But we can't use the parent version that loop on each indices # But we can't use the parent version that loops on each index
# as we also need to loop when set_instead_of_inc is True and the # as we also need to loop when set_instead_of_inc is True and the
# parent don't loop in that case. # parent doesn't loop in that case.
def perform(self, node, inp, out_): def perform(self, node, inp, out_):
# TODO opt to make this inplace # TODO opt to make this inplace
x, y, idx = inp x, y, idx = inp
...@@ -1950,7 +1950,7 @@ class GpuAdvancedIncSubtensor1(tensor.AdvancedIncSubtensor1, GpuOp): ...@@ -1950,7 +1950,7 @@ class GpuAdvancedIncSubtensor1(tensor.AdvancedIncSubtensor1, GpuOp):
if not self.inplace: if not self.inplace:
x = x.copy() x = x.copy()
if self.set_instead_of_inc: if self.set_instead_of_inc:
# CudaNdarray __setitem__ don't do broadcast nor support # CudaNdarray __setitem__ doesn't do broadcast nor support
# list of index. # list of index.
assert y.ndim <= x.ndim # Should be guaranteed by `make_node` assert y.ndim <= x.ndim # Should be guaranteed by `make_node`
if y.ndim == x.ndim: if y.ndim == x.ndim:
......
...@@ -1066,7 +1066,7 @@ CudaNdarray_inplace_elemwise(PyObject* py_self, PyObject * py_other, operator_t ...@@ -1066,7 +1066,7 @@ CudaNdarray_inplace_elemwise(PyObject* py_self, PyObject * py_other, operator_t
PyErr_SetString( PyErr_SetString(
PyExc_ValueError, PyExc_ValueError,
"CudaNdarray_inplace_elemwise cannot work inplace on" "CudaNdarray_inplace_elemwise cannot work inplace on"
" un-initialized array when the new value have more then" " un-initialized array when the new value have more than"
" 0 or 1 broadcastable dimensions"); " 0 or 1 broadcastable dimensions");
Py_XDECREF(new_other); Py_XDECREF(new_other);
return 0; return 0;
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论