提交 835c39f4 authored 作者: Olivier Delalleau's avatar Olivier Delalleau

Minor typo fixes in doc

上级 262f59a2
...@@ -846,18 +846,18 @@ Reductions ...@@ -846,18 +846,18 @@ Reductions
:Parameter: *no_zeros_in_input* - The grad of prod is complicated :Parameter: *no_zeros_in_input* - The grad of prod is complicated
as we need to handle 3 different cases: without zeros in the as we need to handle 3 different cases: without zeros in the
input reduced group, with 1 zeros or with more zeros. input reduced group, with 1 zero or with more zeros.
This could slow you down, but more importantly, we currently This could slow you down, but more importantly, we currently
don't support the second derivative of the 3 cases. So you don't support the second derivative of the 3 cases. So you
can not take the second derivative of the default prod(). cannot take the second derivative of the default prod().
To remove the handling of the special cases of 0 and so get To remove the handling of the special cases of 0 and so get
some small speed up and allow second derivative set some small speed up and allow second derivative set
``no_zeros_in_inputs`` to ``True``. It default to ``False``. ``no_zeros_in_inputs`` to ``True``. It defaults to ``False``.
**It is the user responsability to make sure there is no zeros **It is the user responsibility to make sure there are no zeros
in the inputs. If there is, the grad will be wrong.** in the inputs. If there are, the grad will be wrong.**
:Returns: product of every term in *x* along *axis* :Returns: product of every term in *x* along *axis*
......
// REMEMBER TO RAISE c_code_cache_version when changing this file // REMEMBER TO INCREASE c_code_cache_version when changing this file
// //
//TODO detect SHARED_SIZE dynamically //TODO detect SHARED_SIZE dynamically
#define SHARED_SIZE (16*1024) #define SHARED_SIZE (16*1024)
......
...@@ -31,21 +31,21 @@ class GpuConv(gof.Op): ...@@ -31,21 +31,21 @@ class GpuConv(gof.Op):
imshp=None, imshp=None,
max_threads_dim0=None): max_threads_dim0=None):
""" """
:param version: each version of c_code implement many kernel for the :param version: each version of c_code implements many kernels for the
convolution. By default we try to guess the best one. convolution. By default we try to guess the best one.
You can force one version with this parameter. This You can force one version with this parameter. This
parameter is used by the tests. parameter is used by the tests.
:param verbose: for value of 1,2 and 3. Print more information during :param verbose: for value of 1,2 and 3. Print more information during
the execution of the convolution. Mostly used for the execution of the convolution. Mostly used for
optimization or debugging. optimization or debugging.
:param kshp: The size of the kernel. If provided, can genera :param kshp: The size of the kernel. If provided, can generate
faster code. If the GpuConv op is automatically faster code. If the GpuConv op is automatically
inserted, inserted,
we take its value automatically from the Conv op. we take its value automatically from the Conv op.
:param imshp: The size of the image. Not used for code generation but :param imshp: The size of the image. Not used for code generation but
allow to select an experimental new version in another allows to select an experimental new version in another
repo. repo.
:param max_threads_dim0: The maximum number of thread for the :param max_threads_dim0: The maximum number of threads for the
block size dimensions 0 (blockDim.x) used by the block size dimensions 0 (blockDim.x) used by the
GPU function. GPU function.
......
// REMEMBER TO RAISE c_code_cache_version when changing this file // REMEMBER TO INCREASE c_code_cache_version when changing this file
// //
//implement the valid convolution only //implement the valid convolution only
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论