提交 278ecf5b authored 作者: nouiz's avatar nouiz

Merge pull request #1121 from delallea/minor

Minor stuff
...@@ -770,9 +770,9 @@ Alternative: Anaconda 0.8.3 (Linux VM on Windows) ...@@ -770,9 +770,9 @@ Alternative: Anaconda 0.8.3 (Linux VM on Windows)
################################################# #################################################
ContinuumIO_ was providing a free VM with Theano installed. Now they ContinuumIO_ was providing a free VM with Theano installed. Now they
provide a new installation system that install itself on Windows. We provide a new installation system that installs itself on Windows.
don't have the time now to update the docs, so we remove the old Since we lack time to update this part of the documentation, it has been
documentations that don't work. removed for now. Check back later!
.. _ContinuumIO: http://continuum.io .. _ContinuumIO: http://continuum.io
......
...@@ -199,7 +199,7 @@ import theano and print the config variable, as in: ...@@ -199,7 +199,7 @@ import theano and print the config variable, as in:
Bool value: either True or False Bool value: either True or False
Default: True if the environment variable OMP_NUM_THREADS!=1 or Default: True if the environment variable OMP_NUM_THREADS!=1 or
if we detect more then 1 CPU core. Otherwise False. if we detect more than 1 CPU core. Otherwise False.
Enable or not parallel computation on the CPU with OpenMP. Enable or not parallel computation on the CPU with OpenMP.
It is the default value used when creating an Op that support it. It is the default value used when creating an Op that support it.
......
...@@ -15,7 +15,7 @@ Note that you want SciPy >= 0.7.2 ...@@ -15,7 +15,7 @@ Note that you want SciPy >= 0.7.2
dimensions. There may be more bugs. It also has inconsistent dimensions. There may be more bugs. It also has inconsistent
implementation of sparse matrices. implementation of sparse matrices.
We do not test against version bellow SciPy 0.7.2. We do not test against SciPy versions below 0.7.2.
We describe the details of the compressed sparse matrix types. We describe the details of the compressed sparse matrix types.
``scipy.sparse.csc_matrix`` ``scipy.sparse.csc_matrix``
......
...@@ -2304,7 +2304,7 @@ class GpuAdvancedSubtensor1(tensor.AdvancedSubtensor1, GpuOp): ...@@ -2304,7 +2304,7 @@ class GpuAdvancedSubtensor1(tensor.AdvancedSubtensor1, GpuOp):
x, idx = inp x, idx = inp
out, = out_ out, = out_
x_orig = x x_orig = x
#TODO: if more then 3 dims, reshape the inputs even if not all #TODO: if more than 3 dims, reshape the inputs even if not all
#dimensions are c contiguous #dimensions are c contiguous
if x.ndim > 3 and x.is_c_contiguous(): if x.ndim > 3 and x.is_c_contiguous():
x = x.reshape((x.shape[0], numpy.prod(x.shape[1:]))) x = x.reshape((x.shape[0], numpy.prod(x.shape[1:])))
......
...@@ -395,7 +395,7 @@ def get_valid_shapes(): ...@@ -395,7 +395,7 @@ def get_valid_shapes():
, ((20, 16, 32, 32), (1, 16, 28, 28), (1, 1), (1, 1), (1, 1)) # layer 1 backprop to weights , ((20, 16, 32, 32), (1, 16, 28, 28), (1, 1), (1, 1), (1, 1)) # layer 1 backprop to weights
, ((60,20,28,28), (10,20,5,5), (1, 1), (2,2), (1, 1))#added a test case that fail from test_nnet.py.test_conv_nnet2 , ((60,20,28,28), (10,20,5,5), (1, 1), (2,2), (1, 1))#added a test case that fail from test_nnet.py.test_conv_nnet2
, ((10,5,28,28), (10,5,5,5), (1, 1), (2,2), (1, 1))#test precedent but reduced that triger the error , ((10,5,28,28), (10,5,5,5), (1, 1), (2,2), (1, 1))#test precedent but reduced that triger the error
#Test more then maxThreadsDim0 #Test more than maxThreadsDim0
, ((2,4,13,1050), (3,4,10, 11), (1, 1), (1, 1), (1, 1)) , ((2,4,13,1050), (3,4,10, 11), (1, 1), (1, 1), (1, 1))
, ((2,4,1050,13), (3,4,10, 11), (1, 1), (1, 1), (1, 1)) , ((2,4,1050,13), (3,4,10, 11), (1, 1), (1, 1), (1, 1))
] ]
...@@ -667,7 +667,7 @@ def test_full(): ...@@ -667,7 +667,7 @@ def test_full():
, ((10,30,23,23),(20,30,7,7), (1, 1), (1, 1), (1, 1))#test_lenet_64 full , ((10,30,23,23),(20,30,7,7), (1, 1), (1, 1), (1, 1))#test_lenet_64 full
# , ((20,10,29,29),(30,10,23,23), (1, 1), (1, 1), (1, 1))#test_lenet_64 bprop 1 # , ((20,10,29,29),(30,10,23,23), (1, 1), (1, 1), (1, 1))#test_lenet_64 bprop 1
# , ((1,10,64,64),(20,10,58,58), (1, 1), (1, 1), (1, 1))#test_lenet_64 bprop 2 # , ((1,10,64,64),(20,10,58,58), (1, 1), (1, 1), (1, 1))#test_lenet_64 bprop 2
#Test more then maxThreadsDim0 #Test more than maxThreadsDim0
, ((2,4,13,1050), (3,4,10, 11), (1, 1), (1, 1), (1, 1)) , ((2,4,13,1050), (3,4,10, 11), (1, 1), (1, 1), (1, 1))
, ((2,4,1050,13), (3,4,10, 11), (1, 1), (1, 1), (1, 1)) , ((2,4,1050,13), (3,4,10, 11), (1, 1), (1, 1), (1, 1))
] ]
......
...@@ -2377,8 +2377,8 @@ class SpecifyShape(Op): ...@@ -2377,8 +2377,8 @@ class SpecifyShape(Op):
def c_code(self, node, nodename, inp, out, sub): def c_code(self, node, nodename, inp, out, sub):
if not isinstance(node.inputs[0], TensorVariable): if not isinstance(node.inputs[0], TensorVariable):
# The c code bellow support only Tensor. super.c_code # The C code below supports only Tensor. super.c_code
# will raise an exception to tell that there isn't c code # will raise an exception to tell that there is no C code
# for the other cases. # for the other cases.
return super(SpecifyShape, self).c_code(node, nodename, return super(SpecifyShape, self).c_code(node, nodename,
inp, out, sub) inp, out, sub)
...@@ -2389,8 +2389,8 @@ class SpecifyShape(Op): ...@@ -2389,8 +2389,8 @@ class SpecifyShape(Op):
return """ return """
if (PyArray_NDIM(%(iname)s) != PyArray_DIMS(%(shape)s)[0]) { if (PyArray_NDIM(%(iname)s) != PyArray_DIMS(%(shape)s)[0]) {
PyErr_Format(PyExc_AssertionError, PyErr_Format(PyExc_AssertionError,
"SpecifyShape: vector of shape have %%d element," "SpecifyShape: vector of shape has %%d elements,"
" but the input have %%d dimensions.", " but the input has %%d dimensions.",
PyArray_NDIM(%(iname)s), PyArray_NDIM(%(iname)s),
PyArray_DIMS(%(shape)s)[0]); PyArray_DIMS(%(shape)s)[0]);
%(fail)s; %(fail)s;
...@@ -2400,7 +2400,7 @@ class SpecifyShape(Op): ...@@ -2400,7 +2400,7 @@ class SpecifyShape(Op):
i))[0]; i))[0];
if (PyArray_DIMS(%(iname)s)[i] != shp) { if (PyArray_DIMS(%(iname)s)[i] != shp) {
PyErr_Format(PyExc_AssertionError, PyErr_Format(PyExc_AssertionError,
"SpecifyShape: dim %%d of input have shape %%d," "SpecifyShape: dim %%d of input has shape %%d,"
" expected %%d.", " expected %%d.",
i, PyArray_DIMS(%(iname)s)[i], i, PyArray_DIMS(%(iname)s)[i],
shp); shp);
......
...@@ -491,7 +491,7 @@ def fill_diagonal(a, val): ...@@ -491,7 +491,7 @@ def fill_diagonal(a, val):
2, the main diagonal is the list of locations a[i, i, ..., i] 2, the main diagonal is the list of locations a[i, i, ..., i]
(i.e. with indices all identical).) (i.e. with indices all identical).)
Support rectangular matrix and tensor with more then 2 dimensions Support rectangular matrix and tensor with more than 2 dimensions
if the later have all dimensions are equals. if the later have all dimensions are equals.
.. versionadded:: 0.6 .. versionadded:: 0.6
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论