Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
P
pytensor
项目
项目
详情
活动
周期分析
仓库
仓库
文件
提交
分支
标签
贡献者
图表
比较
统计图
议题
0
议题
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
CI / CD
CI / CD
流水线
作业
日程
统计图
Wiki
Wiki
代码片段
代码片段
成员
成员
折叠边栏
关闭边栏
活动
图像
聊天
创建新问题
作业
提交
问题看板
Open sidebar
testgroup
pytensor
Commits
4f5f7d6f
提交
4f5f7d6f
authored
10月 01, 2012
作者:
Frederic
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
small update to NEWS.txt to be able to update the new candidate release.
上级
7e185685
显示空白字符变更
内嵌
并排
正在显示
2 个修改的文件
包含
22 行增加
和
28 行删除
+22
-28
NEWS.txt
NEWS.txt
+19
-25
NEWS.txt
doc/NEWS.txt
+3
-3
没有找到文件。
NEWS.txt
浏览文件 @
4f5f7d6f
...
@@ -10,10 +10,6 @@ Release Notes
...
@@ -10,10 +10,6 @@ Release Notes
Theano 0.6rc1 (1 October 2012)
Theano 0.6rc1 (1 October 2012)
==============================
==============================
Updates in the Trunk since the last release up to
git log -p rel-0.5... |grep Merge|less
done up to and including merge of PR 976
Highlight:
Highlight:
* Bug fix, crash fix, CPU and GPU speed up.
* Bug fix, crash fix, CPU and GPU speed up.
...
@@ -25,10 +21,10 @@ Highlight:
...
@@ -25,10 +21,10 @@ Highlight:
* Use GPU asynchronous functionality (Frederic B.)
* Use GPU asynchronous functionality (Frederic B.)
* Better Windows support.
* Better Windows support.
Known bug
Known bug
:
* A few crash case that will be fixed by the final release.
* A few crash case that will be fixed by the final release.
Bug fixes
Bug fixes
:
* Outputs of Scan nodes could contain corrupted values: some parts of the
* Outputs of Scan nodes could contain corrupted values: some parts of the
output would be repeated a second time, instead of the correct values.
output would be repeated a second time, instead of the correct values.
It happened randomly, and quite infrequently, but the bug has been present
It happened randomly, and quite infrequently, but the bug has been present
...
@@ -39,7 +35,7 @@ Bug fixes
...
@@ -39,7 +35,7 @@ Bug fixes
was transformed into inc_subtensor on the GPU. Now we have a correct
was transformed into inc_subtensor on the GPU. Now we have a correct
(but slow) GPU implementation.
(but slow) GPU implementation.
Note 1: set_subtensor(x[slice[,...]], new_value) was working correctly
Note 1: set_subtensor(x[slice[,...]], new_value) was working correctly
in all cases as well as
inc_subtensor(*, *)
.
in all cases as well as
all inc_subtensor
.
Note 2: If your code was affected by the incorrect behavior, we now print
Note 2: If your code was affected by the incorrect behavior, we now print
a warning by default (Frederic B.)
a warning by default (Frederic B.)
* Fixed an issue whereby config values were used as default arguments,
* Fixed an issue whereby config values were used as default arguments,
...
@@ -72,11 +68,11 @@ Bug fixes
...
@@ -72,11 +68,11 @@ Bug fixes
This probably didn't cause problem as there is only the UsmmCscDense op
This probably didn't cause problem as there is only the UsmmCscDense op
(used call to Usmm wieh CSC matrix) that could interfere with them.
(used call to Usmm wieh CSC matrix) that could interfere with them.
Deprecation
Deprecation
:
* Deprecated the Module class (Ian G.)
* Deprecated the Module class (Ian G.)
This was a predecessor of SharedVariable with a less pythonic phylosophy.
This was a predecessor of SharedVariable with a less pythonic phylosophy.
Interface changes
Interface changes
:
* Now the base version requirement are numpy >= 1.5.0 and the optional scipy >= 0.8.
* Now the base version requirement are numpy >= 1.5.0 and the optional scipy >= 0.8.
* In Theano 0.5, we removed the deprecated sharedvar.value property.
* In Theano 0.5, we removed the deprecated sharedvar.value property.
Now we raise an error if you access it. (Frederic B.)
Now we raise an error if you access it. (Frederic B.)
...
@@ -123,7 +119,7 @@ New memory output contract(was told about in the release note of Theano 0.5):
...
@@ -123,7 +119,7 @@ New memory output contract(was told about in the release note of Theano 0.5):
* Updated the a few ops to respect this contract (Pascal L.)
* Updated the a few ops to respect this contract (Pascal L.)
New Features
New Features
:
* GPU scan now work (don't crash) when there is a mixture of float32 and other dtypes.
* GPU scan now work (don't crash) when there is a mixture of float32 and other dtypes.
* theano_var.eval({other_var:val[,...]} to simplify the usage of Theano (Ian G.)
* theano_var.eval({other_var:val[,...]} to simplify the usage of Theano (Ian G.)
* debugprint new param ids=["CHAR", "id", "int", ""]
* debugprint new param ids=["CHAR", "id", "int", ""]
...
@@ -190,7 +186,7 @@ New Features
...
@@ -190,7 +186,7 @@ New Features
Now it is applied more frequently. (Pascal L.)
Now it is applied more frequently. (Pascal L.)
New Op/function
New Op/function
:
* Added element-wise operation theano.tensor.{GammaLn,Psi} (John Salvatier, Nicolas Bouchard)
* Added element-wise operation theano.tensor.{GammaLn,Psi} (John Salvatier, Nicolas Bouchard)
* Added element-wise operation theano.tensor.{arcsin,arctan,arccosh,arcsinh,arctanh,exp2,arctan2} (Nicolas Bouchard)
* Added element-wise operation theano.tensor.{arcsin,arctan,arccosh,arcsinh,arctanh,exp2,arctan2} (Nicolas Bouchard)
* Added element-wise operation theano.tensor.{gamma,conj,complex_from_polar,expm1,deg2rad,rad2deg,trunc,gamma} (Nicolas Bouchard)
* Added element-wise operation theano.tensor.{gamma,conj,complex_from_polar,expm1,deg2rad,rad2deg,trunc,gamma} (Nicolas Bouchard)
...
@@ -209,7 +205,7 @@ New Op/function
...
@@ -209,7 +205,7 @@ New Op/function
* theano.sandbox.linalg.kron.py:Kron op. (Eric L.)
* theano.sandbox.linalg.kron.py:Kron op. (Eric L.)
Kronecker product
Kronecker product
Speed up
Speed up
:
* CPU convolution are now parallelized (Frederic B.)
* CPU convolution are now parallelized (Frederic B.)
By default use all cores/hyper-threads
By default use all cores/hyper-threads
To control it, use the `OMP_NUM_THREADS=N` environment variable where N is the number of
To control it, use the `OMP_NUM_THREADS=N` environment variable where N is the number of
...
@@ -230,7 +226,7 @@ Speed up
...
@@ -230,7 +226,7 @@ Speed up
There was warning printed by the subtensor optimization in those cases.
There was warning printed by the subtensor optimization in those cases.
* faster rng_mrg python code. (mostly used for tests) (Frederic B.)
* faster rng_mrg python code. (mostly used for tests) (Frederic B.)
Speed up GPU
Speed up GPU
:
* Convolution on the GPU now check the generation of the card to make
* Convolution on the GPU now check the generation of the card to make
it faster in some cases (especially medium/big ouput image) (Frederic B.)
it faster in some cases (especially medium/big ouput image) (Frederic B.)
* We had hardcoded 512 as the maximum number of thread per block. Newer card
* We had hardcoded 512 as the maximum number of thread per block. Newer card
...
@@ -243,14 +239,14 @@ Speed up GPU
...
@@ -243,14 +239,14 @@ Speed up GPU
* Faster creation of CudaNdarray objects (Frederic B.)
* Faster creation of CudaNdarray objects (Frederic B.)
* Now some Max reduction are implemented on the GPU. (Ian G.)
* Now some Max reduction are implemented on the GPU. (Ian G.)
Sparse Sandbox graduate (moved from theano.sparse.sandbox.sp)
Sparse Sandbox graduate (moved from theano.sparse.sandbox.sp)
:
* sparse.remove0 (Frederic B., Nicolas B.)
* sparse.remove0 (Frederic B., Nicolas B.)
* sparse.sp_sum(a, axis=None) (Nicolas B.)
* sparse.sp_sum(a, axis=None) (Nicolas B.)
* bugfix: the not structured grad was returning a structured grad.
* bugfix: the not structured grad was returning a structured grad.
* sparse.{col_scale,row_scale,ensure_sorted_indices,clean} (Nicolas B.)
* sparse.{col_scale,row_scale,ensure_sorted_indices,clean} (Nicolas B.)
* sparse.{diag,square_diagonal} (Nicolas B.)
* sparse.{diag,square_diagonal} (Nicolas B.)
Sparse
Sparse
:
* Support for uint* dtype.
* Support for uint* dtype.
* Implement theano.sparse.mul(sparse1, sparse2) when both inputs don't
* Implement theano.sparse.mul(sparse1, sparse2) when both inputs don't
have the same sparsity pattern. (Frederic B.)
have the same sparsity pattern. (Frederic B.)
...
@@ -270,7 +266,7 @@ Sparse
...
@@ -270,7 +266,7 @@ Sparse
* Implement the CSMProperties grad method (Yann Dauphin)
* Implement the CSMProperties grad method (Yann Dauphin)
* Move optimizations to theano/sparse/opt.py (Nicolas B.)
* Move optimizations to theano/sparse/opt.py (Nicolas B.)
New flags
New flags
:
* `profile=True` flag now print a printing of the sum of all printed profile.(Frederic B.)
* `profile=True` flag now print a printing of the sum of all printed profile.(Frederic B.)
* It work with the linker vm/cvm(default).
* It work with the linker vm/cvm(default).
* Also print compile time, optimizer time and linker time.
* Also print compile time, optimizer time and linker time.
...
@@ -295,7 +291,7 @@ New flags
...
@@ -295,7 +291,7 @@ New flags
* new flag `cxx`. This is the c++ compiler to use. If empty do not compile c code. (Frederic B.)
* new flag `cxx`. This is the c++ compiler to use. If empty do not compile c code. (Frederic B.)
* New flag `print_active_device` flag that default to True. (Matthew R.)
* New flag `print_active_device` flag that default to True. (Matthew R.)
Documentation
Documentation
:
* Added in the tutorial documentation on how to extend Theano.
* Added in the tutorial documentation on how to extend Theano.
This explains how to make a Theano Op from a Python function.
This explains how to make a Theano Op from a Python function.
http://deeplearning.net/software/theano/tutorial/extending_theano.html
http://deeplearning.net/software/theano/tutorial/extending_theano.html
...
@@ -310,11 +306,11 @@ Documentation
...
@@ -310,11 +306,11 @@ Documentation
* Doc type fix, Doc update, Better error messag: Olivier D., David W.F., Frederic B., James B., Matthew Rocklin, Ian G.
* Doc type fix, Doc update, Better error messag: Olivier D., David W.F., Frederic B., James B., Matthew Rocklin, Ian G.
* Python Memory Management (Steven Pigeon, Olivier D.)
* Python Memory Management (Steven Pigeon, Olivier D.)
Proposal
Proposal
:
* Math framework for complex gradien (Pascal L.)
* Math framework for complex gradien (Pascal L.)
Internal changes
Internal changes
:
* Define new exceptions MissingInputError and UnusedInputError, and use them
* Define new exceptions MissingInputError and UnusedInputError, and use them
in theano.function, instead of TypeError and ValueError. (Pascal L.)
in theano.function, instead of TypeError and ValueError. (Pascal L.)
* Better handling of bitwidth and max values of integers and pointers
* Better handling of bitwidth and max values of integers and pointers
...
@@ -344,7 +340,7 @@ Internal changes
...
@@ -344,7 +340,7 @@ Internal changes
* New theano.gof.sched.sort_apply_nodes() that will allow other execution ordering. (Matthew R.)
* New theano.gof.sched.sort_apply_nodes() that will allow other execution ordering. (Matthew R.)
* New attribute sort_schedule_fn, a way to specify a scheduler to use. (Matthew R.)
* New attribute sort_schedule_fn, a way to specify a scheduler to use. (Matthew R.)
Crash Fix
Crash Fix
:
* Fix import conflict name (usaar33, Frederic B.)
* Fix import conflict name (usaar33, Frederic B.)
* This make Theano work with PiCloud.
* This make Theano work with PiCloud.
* Do not try to use the BLAS library when blas.ldflags is manually set to an
* Do not try to use the BLAS library when blas.ldflags is manually set to an
...
@@ -388,7 +384,7 @@ Crash Fix
...
@@ -388,7 +384,7 @@ Crash Fix
* MaxArgmax.grad() when one of the gradient it receive is None. (Razvan P, reported by Mark Fenner)
* MaxArgmax.grad() when one of the gradient it receive is None. (Razvan P, reported by Mark Fenner)
* Fix crash of GpuSum when some dimensions shape was 0. (Frederic B.)
* Fix crash of GpuSum when some dimensions shape was 0. (Frederic B.)
Tests
Tests
:
* Use less memory (Olivier D.)(frix crash on 32-bits computers)
* Use less memory (Olivier D.)(frix crash on 32-bits computers)
* Fix test with Theano flag "blas.ldflags=". (Frederic B., Pascal L.)
* Fix test with Theano flag "blas.ldflags=". (Frederic B., Pascal L.)
* Fix crash with advanced subtensor and numpy constant.
* Fix crash with advanced subtensor and numpy constant.
...
@@ -398,7 +394,7 @@ Tests
...
@@ -398,7 +394,7 @@ Tests
* DebugMode now check the view_map for all type of Theano variable.
* DebugMode now check the view_map for all type of Theano variable.
It was doing only variable of tensor type. (Frederic B.)
It was doing only variable of tensor type. (Frederic B.)
Others
Others
:
* Remove python warning for some python version. (Gabe Schwartz)
* Remove python warning for some python version. (Gabe Schwartz)
* Remove useless fill op in fast_compile mode to make the graph more readable. (Fredric B.)
* Remove useless fill op in fast_compile mode to make the graph more readable. (Fredric B.)
* Remove GpuOuter as it is a subset of the new GpuGer (Frederic B.)
* Remove GpuOuter as it is a subset of the new GpuGer (Frederic B.)
...
@@ -411,8 +407,6 @@ Others
...
@@ -411,8 +407,6 @@ Others
Other thanks:
Other thanks:
* blaxill reported an error introduced into the trunk.
* blaxill reported an error introduced into the trunk.
New stuff that will probably be reworked/removed before the release
New stuff that will probably be reworked/removed before the release:
* new flag "time_seq_optimizer" (Frederic B.)
* new flag "time_eq_optimizer" (Frederic B.)
* Better PyCUDA sharing of the GPU context.(fix crash at exit) (Frederic B.)
* Better PyCUDA sharing of the GPU context.(fix crash at exit) (Frederic B.)
TODO: there is still a crash at exit!
TODO: there is still a crash at exit!
doc/NEWS.txt
浏览文件 @
4f5f7d6f
...
@@ -13,7 +13,7 @@ Highlight:
...
@@ -13,7 +13,7 @@ Highlight:
* Theano vision: http://deeplearning.net/software/theano/introduction.html#theano-vision (Many people)
* Theano vision: http://deeplearning.net/software/theano/introduction.html#theano-vision (Many people)
* Theano with GPU works in some cases on Windows now. Still experimental. (Sebastian Urban)
* Theano with GPU works in some cases on Windows now. Still experimental. (Sebastian Urban)
* Faster dot() call: New/Better direct call to cpu and gpu ger, gemv, gemm
* Faster dot() call: New/Better direct call to cpu and gpu ger, gemv, gemm
and dot(vector, vector). (James, Fr
édé
ric, Pascal)
and dot(vector, vector). (James, Fr
ede
ric, Pascal)
* C implementation of Alloc. (James, Pascal)
* C implementation of Alloc. (James, Pascal)
* theano.grad() now also work with sparse variable. (Arnaud)
* theano.grad() now also work with sparse variable. (Arnaud)
* Macro to implement the Jacobian/Hessian with theano.tensor.{jacobian,hessian} (Razvan)
* Macro to implement the Jacobian/Hessian with theano.tensor.{jacobian,hessian} (Razvan)
...
@@ -24,7 +24,7 @@ Interface Behavior Changes:
...
@@ -24,7 +24,7 @@ Interface Behavior Changes:
* The current default value of the parameter axis of
* The current default value of the parameter axis of
theano.{max,min,argmax,argmin,max_and_argmax} is now the same as
theano.{max,min,argmax,argmin,max_and_argmax} is now the same as
numpy: None. i.e. operate on all dimensions of the tensor.
numpy: None. i.e. operate on all dimensions of the tensor.
(Fr
édé
ric Bastien, Olivier Delalleau) (was deprecated and generated
(Fr
ede
ric Bastien, Olivier Delalleau) (was deprecated and generated
a warning since Theano 0.3 released Nov. 23rd, 2010)
a warning since Theano 0.3 released Nov. 23rd, 2010)
* The current output dtype of sum with input dtype [u]int* is now always [u]int64.
* The current output dtype of sum with input dtype [u]int* is now always [u]int64.
You can specify the output dtype with a new dtype parameter to sum.
You can specify the output dtype with a new dtype parameter to sum.
...
@@ -209,7 +209,7 @@ Crashes fixed:
...
@@ -209,7 +209,7 @@ Crashes fixed:
* When the subtensor inputs had 0 dimensions and the outputs 0 dimensions. (Frederic)
* When the subtensor inputs had 0 dimensions and the outputs 0 dimensions. (Frederic)
* Crash when the step to subtensor was not 1 in conjunction with some optimization. (Frederic, reported by Olivier Chapelle)
* Crash when the step to subtensor was not 1 in conjunction with some optimization. (Frederic, reported by Olivier Chapelle)
* Runtime crash related to an optimization with subtensor of alloc (reported by Razvan, fixed by Frederic)
* Runtime crash related to an optimization with subtensor of alloc (reported by Razvan, fixed by Frederic)
* Fix dot22scalar cast of integer scalars (Justin Bayer, Fr
édé
ric, Olivier)
* Fix dot22scalar cast of integer scalars (Justin Bayer, Fr
ede
ric, Olivier)
* Fix runtime crash in gemm, dot22. FB
* Fix runtime crash in gemm, dot22. FB
* Fix on 32bits computer: make sure all shape are int64.(Olivier)
* Fix on 32bits computer: make sure all shape are int64.(Olivier)
* Fix to deque on python 2.4 (Olivier)
* Fix to deque on python 2.4 (Olivier)
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论