提交 496d2727 authored 作者: Frederic Bastien's avatar Frederic Bastien

fix test following change to the grad again variable not in the graph.

上级 db7b0033
...@@ -2637,20 +2637,20 @@ def test_make_vector(): ...@@ -2637,20 +2637,20 @@ def test_make_vector():
assert mv.dtype == dtype assert mv.dtype == dtype
f = theano.function([b,i,d], mv) f = theano.function([b,i,d], mv)
f_val = f(val[b], val[i], val[d]) f_val = f(val[b], val[i], val[d])
print 'f_val =', f_val #print 'f_val =', f_val
s = mv.sum() s = mv.sum()
gb = T.grad(s, b) gb = T.grad(s, b, assume_continuously_differentiable=True)
gi = T.grad(s, i) gi = T.grad(s, i, assume_continuously_differentiable=True)
gd = T.grad(s, d) gd = T.grad(s, d, assume_continuously_differentiable=True)
print 'gb =', gb #print 'gb =', gb
print 'gi =', gi #print 'gi =', gi
print 'gd =', gd #print 'gd =', gd
g = theano.function([b,i,d], [gb, gi, gd]) g = theano.function([b,i,d], [gb, gi, gd])
g_val = g(val[b], val[i], val[d]) g_val = g(val[b], val[i], val[d])
print 'g_val =', g_val #print 'g_val =', g_val
if dtype.startswith('int'): if dtype.startswith('int'):
# The gradient should be 0 # The gradient should be 0
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论