提交 a3424449 authored 作者: Frédéric Bastien's avatar Frédéric Bastien 提交者: GitHub

Merge pull request #5152 from adbrebs/h_softmax_doc

Fix h_softmax example data types
...@@ -2328,16 +2328,16 @@ def h_softmax(x, batch_size, n_outputs, n_classes, n_outputs_per_class, ...@@ -2328,16 +2328,16 @@ def h_softmax(x, batch_size, n_outputs, n_classes, n_outputs_per_class,
>>> output_size = n_outputs_per_class * n_outputs_per_class >>> output_size = n_outputs_per_class * n_outputs_per_class
>>> >>>
>>> # First level of h_softmax >>> # First level of h_softmax
>>> W1 = theano.shared(np.asarray( >>> floatX = theano.config.floatX
... np.random.normal(0, 0.001, (dim_x, n_classes)))) >>> W1 = theano.shared(
>>> b1 = theano.shared(np.asarray(np.zeros((n_classes,)))) ... np.random.normal(0, 0.001, (dim_x, n_classes)).astype(floatX))
>>> b1 = theano.shared(np.zeros((n_classes,), floatX))
>>> >>>
>>> # Second level of h_softmax >>> # Second level of h_softmax
>>> W2 = np.asarray(np.random.normal(0, 0.001, >>> W2 = np.random.normal(0, 0.001,
... size=(n_classes, dim_x, n_outputs_per_class))) ... size=(n_classes, dim_x, n_outputs_per_class)).astype(floatX)
>>> W2 = theano.shared(W2) >>> W2 = theano.shared(W2)
>>> b2 = theano.shared( >>> b2 = theano.shared(np.zeros((n_classes, n_outputs_per_class), floatX))
... np.asarray(np.zeros((n_classes, n_outputs_per_class))))
>>> >>>
>>> # We can now build the graph to compute a loss function, typically the >>> # We can now build the graph to compute a loss function, typically the
>>> # negative log-likelihood: >>> # negative log-likelihood:
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论