提交 4f44bfa7 authored 作者: Frederic's avatar Frederic

test less verbose

上级 542ff218
...@@ -264,7 +264,7 @@ def test_mlp(): ...@@ -264,7 +264,7 @@ def test_mlp():
###################### ######################
# BUILD ACTUAL MODEL # # BUILD ACTUAL MODEL #
###################### ######################
print '... building the model' #print '... building the model'
# allocate symbolic variables for the data # allocate symbolic variables for the data
index = T.lscalar() # index to a [mini]batch index = T.lscalar() # index to a [mini]batch
...@@ -302,8 +302,8 @@ def test_mlp(): ...@@ -302,8 +302,8 @@ def test_mlp():
x:train_set_x[index*batch_size:(index+1)*batch_size], x:train_set_x[index*batch_size:(index+1)*batch_size],
y:train_set_y[index*batch_size:(index+1)*batch_size]}, y:train_set_y[index*batch_size:(index+1)*batch_size]},
mode=mode) mode=mode)
print 'MODEL 1' #print 'MODEL 1'
theano.printing.debugprint(train_model, print_type=True) #theano.printing.debugprint(train_model, print_type=True)
assert any([isinstance(i.op,T.nnet.CrossentropySoftmax1HotWithBiasDx) for i in train_model.maker.env.toposort()]) assert any([isinstance(i.op,T.nnet.CrossentropySoftmax1HotWithBiasDx) for i in train_model.maker.env.toposort()])
# Even without FeatureShape # Even without FeatureShape
...@@ -313,9 +313,9 @@ def test_mlp(): ...@@ -313,9 +313,9 @@ def test_mlp():
givens={ givens={
x:train_set_x[index*batch_size:(index+1)*batch_size], x:train_set_x[index*batch_size:(index+1)*batch_size],
y:train_set_y[index*batch_size:(index+1)*batch_size]}) y:train_set_y[index*batch_size:(index+1)*batch_size]})
print #print
print 'MODEL 2' #print 'MODEL 2'
theano.printing.debugprint(train_model, print_type=True) #theano.printing.debugprint(train_model, print_type=True)
assert any([isinstance(i.op,T.nnet.CrossentropySoftmax1HotWithBiasDx) for i in train_model.maker.env.toposort()]) assert any([isinstance(i.op,T.nnet.CrossentropySoftmax1HotWithBiasDx) for i in train_model.maker.env.toposort()])
if __name__ == '__main__': if __name__ == '__main__':
......
...@@ -185,7 +185,7 @@ class QuadraticDenoisingAA(module.Module): ...@@ -185,7 +185,7 @@ class QuadraticDenoisingAA(module.Module):
#self.validate = theano.Method(self.input, [self.cost, self.output]) #self.validate = theano.Method(self.input, [self.cost, self.output])
def _instance_initialize(self, obj, input_size, hidden_size, seed, lr, qfilter_relscale): def _instance_initialize(self, obj, input_size, hidden_size, seed, lr, qfilter_relscale):
print 'QDAA init' #print 'QDAA init'
""" """
qfilter_relscale is the initial range for any quadratic filters (relative to the linear qfilter_relscale is the initial range for any quadratic filters (relative to the linear
filter's initial range) filter's initial range)
...@@ -454,11 +454,11 @@ class ConvolutionalMLP(module.FancyModule): ...@@ -454,11 +454,11 @@ class ConvolutionalMLP(module.FancyModule):
i.initialize(input_size=self.input_size, i.initialize(input_size=self.input_size,
hidden_size=self.input_representation_size, noise_level=noise_level, hidden_size=self.input_representation_size, noise_level=noise_level,
seed=int(R.random_integers(2**30)), lr=lr, qfilter_relscale=qfilter_relscale) seed=int(R.random_integers(2**30)), lr=lr, qfilter_relscale=qfilter_relscale)
print type(i.w1) #print type(i.w1)
assert isinstance(i.w1, N.ndarray) assert isinstance(i.w1, N.ndarray)
for i in self.input_representations[1:]: for i in self.input_representations[1:]:
print type(i.w1) #print type(i.w1)
assert isinstance(i.w1, N.ndarray) assert isinstance(i.w1, N.ndarray)
assert (i.w1 == self.input_representations[0].w1).all() assert (i.w1 == self.input_representations[0].w1).all()
assert (i.w2 == self.input_representations[0].w2).all() assert (i.w2 == self.input_representations[0].w2).all()
...@@ -528,7 +528,7 @@ def create_realistic(window_size=3,#7, ...@@ -528,7 +528,7 @@ def create_realistic(window_size=3,#7,
def test_naacl_model(iters_per_unsup=3, iters_per_sup=3, def test_naacl_model(iters_per_unsup=3, iters_per_sup=3,
optimizer=None, realistic=False): optimizer=None, realistic=False):
print "BUILDING MODEL" #print "BUILDING MODEL"
import time import time
t = time.time() t = time.time()
...@@ -545,7 +545,7 @@ def test_naacl_model(iters_per_unsup=3, iters_per_sup=3, ...@@ -545,7 +545,7 @@ def test_naacl_model(iters_per_unsup=3, iters_per_sup=3,
else: else:
m = create(compile_mode=mode) m = create(compile_mode=mode)
print 'BUILD took %.3fs'%(time.time() - t) #print 'BUILD took %.3fs'%(time.time() - t)
prog_str = [] prog_str = []
idx_of_node = {} idx_of_node = {}
for i, node in enumerate(m.pretraining_update.maker.env.toposort()): for i, node in enumerate(m.pretraining_update.maker.env.toposort()):
...@@ -557,7 +557,7 @@ def test_naacl_model(iters_per_unsup=3, iters_per_sup=3, ...@@ -557,7 +557,7 @@ def test_naacl_model(iters_per_unsup=3, iters_per_sup=3,
#print input_pretraining_gradients[4].owner.inputs[1].owner.inputs #print input_pretraining_gradients[4].owner.inputs[1].owner.inputs
#sys.exit() #sys.exit()
print "PROGRAM LEN %i HASH %i"% (len(m.pretraining_update.maker.env.nodes), reduce(lambda a, b: hash(a) ^ hash(b),prog_str)) #print "PROGRAM LEN %i HASH %i"% (len(m.pretraining_update.maker.env.nodes), reduce(lambda a, b: hash(a) ^ hash(b),prog_str))
rng = N.random.RandomState(unittest_tools.fetch_seed(23904)) rng = N.random.RandomState(unittest_tools.fetch_seed(23904))
...@@ -565,35 +565,35 @@ def test_naacl_model(iters_per_unsup=3, iters_per_sup=3, ...@@ -565,35 +565,35 @@ def test_naacl_model(iters_per_unsup=3, iters_per_sup=3,
targets = N.asarray([0,3,4,2,3,4,4,2,1,0]) targets = N.asarray([0,3,4,2,3,4,4,2,1,0])
#print inputs #print inputs
print 'UNSUPERVISED PHASE' #print 'UNSUPERVISED PHASE'
t = time.time() t = time.time()
for i in xrange(3): for i in xrange(3):
for j in xrange(iters_per_unsup): for j in xrange(iters_per_unsup):
m.pretraining_update(*inputs) m.pretraining_update(*inputs)
s0, s1 = [str(j) for j in m.pretraining_update(*inputs)] s0, s1 = [str(j) for j in m.pretraining_update(*inputs)]
print 'huh?', i, iters_per_unsup, iters_per_unsup * (i+1), s0, s1 #print 'huh?', i, iters_per_unsup, iters_per_unsup * (i+1), s0, s1
if iters_per_unsup == 3: if iters_per_unsup == 3:
assert s0.startswith('0.927793')#'0.403044') assert s0.startswith('0.927793')#'0.403044')
assert s1.startswith('0.068035')#'0.074898') assert s1.startswith('0.068035')#'0.074898')
print 'UNSUPERVISED took %.3fs'%(time.time() - t) #print 'UNSUPERVISED took %.3fs'%(time.time() - t)
print 'FINETUNING GRAPH' #print 'FINETUNING GRAPH'
print 'SUPERVISED PHASE COSTS (%s)'%optimizer #print 'SUPERVISED PHASE COSTS (%s)'%optimizer
t = time.time() t = time.time()
for i in xrange(3): for i in xrange(3):
for j in xrange(iters_per_unsup): for j in xrange(iters_per_unsup):
m.finetuning_update(*(inputs + [targets])) m.finetuning_update(*(inputs + [targets]))
s0 = str(m.finetuning_update(*(inputs + [targets]))) s0 = str(m.finetuning_update(*(inputs + [targets])))
print iters_per_sup * (i+1), s0 #print iters_per_sup * (i+1), s0
if iters_per_sup == 10: if iters_per_sup == 10:
s0f = float(s0) s0f = float(s0)
assert 19.7042 < s0f and s0f < 19.7043 assert 19.7042 < s0f and s0f < 19.7043
print 'SUPERVISED took %.3fs'%( time.time() - t) #print 'SUPERVISED took %.3fs'%( time.time() - t)
def jtest_main(): def jtest_main():
from theano import gof from theano import gof
JTEST = theano.compile.mode.optdb.query(*sys.argv[2:]) JTEST = theano.compile.mode.optdb.query(*sys.argv[2:])
print 'JTEST', JTEST #print 'JTEST', JTEST
theano.compile.register_optimizer('JTEST', JTEST) theano.compile.register_optimizer('JTEST', JTEST)
optimizer = eval(sys.argv[1]) optimizer = eval(sys.argv[1])
test_naacl_model(optimizer, 10, 10, realistic=False) test_naacl_model(optimizer, 10, 10, realistic=False)
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论