text
stringlengths 0
93.6k
|
---|
for batch in utils.iterate_minibatches(X_train, Y_train, masks=mask_train, char_inputs=C_train,
|
batch_size=batch_size, shuffle=True):
|
inputs, targets, masks, char_inputs = batch
|
err, corr, num = train_fn(inputs, targets, masks, char_inputs)
|
train_err += err * inputs.shape[0]
|
train_corr += corr
|
train_total += num
|
train_inst += inputs.shape[0]
|
train_batches += 1
|
time_ave = (time.time() - start_time) / train_batches
|
time_left = (num_batches - train_batches) * time_ave
|
# update log
|
sys.stdout.write("\b" * num_back)
|
log_info = 'train: %d/%d loss: %.4f, acc: %.2f%%, time left (estimated): %.2fs' % (
|
min(train_batches * batch_size, num_data), num_data,
|
train_err / train_inst, train_corr * 100 / train_total, time_left)
|
sys.stdout.write(log_info)
|
num_back = len(log_info)
|
# update training log after each epoch
|
assert train_inst == num_data
|
sys.stdout.write("\b" * num_back)
|
print 'train: %d/%d loss: %.4f, acc: %.2f%%, time: %.2fs' % (
|
min(train_batches * batch_size, num_data), num_data,
|
train_err / num_data, train_corr * 100 / train_total, time.time() - start_time)
|
# evaluate performance on dev data
|
dev_err = 0.0
|
dev_corr = 0.0
|
dev_total = 0
|
dev_inst = 0
|
for batch in utils.iterate_minibatches(X_dev, Y_dev, masks=mask_dev, char_inputs=C_dev, batch_size=batch_size):
|
inputs, targets, masks, char_inputs = batch
|
err, corr, num, predictions = eval_fn(inputs, targets, masks, char_inputs)
|
dev_err += err * inputs.shape[0]
|
dev_corr += corr
|
dev_total += num
|
dev_inst += inputs.shape[0]
|
if output_predict:
|
utils.output_predictions(predictions, targets, masks, 'tmp/dev%d' % epoch, label_alphabet,
|
is_flattened=False)
|
print 'dev loss: %.4f, corr: %d, total: %d, acc: %.2f%%' % (
|
dev_err / dev_inst, dev_corr, dev_total, dev_corr * 100 / dev_total)
|
if best_loss < dev_err and best_acc > dev_corr / dev_total:
|
stop_count += 1
|
else:
|
update_loss = False
|
update_acc = False
|
stop_count = 0
|
if best_loss > dev_err:
|
update_loss = True
|
best_loss = dev_err
|
best_epoch_loss = epoch
|
if best_acc < dev_corr / dev_total:
|
update_acc = True
|
best_acc = dev_corr / dev_total
|
best_epoch_acc = epoch
|
# evaluate on test data when better performance detected
|
test_err = 0.0
|
test_corr = 0.0
|
test_total = 0
|
test_inst = 0
|
for batch in utils.iterate_minibatches(X_test, Y_test, masks=mask_test, char_inputs=C_test,
|
batch_size=batch_size):
|
inputs, targets, masks, char_inputs = batch
|
err, corr, num, predictions = eval_fn(inputs, targets, masks, char_inputs)
|
test_err += err * inputs.shape[0]
|
test_corr += corr
|
test_total += num
|
test_inst += inputs.shape[0]
|
if output_predict:
|
utils.output_predictions(predictions, targets, masks, 'tmp/test%d' % epoch, label_alphabet,
|
is_flattened=False)
|
print 'test loss: %.4f, corr: %d, total: %d, acc: %.2f%%' % (
|
test_err / test_inst, test_corr, test_total, test_corr * 100 / test_total)
|
if update_loss:
|
best_loss_test_err = test_err
|
best_loss_test_corr = test_corr
|
if update_acc:
|
best_acc_test_err = test_err
|
best_acc_test_corr = test_corr
|
# stop if dev acc decrease 3 time straightly.
|
if stop_count == patience:
|
break
|
# re-compile a function with new learning rate for training
|
if update_algo != 'adadelta':
|
lr = learning_rate / (1.0 + epoch * decay_rate)
|
updates = utils.create_updates(loss_train, params, update_algo, lr, momentum=momentum)
|
train_fn = theano.function([input_var, target_var, mask_var, char_input_var],
|
[loss_train, corr_train, num_tokens],
|
updates=updates)
|
# print best performance on test data.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.