-
Notifications
You must be signed in to change notification settings - Fork 198
/
predict.py
84 lines (67 loc) · 3.12 KB
/
predict.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
import os
import sys
import json
import logging
import data_helper
import numpy as np
import pandas as pd
import tensorflow as tf
from tensorflow.contrib import learn
logging.getLogger().setLevel(logging.INFO)
def predict_unseen_data():
"""Step 0: load trained model and parameters"""
params = json.loads(open('./parameters.json').read())
checkpoint_dir = sys.argv[1]
if not checkpoint_dir.endswith('/'):
checkpoint_dir += '/'
checkpoint_file = tf.train.latest_checkpoint(checkpoint_dir + 'checkpoints')
logging.critical('Loaded the trained model: {}'.format(checkpoint_file))
"""Step 1: load data for prediction"""
test_file = sys.argv[2]
test_examples = json.loads(open(test_file).read())
# labels.json was saved during training, and it has to be loaded during prediction
labels = json.loads(open('./labels.json').read())
one_hot = np.zeros((len(labels), len(labels)), int)
np.fill_diagonal(one_hot, 1)
label_dict = dict(zip(labels, one_hot))
x_raw = [example['consumer_complaint_narrative'] for example in test_examples]
x_test = [data_helper.clean_str(x) for x in x_raw]
logging.info('The number of x_test: {}'.format(len(x_test)))
y_test = None
if 'product' in test_examples[0]:
y_raw = [example['product'] for example in test_examples]
y_test = [label_dict[y] for y in y_raw]
logging.info('The number of y_test: {}'.format(len(y_test)))
vocab_path = os.path.join(checkpoint_dir, "vocab.pickle")
vocab_processor = learn.preprocessing.VocabularyProcessor.restore(vocab_path)
x_test = np.array(list(vocab_processor.transform(x_test)))
"""Step 2: compute the predictions"""
graph = tf.Graph()
with graph.as_default():
session_conf = tf.ConfigProto(allow_soft_placement=True, log_device_placement=False)
sess = tf.Session(config=session_conf)
with sess.as_default():
saver = tf.train.import_meta_graph("{}.meta".format(checkpoint_file))
saver.restore(sess, checkpoint_file)
input_x = graph.get_operation_by_name("input_x").outputs[0]
dropout_keep_prob = graph.get_operation_by_name("dropout_keep_prob").outputs[0]
predictions = graph.get_operation_by_name("output/predictions").outputs[0]
batches = data_helper.batch_iter(list(x_test), params['batch_size'], 1, shuffle=False)
all_predictions = []
for x_test_batch in batches:
batch_predictions = sess.run(predictions, {input_x: x_test_batch, dropout_keep_prob: 1.0})
all_predictions = np.concatenate([all_predictions, batch_predictions])
if y_test is not None:
y_test = np.argmax(y_test, axis=1)
correct_predictions = sum(all_predictions == y_test)
# Save the actual labels back to file
actual_labels = [labels[int(prediction)] for prediction in all_predictions]
for idx, example in enumerate(test_examples):
example['new_prediction'] = actual_labels[idx]
with open('./data/small_samples_prediction.json', 'w') as outfile:
json.dump(test_examples, outfile, indent=4)
logging.critical('The accuracy is: {}'.format(correct_predictions / float(len(y_test))))
logging.critical('The prediction is complete')
if __name__ == '__main__':
# python3 predict.py ./trained_model_1478649295/ ./data/small_samples.json
predict_unseen_data()