コード例 #1
0
def predict(images_dir, model_dir, bottlenecks_dir, input_layer_name,
            output_layer_name, model_name):
    model_config = pretrain_model.get_model_config()

    image_lists = utils.create_image_lists(images_dir)

    with tf.Session() as sess:
        jpeg_data, decoded_image = pretrain_model.decode_jpeg(
            model_config["input_width"], model_config["input_height"],
            model_config["input_depth"], model_config["input_mean"],
            model_config["input_std"])

        bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_mobilenet_v2(
            model_dir, sess)

        load_graph(FLAGS.graph)

        test_bottlenecks, test_labels = (bottleneck.get_batch_of_bottlenecks(
            sess, image_lists, -1, "testing", bottlenecks_dir, images_dir,
            model_name, jpeg_data, decoded_image, resized_image, bottlenecks))

        softmax_tensor = sess.graph.get_tensor_by_name(output_layer_name)
        bottlenecks_input = sess.graph.get_tensor_by_name(input_layer_name)

        prediction = sess.run(softmax_tensor,
                              feed_dict={bottlenecks_input: test_bottlenecks})
        # print(sess.run(tf.shape(prediction)))

        y_pred = tf.placeholder(tf.float32, [None, len(image_lists.keys())])
        y_truth = tf.placeholder(tf.float32, [None, len(image_lists.keys())])

        correct_prediction = tf.equal(tf.argmax(y_pred, 1),
                                      tf.argmax(y_truth, 1))
        acc = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))

        cross_entropy = tf.reduce_mean(
            tf.nn.softmax_cross_entropy_with_logits(labels=y_truth,
                                                    logits=y_pred))

        accuracy, loss = sess.run([acc, cross_entropy],
                                  feed_dict={
                                      y_pred: prediction,
                                      y_truth: test_labels
                                  })

        print("test accuracy: {0:.7f}, test loss: {1:.7f}".format(
            accuracy, loss))

        pre_data_dir = os.path.join("../predictData", model_name)

        if not os.path.exists(pre_data_dir):
            os.makedirs(pre_data_dir)

        with h5py.File(os.path.join(pre_data_dir, "prediction_and_labels.h5"),
                       "w") as f:
            f["prediction"] = prediction
            f["truth"] = test_labels

    return 0
コード例 #2
0
def main(_):
    since = time.time()

    tf.logging.set_verbosity(tf.logging.INFO)

    # Create directories to store TensorBoard summaries
    if tf.gfile.Exists(FLAGS.summaries_dir):
        tf.gfile.DeleteRecursively(FLAGS.summaries_dir)
    tf.gfile.MakeDirs(FLAGS.summaries_dir)

    # Set model Hyperparameter
    model_config = pretrain_model.get_model_config()

    # Read the folder structure, and create lists of all the images
    image_lists = utils.create_image_lists(FLAGS.images_dir)
    class_count = len(image_lists.keys())

    if class_count == 0:
        tf.logging.error("No valid folders of images found at " + FLAGS.images_dir)
        return -1
    if class_count == 1:
        tf.logging.error("Only one valid folder of images found at " + FLAGS.images_dir)
        return -1

    # Create output_labels.txt displaying classes being trained
    with open(FLAGS.output_labels, "w") as f:
        f.write("\n".join(image_lists.keys()) + "\n")

    with tf.Session() as sess:
        # Set up the image decoding
        jpeg_data, decoded_image = pretrain_model.decode_jpeg(
            model_config["input_width"], model_config["input_height"],
            model_config["input_depth"], model_config["input_mean"],
            model_config["input_std"])

        # Load DenseNet model
        bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_mobilenet_v2(
            FLAGS.model_dir, sess)

        # store pretrained model bottlenecks
        bottleneck.store_bottlenecks(
            sess, image_lists, FLAGS.images_dir, FLAGS.bottlenecks_dir, model_name,
            jpeg_data, decoded_image, resized_image,
            bottlenecks, model=None)

        bottlenecks = tf.stop_gradient(bottlenecks)

        # Initialized final layer
        (train_step, cross_entropy, accuracy, bottlenecks_input, labels_input,
         final_result) = train.final_layer(
            len(image_lists.keys()), FLAGS.final_name, bottlenecks,
            bottlenecks_size, FLAGS.learning_rate)

        merged = tf.summary.merge_all()
        train_writer = tf.summary.FileWriter(FLAGS.summaries_dir + "/train",
                                             sess.graph)

        validation_writer = tf.summary.FileWriter(FLAGS.summaries_dir + "/validation")

        # Initialize all variables
        init = tf.global_variables_initializer()
        sess.run(init)

        # Get validation bottlenecks for evaluation
        validation_bottlenecks, validation_labels = (
            bottleneck.get_batch_of_bottlenecks(
                sess, image_lists, FLAGS.validation_batch_size, "validation",
                FLAGS.bottlenecks_dir, FLAGS.images_dir, model_name, jpeg_data,
                decoded_image, resized_image, bottlenecks))

        # Get test bottlenecks for evaluation
        test_bottlenecks, test_labels = (
            bottleneck.get_batch_of_bottlenecks(
                sess, image_lists, FLAGS.test_batch_size, "testing",
                FLAGS.bottlenecks_dir, FLAGS.images_dir, model_name, jpeg_data,
                decoded_image, resized_image, bottlenecks))

        best_acc = 0.0

        for i in range(FLAGS.iterations):
            # Get training bottlenecks
            (train_bottlenecks, train_labels) = bottleneck.get_batch_of_bottlenecks(
                sess, image_lists, FLAGS.train_batch_size, "training",
                FLAGS.bottlenecks_dir, FLAGS.images_dir, model_name, jpeg_data,
                decoded_image, resized_image, bottlenecks)
            # Training step
            train_summary, _ = sess.run(
                [merged, train_step],
                feed_dict={bottlenecks_input: train_bottlenecks,
                           labels_input: train_labels})
            train_writer.add_summary(train_summary, i)

            # Show evaluation based on specified frequency
            final_step = (i + 1 == FLAGS.iterations)
            if (i % FLAGS.eval_interval) == 0 or final_step:
                # Evaluation
                train_accuracy, train_loss = sess.run(
                    [accuracy, cross_entropy],
                    feed_dict={bottlenecks_input: train_bottlenecks,
                               labels_input: train_labels})

                # Run evaluation step on validation bottlenecks
                validation_summary, validation_accuracy, validation_loss = sess.run(
                    [merged, accuracy, cross_entropy],
                    feed_dict={bottlenecks_input: validation_bottlenecks,
                               labels_input: validation_labels})
                validation_writer.add_summary(validation_summary, i)

                # Save best accuracy and store model
                if validation_accuracy > best_acc:
                    best_acc = validation_accuracy

                    # Calculate the test accuracy with best validation on test bottlenecks
                    test_accuracy1, test_loss1 = sess.run(
                        [accuracy, cross_entropy],
                        feed_dict={bottlenecks_input: test_bottlenecks,
                                   labels_input: test_labels})

                    train.save_graph_to_file(sess, FLAGS.output_graph, FLAGS.final_name)
                    train.save_checkpoint_to_file(sess, FLAGS.output_checkpoint_dir)

                tf.logging.info("Iteration {}: train loss = {}, train acc = {}, val loss = {}, val acc = {}.".format(
                    i, train_loss, train_accuracy, validation_loss, validation_accuracy))

        # Calculate the final test accuracy on test bottlenecks
        test_accuracy2, test_loss2 = sess.run(
            [accuracy, cross_entropy],
            feed_dict={bottlenecks_input: test_bottlenecks,
                       labels_input: test_labels})

        tf.logging.info("Best validation accuracy = {}".format(best_acc * 100))
        tf.logging.info("Test accuracy with best validation =  {}".format(test_accuracy1 * 100))
        tf.logging.info("Final test accuracy =  {}".format(test_accuracy2 * 100))

    time_elapsed = time.time() - since

    print("Runtime: {}min, {:0.2f}sec".format(int(time_elapsed // 60), time_elapsed % 60))

    with open(os.path.join(path, model_name, "results.txt"), "w") as f:
        f.write("Best validation accuracy: " + str(best_acc) + "\n")
        f.write("Test accuracy with best validation: " + str(test_accuracy1) + "\n")
        f.write("Final test accuracy =  {}".format(test_accuracy2 * 100) + "\n")
        f.write("Runtime: " + str(int(time_elapsed // 60)) + "min," + str(time_elapsed % 60) + "sec")