예제 #1
0
def predict(images_dir, model_dir, bottlenecks_dir, input_layer_name,
            output_layer_name):
    model_config = pretrain_model.get_model_config()

    image_lists = utils.create_image_lists(images_dir)

    with tf.Session() as sess:
        jpeg_data, decoded_image = pretrain_model.decode_jpeg(
            model_config["input_width"], model_config["input_height"],
            model_config["input_depth"], model_config["input_mean"],
            model_config["input_std"])

        if (model_name == "inception_v3"):
            _, bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_inception_v3(
                FLAGS.model_dir)
        elif (model_name == "vgg16"):
            bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_vgg_16(
                model_dir, sess)
        elif (model_name == "resnet_v2_50"):
            bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_resnet_v2_50(
                model_dir, sess)
        elif (model_name == "resnet_v2_101"):
            bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_resnet_v2_101(
                model_dir, sess)
        elif (model_name == "resnet_v2_152"):
            bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_resnet_v2_152(
                model_dir, sess)
        elif (model_name == "mobilenet_v2"):
            bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_mobilenet_v2(
                model_dir, sess)
        elif (model_name == "densenet121"):
            densenet_model, bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_densenet_121(
                model_dir)
        elif (model_name == "densenet169"):
            densenet_model, bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_densenet_169(
                model_dir)
        elif (model_name == "densenet201"):
            densenet_model, bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_densenet_201(
                model_dir)

        load_graph(FLAGS.graph)

        test_bottlenecks, test_labels = (bottleneck.get_batch_of_bottlenecks(
            sess, image_lists, -1, "testing", bottlenecks_dir, images_dir,
            model_name, jpeg_data, decoded_image, resized_image, bottlenecks))

        softmax_tensor = sess.graph.get_tensor_by_name(output_layer_name)
        bottlenecks_input = sess.graph.get_tensor_by_name(input_layer_name)

        prediction = sess.run(softmax_tensor,
                              feed_dict={bottlenecks_input: test_bottlenecks})
        #print(sess.run(tf.shape(prediction)))

        y_pred = tf.placeholder(tf.float32, [None, len(image_lists.keys())])
        y_truth = tf.placeholder(tf.float32, [None, len(image_lists.keys())])

        correct_prediction = tf.equal(tf.argmax(y_pred, 1),
                                      tf.argmax(y_truth, 1))
        acc = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))

        cross_entropy = tf.reduce_mean(
            tf.nn.softmax_cross_entropy_with_logits(labels=y_truth,
                                                    logits=y_pred))

        accuracy, loss = sess.run([acc, cross_entropy],
                                  feed_dict={
                                      y_pred: prediction,
                                      y_truth: test_labels
                                  })

        print("test accuracy: {0:.7f}, test loss: {1:.7f}".format(
            accuracy, loss))

        pre_data_dir = os.path.join(path, "predictData", model_name)

        if not os.path.exists(pre_data_dir):
            os.makedirs(pre_data_dir)

        with h5py.File(os.path.join(pre_data_dir, "prediction_and_labels.h5"),
                       "w") as f:
            f["prediction"] = prediction
            f["truth"] = test_labels

    return 0
예제 #2
0
def main(_):
    since = time.time()

    tf.logging.set_verbosity(tf.logging.INFO)

    # Create directories to store TensorBoard summaries
    if tf.gfile.Exists(FLAGS.summaries_dir):
        tf.gfile.DeleteRecursively(FLAGS.summaries_dir)
    tf.gfile.MakeDirs(FLAGS.summaries_dir)

    # Set model Hyperparameter
    model_config = pretrain_model.get_model_config()

    # Read the folder structure, and create lists of all the images
    image_lists = utils.create_image_lists(FLAGS.images_dir)
    class_count = len(image_lists.keys())

    if class_count == 0:
        tf.logging.error("No valid folders of images found at " +
                         FLAGS.images_dir)
        return -1
    if class_count == 1:
        tf.logging.error("Only one valid folder of images found at " +
                         FLAGS.images_dir)
        return -1

    # Create output_labels.txt displaying classes being trained
    with open(FLAGS.output_labels, "w") as f:
        f.write("\n".join(image_lists.keys()) + "\n")

    with tf.Session() as sess:
        # Set up the image decoding
        jpeg_data, decoded_image = pretrain_model.decode_jpeg(
            model_config["input_width"], model_config["input_height"],
            model_config["input_depth"], model_config["input_mean"],
            model_config["input_std"])

        # Load DenseNet model
        bottlenecks, resized_image, bottlenecks_size = pretrain_model.load_vgg_16(
            FLAGS.model_dir, sess)

        # store pretrained model bottlenecks
        bottleneck.store_bottlenecks(sess,
                                     image_lists,
                                     FLAGS.images_dir,
                                     FLAGS.bottlenecks_dir,
                                     FLAGS.model_name,
                                     jpeg_data,
                                     decoded_image,
                                     resized_image,
                                     bottlenecks,
                                     model=None)

        bottlenecks = tf.stop_gradient(bottlenecks)

        global_step = tf.Variable(tf.constant(0), trainable=False)

        # Initialized final layer
        (train_step, cross_entropy, accuracy, bottlenecks_input, labels_input,
         final_result) = train.final_layer(len(image_lists.keys()),
                                           FLAGS.final_name, bottlenecks,
                                           bottlenecks_size,
                                           FLAGS.learning_rate, global_step)

        merged = tf.summary.merge_all()
        train_writer = tf.summary.FileWriter(FLAGS.summaries_dir + "/train",
                                             sess.graph)

        validation_writer = tf.summary.FileWriter(FLAGS.summaries_dir +
                                                  "/validation")

        # Initialize all variables
        init = tf.global_variables_initializer()
        sess.run(init)

        # Get validation bottlenecks for evaluation
        validation_bottlenecks, validation_labels = (
            bottleneck.get_batch_of_bottlenecks(
                sess, image_lists, FLAGS.validation_batch_size, "validation",
                FLAGS.bottlenecks_dir, FLAGS.images_dir, FLAGS.model_name,
                jpeg_data, decoded_image, resized_image, bottlenecks))

        # Get test bottlenecks for evaluation
        test_bottlenecks, test_labels = (bottleneck.get_batch_of_bottlenecks(
            sess, image_lists, FLAGS.test_batch_size, "testing",
            FLAGS.bottlenecks_dir, FLAGS.images_dir, FLAGS.model_name,
            jpeg_data, decoded_image, resized_image, bottlenecks))

        best_acc = 0.0

        for i in range(FLAGS.iterations):
            # Get training bottlenecks
            (train_bottlenecks,
             train_labels) = bottleneck.get_batch_of_bottlenecks(
                 sess, image_lists, FLAGS.train_batch_size, "training",
                 FLAGS.bottlenecks_dir, FLAGS.images_dir, FLAGS.model_name,
                 jpeg_data, decoded_image, resized_image, bottlenecks)
            # Training step
            train_summary, _ = sess.run(
                [merged, train_step],
                feed_dict={
                    bottlenecks_input: train_bottlenecks,
                    labels_input: train_labels,
                    global_step: i
                })
            train_writer.add_summary(train_summary, i)

            # Show evaluation based on specified frequency
            final_step = (i + 1 == FLAGS.iterations)
            if (i % FLAGS.eval_interval) == 0 or final_step:
                # Evaluation
                train_accuracy, train_loss = sess.run(
                    [accuracy, cross_entropy],
                    feed_dict={
                        bottlenecks_input: train_bottlenecks,
                        labels_input: train_labels
                    })

                # Run evaluation step on validation bottlenecks
                validation_summary, validation_accuracy, validation_loss = sess.run(
                    [merged, accuracy, cross_entropy],
                    feed_dict={
                        bottlenecks_input: validation_bottlenecks,
                        labels_input: validation_labels
                    })
                validation_writer.add_summary(validation_summary, i)

                # Save best accuracy and store model
                if validation_accuracy > best_acc:
                    best_acc = validation_accuracy

                    # Calculate the test accuracy with best validation on test bottlenecks
                    test_accuracy1, test_loss1 = sess.run(
                        [accuracy, cross_entropy],
                        feed_dict={
                            bottlenecks_input: test_bottlenecks,
                            labels_input: test_labels
                        })

                    train.save_graph_to_file(sess, FLAGS.output_graph,
                                             FLAGS.final_name)
                    train.save_checkpoint_to_file(sess,
                                                  FLAGS.output_checkpoint_dir)

                tf.logging.info(
                    "Iteration {}: train loss = {}, train acc = {}, val loss = {}, val acc = {}."
                    .format(i, train_loss, train_accuracy, validation_loss,
                            validation_accuracy))

        # Calculate the final test accuracy on test bottlenecks
        test_accuracy2, test_loss2 = sess.run([accuracy, cross_entropy],
                                              feed_dict={
                                                  bottlenecks_input:
                                                  test_bottlenecks,
                                                  labels_input: test_labels
                                              })

        tf.logging.info("Best validation accuracy = {}".format(best_acc * 100))
        tf.logging.info("Test accuracy with best validation =  {}".format(
            test_accuracy1 * 100))
        tf.logging.info("Final test accuracy =  {}".format(test_accuracy2 *
                                                           100))

    time_elapsed = time.time() - since

    print("Runtime: {}min, {:0.2f}sec".format(int(time_elapsed // 60),
                                              time_elapsed % 60))

    with open(os.path.join("..", FLAGS.model_name, "results.txt"), "w") as f:
        f.write("Best validation accuracy: " + str(best_acc) + "\n")
        f.write("Test accuracy with best validation: " + str(test_accuracy1) +
                "\n")
        f.write("Final test accuracy =  {}".format(test_accuracy2 * 100) +
                "\n")
        f.write("Runtime: " + str(int(time_elapsed // 60)) + "min," +
                str(time_elapsed % 60) + "sec")