def __init__(self, args, seq_length, reuse=False):
        self.x = tf.placeholder(name='x', dtype=tf.float32, shape=[args.batch_size, seq_length, args.vector_dim])
        self.y = self.x
        eof = np.zeros([args.batch_size, args.vector_dim + 1])
        eof[:, args.vector_dim] = np.ones([args.batch_size])
        eof = tf.constant(eof, dtype=tf.float32)
        zero = tf.constant(np.zeros([args.batch_size, args.vector_dim + 1]), dtype=tf.float32)

        if args.model == 'LSTM':
            # single_cell = tf.nn.rnn_cell.BasicLSTMCell(args.rnn_size)
            # cannot use [single_cell] * 3 in tensorflow 1.2
            def rnn_cell(rnn_size):
                return tf.nn.rnn_cell.BasicLSTMCell(rnn_size, reuse=reuse)
            cell = tf.nn.rnn_cell.MultiRNNCell([rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)])
        elif args.model == 'NTM':
            import ntm.ntm_cell as ntm_cell
            cell = ntm_cell.NTMCell(args.rnn_size, args.memory_size, args.memory_vector_dim, 1, 1,
                                    addressing_mode='content_and_location',
                                    reuse=reuse,
                                    output_dim=args.vector_dim)

        state = cell.zero_state(args.batch_size, tf.float32)
        self.state_list = [state]
        for t in range(seq_length):
            output, state = cell(tf.concat([self.x[:, t, :], np.zeros([args.batch_size, 1])], axis=1), state)
            self.state_list.append(state)
        output, state = cell(eof, state)
        self.state_list.append(state)

        self.o = []
        for t in range(seq_length):
            output, state = cell(zero, state)
            self.o.append(output[:, 0:args.vector_dim])
            self.state_list.append(state)
        self.o = tf.sigmoid(tf.transpose(self.o, perm=[1, 0, 2]))

        # self.copy_loss = tf.reduce_mean(tf.reduce_sum(tf.square(self.y - self.o), reduction_indices=[1, 2]))
        eps = 1e-8
        self.copy_loss = -tf.reduce_mean(   # cross entropy function
            self.y * tf.log(self.o + eps) + (1 - self.y) * tf.log(1 - self.o + eps)
        )
        with tf.variable_scope('optimizer', reuse=reuse):
            self.optimizer = tf.train.RMSPropOptimizer(learning_rate=args.learning_rate, momentum=0.9, decay=0.95)
            gvs = self.optimizer.compute_gradients(self.copy_loss)
            capped_gvs = [(tf.clip_by_value(grad, -10., 10.), var) for grad, var in gvs]
            self.train_op = self.optimizer.apply_gradients(capped_gvs)
        self.copy_loss_summary = tf.summary.scalar('copy_loss_%d' % seq_length, self.copy_loss)
Esempio n. 2
0
    def __init__(self, args):
        if args.label_type == 'one_hot':
            args.output_dim = args.n_classes
        elif args.label_type == 'five_hot':
            args.output_dim = 25

        self.x_image = tf.placeholder(dtype=tf.float32, shape=[args.batch_size,
                                                               args.seq_length,
                                                               args.image_width * args.image_height])
        self.x_label = tf.placeholder(dtype=tf.float32, shape=[args.batch_size,
                                                               args.seq_length,
                                                               args.output_dim])
        self.y = tf.placeholder(dtype=tf.float32, shape=[args.batch_size,
                                                         args.seq_length,
                                                         args.output_dim])

        if args.model == 'LSTM':
            def rnn_cell(rnn_size):
                return tf.contrib.rnn.BasicLSTMCell(rnn_size)

            cell = tf.contrib.rnn.MultiRNNCell([rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)])
        elif args.model == 'NTM':
            import ntm.ntm_cell as ntm_cell
            cell = ntm_cell.NTMCell(args.rnn_size, args.memory_size, args.memory_vector_dim,
                                    read_head_num=args.read_head_num,
                                    write_head_num=args.write_head_num,
                                    addressing_mode='content_and_location',
                                    output_dim=args.output_dim)
        elif args.model == 'MANN':
            import ntm.mann_cell as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size, args.memory_size, args.memory_vector_dim,
                                      head_num=args.read_head_num)
        elif args.model == 'MANN2':
            import ntm.mann_cell_2 as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size, args.memory_size, args.memory_vector_dim,
                                      head_num=args.read_head_num)

        state = cell.zero_state(args.batch_size, tf.float32)
        self.state_list = [state]  # For debugging
        self.o = []
        for t in range(args.seq_length):
            output, state = cell(tf.concat([self.x_image[:, t, :], self.x_label[:, t, :]], axis=1), state)
            # output, state = cell(self.y[:, t, :], state)
            with tf.variable_scope("o2o", reuse=(t > 0)):
                o2o_w = tf.get_variable('o2o_w', [output.get_shape()[1], args.output_dim],
                                        initializer=tf.random_uniform_initializer(minval=-0.1, maxval=0.1))
                # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                o2o_b = tf.get_variable('o2o_b', [args.output_dim],
                                        initializer=tf.random_uniform_initializer(minval=-0.1, maxval=0.1))
                # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                output = tf.nn.xw_plus_b(output, o2o_w, o2o_b)
            if args.label_type == 'one_hot':
                output = tf.nn.softmax(output, dim=1)
            elif args.label_type == 'five_hot':
                output = tf.stack([tf.nn.softmax(o) for o in tf.split(output, 5, axis=1)], axis=1)
            self.o.append(output)
            self.state_list.append(state)
        self.o = tf.stack(self.o, axis=1)
        self.state_list.append(state)

        eps = 1e-8
        if args.label_type == 'one_hot':
            self.loss = -tf.reduce_mean(tf.reduce_sum(self.y * tf.log(self.o + eps), axis=[1, 2]))
        elif args.label_type == 'five_hot':
            self.loss = -tf.reduce_mean(
                tf.reduce_sum(tf.stack(tf.split(self.y, 5, axis=2), axis=2) * tf.log(self.o + eps)
                              , axis=[1, 2, 3]))
        self.o = tf.reshape(self.o, shape=[args.batch_size, args.seq_length, -1])
        self.loss_summary = tf.summary.scalar('Loss', self.loss)

        with tf.variable_scope('optimizer'):
            self.optimizer = tf.train.AdamOptimizer(learning_rate=args.learning_rate)
            self.train_op = self.optimizer.minimize(self.loss)
    def __init__(self, args):
        if args.label_type == 'one_hot':
            args.output_dim = args.n_classes
        elif args.label_type == 'five_hot':
            args.output_dim = 25

        eprint("Creating Placeholders")
        if args.dataset_type == 'omniglot':
            self.x_image = tf.placeholder(dtype=tf.float32,
                                          shape=[args.batch_size, args.seq_length, args.image_width * args.image_height])
        elif args.dataset_type == 'kinetics_dynamic':
            self.x_image = tf.placeholder(dtype=tf.float32,
                                          shape=[args.batch_size, args.seq_length, args.image_width, args.image_height, 3])
        elif args.dataset_type == 'kinetics_video':
            #  self.x_image = tf.placeholder(dtype=tf.float32,
                                          #  shape=[args.batch_size, args.seq_length, args.sample_nframes, args.image_width, args.image_height, 3])
            self.x_image = tf.placeholder(dtype=tf.float32,
                                          shape=[args.batch_size, args.seq_length, args.sample_nframes, args.image_width, args.image_height, 3])
        elif args.dataset_type == 'kinetics_single_frame':
            self.x_image = tf.placeholder(dtype=tf.float32,
                                          shape=[args.batch_size, args.seq_length, args.image_width, args.image_height, 3])


        self.is_training = tf.placeholder(tf.bool)

        self.x_label = tf.placeholder(dtype=tf.float32,
                                      shape=[args.batch_size, args.seq_length, args.output_dim])
        self.y = tf.placeholder(dtype=tf.float32,
                                shape=[args.batch_size, args.seq_length, args.output_dim])

        eprint("Creating Cells")
        if args.model == 'LSTM':
            def rnn_cell(rnn_size):
                return tf.nn.rnn_cell.BasicLSTMCell(rnn_size)
            cell = tf.nn.rnn_cell.MultiRNNCell([rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)])
        elif args.model == 'NTM':
            import ntm.ntm_cell as ntm_cell
            cell = ntm_cell.NTMCell(args.rnn_size, args.memory_size, args.memory_vector_dim,
                                    read_head_num=args.read_head_num,
                                    write_head_num=args.write_head_num,
                                    addressing_mode='content_and_location',
                                    output_dim=args.output_dim)
        elif args.model == 'MANN':
            import ntm.mann_cell as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size, args.memory_size, args.memory_vector_dim, is_training=self.is_training,
                                    head_num=args.read_head_num, args=args)
        elif args.model == 'MANN2':
            import ntm.mann_cell_2 as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size, args.memory_size, args.memory_vector_dim,
                                    head_num=args.read_head_num)

        eprint("Looping through seq")
        # Zero out the memory state
        state = cell.zero_state(args.batch_size, tf.float32)
        self.state_list = [state]   # For debugging. Keep track of previous states
        self.o = []
        for t in range(args.seq_length):
            # So this is going and calling __call__
            # it is passing both the image and x_label
            #  output, state = cell(tf.concat([self.x_image[:, t, :], self.x_label[:, t, :]], axis=1), state)
            #  output, state = cell(self.x_image[:, t, :], state)
            # Q: the labels are passed in here as part of the input....
            eprint("Seq: [{}] Passing data into the cell".format(t))
            if args.dataset_type == 'omniglot':
                output, state = cell(self.x_image[:, t, :], self.x_label[:, t, :], state)
            elif args.dataset_type == 'kinetics_video':
                output, state = cell(self.x_image[:, t, :, :, :, :], self.x_label[:, t, :], state)
            else:
                # in the case of kinetics dynamic.. aka default
                output, state = cell(self.x_image[:, t, :, :, :], self.x_label[:, t, :], state)

            # output, state = cell(self.y[:, t, :], state)
            # go from the memory stored dimensionality to the number of classes / predictions
            with tf.variable_scope("o2o", reuse=(t > 0)):
                o2o_w = tf.get_variable('o2o_w', [output.get_shape()[1], args.output_dim],
                                        initializer=tf.random_uniform_initializer(minval=-0.1, maxval=0.1))
                                        # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                o2o_b = tf.get_variable('o2o_b', [args.output_dim],
                                        initializer=tf.random_uniform_initializer(minval=-0.1, maxval=0.1))
                                        # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                # matmul + bias
                output = tf.nn.xw_plus_b(output, o2o_w, o2o_b)
            if args.label_type == 'one_hot':
                # Made a change here for making it dim = -1
                output = tf.nn.softmax(output, dim=-1)
            elif args.label_type == 'five_hot':
                output = tf.stack([tf.nn.softmax(o) for o in tf.split(output, 5, axis=1)], axis=1)
            self.o.append(output)
            self.state_list.append(state)
        self.o = tf.stack(self.o, axis=1)
        self.state_list.append(state)

        eprint("Defining Loss")
        eps = 1e-8
        if args.label_type == 'one_hot':
            self.learning_loss = -tf.reduce_mean(  # cross entropy function
                tf.reduce_sum(self.y * tf.log(self.o + eps), axis=[1, 2])
            )
        elif args.label_type == 'five_hot':
            self.learning_loss = -tf.reduce_mean(  # cross entropy function
                tf.reduce_sum(tf.stack(tf.split(self.y, 5, axis=2), axis=2) * tf.log(self.o + eps), axis=[1, 2, 3])
            )
        self.o = tf.reshape(self.o, shape=[args.batch_size, args.seq_length, -1])
        self.learning_loss_summary = tf.summary.scalar('learning_loss', self.learning_loss)

        eprint( "Total number of variables used ", np.sum([v.get_shape().num_elements() for v in tf.trainable_variables()]) )
        eprint("Defining optimizer")
        with tf.variable_scope('optimizer'):
            if args.optimizer == 'adam':
                self.optimizer = tf.train.AdamOptimizer(learning_rate=args.learning_rate)
            elif args.optimizer == 'rms':
                self.optimizer = tf.train.RMSPropOptimizer(learning_rate=args.learning_rate)
            else:
                self.optimizer = tf.train.GradientDescentOptimizer(learning_rate=args.learning_rate)
            # self.optimizer = tf.train.RMSPropOptimizer(
            #     learning_rate=args.learning_rate, momentum=0.9, decay=0.95
            # )
            # gvs = self.optimizer.compute_gradients(self.learning_loss)
            # capped_gvs = [(tf.clip_by_value(grad, -10., 10.), var) for grad, var in gvs]
            # self.train_op = self.optimizer.apply_gradients(gvs)
            self.train_op = self.optimizer.minimize(self.learning_loss)

        eprint("Finished Definining Model")
Esempio n. 4
0
    def __init__(self, args):
        if args.label_type == 'one_hot':
            args.output_dim = args.n_classes
        elif args.label_type == 'five_hot':
            args.output_dim = 25

        self.x_image = tf.placeholder(dtype=tf.float32,
                                      shape=[
                                          args.batch_size, args.seq_length,
                                          args.image_width * args.image_height
                                      ])
        self.x_label = tf.placeholder(
            dtype=tf.float32,
            shape=[args.batch_size, args.seq_length, args.output_dim])
        self.y = tf.placeholder(
            dtype=tf.float32,
            shape=[args.batch_size, args.seq_length, args.output_dim])

        if args.model == 'LSTM':

            def rnn_cell(rnn_size):
                return tf.nn.rnn_cell.BasicLSTMCell(rnn_size)

            cell = tf.nn.rnn_cell.MultiRNNCell(
                [rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)])
        elif args.model == 'NTM':
            import ntm.ntm_cell as ntm_cell
            cell = ntm_cell.NTMCell(args.rnn_size,
                                    args.memory_size,
                                    args.memory_vector_dim,
                                    read_head_num=args.read_head_num,
                                    write_head_num=args.write_head_num,
                                    addressing_mode='content_and_location',
                                    output_dim=args.output_dim)
        elif args.model == 'MANN':
            import ntm.mann_cell as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size,
                                      args.memory_size,
                                      args.memory_vector_dim,
                                      head_num=args.read_head_num)
        elif args.model == 'MANN2':
            import ntm.mann_cell_2 as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size,
                                      args.memory_size,
                                      args.memory_vector_dim,
                                      head_num=args.read_head_num)
        elif args.model == 'ACT':
            from tf_rnn_adaptive.act_wrapper import ACTWrapper

            def rnn_cell(rnn_size):
                return tf.nn.rnn_cell.BasicLSTMCell(rnn_size)

            # inner_cell = tf.nn.rnn_cell.MultiRNNCell([rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)])
            inner_cell = rnn_cell(args.rnn_size)
            cell = ACTWrapper(inner_cell, ponder_limit=10)
        elif args.model == 'DNC':
            from dnc import dnc
            access_config = {
                "memory_size": args.memory_size,
                "word_size": args.memory_vector_dim,
                "num_reads": args.read_head_num,
                "num_writes": args.write_head_num,
            }
            controller_config = {
                "hidden_size": args.rnn_size,
            }
            clip_value = args.clip_value
            cell = dnc.DNC(access_config, controller_config, args.output_dim,
                           clip_value)
        elif args.model == 'ACT-DNC':
            from tf_rnn_adaptive.act_wrapper import ACTWrapper
            from dnc import dnc
            access_config = {
                "memory_size": args.memory_size,
                "word_size": args.memory_vector_dim,
                "num_reads": args.read_head_num,
                "num_writes": args.write_head_num,
            }
            controller_config = {
                "hidden_size": args.rnn_size,
            }
            clip_value = args.clip_value

            dnc_core = dnc.DNC(access_config, controller_config,
                               args.output_dim, clip_value)
            cell = ACTWrapper(dnc_core, ponder_limit=10)
        elif args.model == 'MY-ACT':
            from my_act.act_wrapper import ACTWrapper
            from dnc import dnc

            # main dnc
            with tf.variable_scope('act_wrapper'):
                access_config = {
                    "memory_size": args.memory_size,
                    "word_size": args.memory_vector_dim,
                    "num_reads": args.read_head_num,
                    "num_writes": args.write_head_num,
                }
                controller_config = {
                    "hidden_size": args.rnn_size,
                }
                clip_value = args.clip_value
                main_dnc = dnc.DNC(access_config,
                                   controller_config,
                                   args.output_dim,
                                   clip_value,
                                   name='main_dnc')

            # auxiliary dnc
            # use memory_vector_dim as aux's output size
            # so that info does not lose when communicating with main dnc
            with tf.variable_scope('act_wrapper'):
                aux_access_config = {
                    "memory_size": 8,
                    "word_size": args.memory_vector_dim,
                    "num_reads": 1,
                    "num_writes": 1
                }
                aux_controller_config = {
                    "hidden_size": args.rnn_size,
                }
                aux_dnc = dnc.DNC(aux_access_config,
                                  aux_controller_config,
                                  args.memory_vector_dim,
                                  clip_value,
                                  name='aux_dnc')

            cell = ACTWrapper(main_dnc,
                              aux_dnc,
                              ponder_limit=10,
                              divergence_type=args.divergence_loss)
        else:
            raise Exception('Unknown model: `{}`'.format(args.model))

        if args.model == 'ACT-DNC':
            state = dnc_core.initial_state(args.batch_size)
        elif args.model == 'DNC':
            state = cell.initial_state(args.batch_size)
        elif args.model == 'MY-ACT':
            state = main_dnc.initial_state(args.batch_size)
        else:
            state = cell.zero_state(args.batch_size, tf.float32)
        self.state_list = [state]  # For debugging
        self.o = []
        for t in range(args.seq_length):
            output, state = cell(
                tf.concat([self.x_image[:, t, :], self.x_label[:, t, :]],
                          axis=1), state)
            # output, state = cell(self.y[:, t, :], state)
            with tf.variable_scope("o2o", reuse=(t > 0)):
                o2o_w = tf.get_variable(
                    'o2o_w', [output.get_shape()[1], args.output_dim],
                    initializer=tf.random_uniform_initializer(minval=-0.1,
                                                              maxval=0.1))
                # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                o2o_b = tf.get_variable(
                    'o2o_b', [args.output_dim],
                    initializer=tf.random_uniform_initializer(minval=-0.1,
                                                              maxval=0.1))
                # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                output = tf.nn.xw_plus_b(output, o2o_w, o2o_b)
            if args.label_type == 'one_hot':
                output = tf.nn.softmax(output, dim=1)
            elif args.label_type == 'five_hot':
                output = tf.stack(
                    [tf.nn.softmax(o) for o in tf.split(output, 5, axis=1)],
                    axis=1)
            self.o.append(output)
            self.state_list.append(state)
        self.o = tf.stack(self.o, axis=1)
        self.state_list.append(state)

        eps = 1e-8
        if args.label_type == 'one_hot':
            self.learning_loss = -tf.reduce_mean(  # cross entropy function
                tf.reduce_sum(self.y * tf.log(self.o + eps), axis=[1, 2]))
        elif args.label_type == 'five_hot':
            self.learning_loss = -tf.reduce_mean(  # cross entropy function
                tf.reduce_sum(tf.stack(tf.split(self.y, 5, axis=2), axis=2) *
                              tf.log(self.o + eps),
                              axis=[1, 2, 3]))
        self.o = tf.reshape(self.o,
                            shape=[args.batch_size, args.seq_length, -1])
        self.learning_loss_summary = tf.summary.scalar('learning_loss',
                                                       self.learning_loss)
        """ ponder loss """
        if 'ACT' in args.model:
            time_penalty = 0.001
            self._ponder_loss = time_penalty * cell.get_ponder_cost(
                args.seq_length)
            self.learning_loss += self._ponder_loss
            self.ponder_steps = cell.get_ponder_steps(args.seq_length)
            self.mean_ponder_steps = tf.reduce_mean(self.ponder_steps)
        else:
            self.mean_ponder_steps = tf.constant(0)
        """ memory divergence loss """
        if args.model == 'MY-ACT':
            if args.divergence_loss is not None:
                divergence_penalty = 0.001
                self.learning_loss += cell.get_memory_divergence_loss(
                ) * divergence_penalty

        with tf.variable_scope('optimizer'):
            self.optimizer = tf.train.AdamOptimizer(
                learning_rate=args.learning_rate)
            # self.optimizer = tf.train.RMSPropOptimizer(
            #     learning_rate=args.learning_rate, momentum=0.9, decay=0.95
            # )
            # gvs = self.optimizer.compute_gradients(self.learning_loss)
            # capped_gvs = [(tf.clip_by_value(grad, -10., 10.), var) for grad, var in gvs]
            # self.train_op = self.optimizer.apply_gradients(gvs)
            self.train_op = self.optimizer.minimize(self.learning_loss)
Esempio n. 5
0
    def __init__(self, args):

        self.x_data = tf.placeholder(dtype=tf.float32,
                                     shape=[args.batch_size, args.seq_length],
                                     name="x_squences")
        self.x_label = tf.placeholder(dtype=tf.float32,
                                      shape=[args.batch_size, args.output_dim],
                                      name="x_label")
        self.y = tf.placeholder(dtype=tf.float32,
                                shape=[args.batch_size, args.output_dim],
                                name="y")

        if args.model == 'LSTM':

            def rnn_cell(rnn_size):
                return tf.nn.rnn_cell.BasicLSTMCell(rnn_size)

            cell = tf.nn.rnn_cell.MultiRNNCell(
                [rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)])
        elif args.model == 'NTM':
            import ntm.ntm_cell as ntm_cell
            cell = ntm_cell.NTMCell(args.rnn_size,
                                    args.memory_size,
                                    args.memory_vector_dim,
                                    read_head_num=args.read_head_num,
                                    write_head_num=args.write_head_num,
                                    addressing_mode='content_and_location',
                                    output_dim=args.output_dim)
        elif args.model == 'MANN':
            import ntm.mann_cell as mann_cell
            cell = mann_cell.MANNCell(rnn_size=args.rnn_size,
                                      memory_size=args.memory_size,
                                      memory_vector_dim=args.memory_vector_dim,
                                      head_num=args.read_head_num,
                                      rnn_layers=args.rnn_num_layers)

        state = cell.zero_state(args.batch_size, tf.float32)
        self.state_list = [state]  # For debugging
        b = tf.concat([self.x_data[:, :], self.x_label[:, :]], axis=1)
        output, state = cell(b, state)
        with tf.variable_scope("o2o", reuse=tf.AUTO_REUSE):
            o2o_w = tf.get_variable(
                'o2o_w', [output.get_shape()[1], args.output_dim],
                initializer=tf.random_uniform_initializer(minval=-0.1,
                                                          maxval=0.1))
            o2o_b = tf.get_variable('o2o_b', [args.output_dim],
                                    initializer=tf.random_uniform_initializer(
                                        minval=-0.1, maxval=0.1))
            output = tf.nn.xw_plus_b(output, o2o_w, o2o_b)

        output = tf.nn.softmax(output)
        self.state_list.append(state)
        self.o = tf.squeeze(output, name="output")

        eps = 1e-8
        self.learning_loss = -tf.reduce_mean(  # cross entropy function
            tf.reduce_sum(self.y * tf.log(self.o + eps), axis=[1]))
        self.accuracy, self.acc_op = tf.metrics.accuracy(
            labels=tf.argmax(self.y, 1),
            predictions=tf.argmax(self.o, 1),
            name="accuracy")
        self.recall, self.rec_op = tf.metrics.recall_at_k(labels=tf.cast(
            self.y, tf.int64),
                                                          predictions=self.o,
                                                          k=100)
        self.precision, self.pre_op = tf.metrics.precision(
            labels=tf.argmax(self.y, 1),
            predictions=tf.argmax(self.o, 1),
            name="precision")

        tf.summary.scalar('learning_loss', self.learning_loss)
        tf.summary.scalar('Accuracy', self.accuracy)
        tf.summary.scalar('Recall_k', self.recall)
        tf.summary.scalar('precision', self.precision)
        self.merged_summary_op = tf.summary.merge_all()

        with tf.variable_scope('optimizer'):
            self.optimizer = tf.train.AdamOptimizer(
                learning_rate=args.learning_rate)
            # self.optimizer = tf.train.RMSPropOptimizer(
            #     learning_rate=args.learning_rate, momentum=0.9, decay=0.95
            # )
            # gvs = self.optimizer.compute_gradients(self.learning_loss)
            # capped_gvs = [(tf.clip_by_value(grad, -10., 10.), var) for grad, var in gvs]
            # self.train_op = self.optimizer.apply_gradients(gvs)
            self.train_op = self.optimizer.minimize(self.learning_loss,
                                                    name="train_op")
Esempio n. 6
0
    def __init__(self, args):
        if args.label_type == 'one_hot':
            args.output_dim = args.n_classes
        elif args.label_type == 'five_hot':
            args.output_dim = 25

        self.x_image = tf.placeholder(dtype=tf.float32,
                                      shape=[
                                          args.batch_size, args.seq_length,
                                          args.image_width * args.image_height
                                      ])
        self.x_label = tf.placeholder(
            dtype=tf.float32,
            shape=[args.batch_size, args.seq_length, args.output_dim])
        self.y = tf.placeholder(
            dtype=tf.float32,
            shape=[args.batch_size, args.seq_length, args.output_dim])

        if args.model == 'LSTM':

            def rnn_cell(rnn_size):
                return tf.nn.rnn_cell.BasicLSTMCell(rnn_size)

            cell = tf.nn.rnn_cell.MultiRNNCell(
                [rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)])
        elif args.model == 'NTM':
            import ntm.ntm_cell as ntm_cell
            cell = ntm_cell.NTMCell(args.rnn_size,
                                    args.memory_size,
                                    args.memory_vector_dim,
                                    read_head_num=args.read_head_num,
                                    write_head_num=args.write_head_num,
                                    addressing_mode='content_and_location',
                                    output_dim=args.output_dim)
        elif args.model == 'MANN':
            import ntm.mann_cell as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size,
                                      args.memory_size,
                                      args.memory_vector_dim,
                                      head_num=args.read_head_num)
        elif args.model == 'MANN2':
            import ntm.mann_cell_2 as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size,
                                      args.memory_size,
                                      args.memory_vector_dim,
                                      head_num=args.read_head_num)

        # Zero out the memory state
        state = cell.zero_state(args.batch_size, tf.float32)
        self.state_list = [state
                           ]  # For debugging. Keep track of previous states
        self.o = []
        for t in range(args.seq_length):
            # So this is going and calling __call__
            # it is passing both the image and x_label
            output, state = cell(
                tf.concat([self.x_image[:, t, :], self.x_label[:, t, :]],
                          axis=1), state)
            # output, state = cell(self.y[:, t, :], state)
            # go from the memory stored dimensionality to the number of classes / predictions
            with tf.variable_scope("o2o", reuse=(t > 0)):
                o2o_w = tf.get_variable(
                    'o2o_w', [output.get_shape()[1], args.output_dim],
                    initializer=tf.random_uniform_initializer(minval=-0.1,
                                                              maxval=0.1))
                # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                o2o_b = tf.get_variable(
                    'o2o_b', [args.output_dim],
                    initializer=tf.random_uniform_initializer(minval=-0.1,
                                                              maxval=0.1))
                # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                # matmul + bias
                output = tf.nn.xw_plus_b(output, o2o_w, o2o_b)
            if args.label_type == 'one_hot':
                output = tf.nn.softmax(output, dim=1)
            elif args.label_type == 'five_hot':
                output = tf.stack(
                    [tf.nn.softmax(o) for o in tf.split(output, 5, axis=1)],
                    axis=1)
            self.o.append(output)
            self.state_list.append(state)
        self.o = tf.stack(self.o, axis=1)
        self.state_list.append(state)

        eps = 1e-8
        if args.label_type == 'one_hot':
            self.learning_loss = -tf.reduce_mean(  # cross entropy function
                tf.reduce_sum(self.y * tf.log(self.o + eps), axis=[1, 2]))
        elif args.label_type == 'five_hot':
            self.learning_loss = -tf.reduce_mean(  # cross entropy function
                tf.reduce_sum(tf.stack(tf.split(self.y, 5, axis=2), axis=2) *
                              tf.log(self.o + eps),
                              axis=[1, 2, 3]))
        self.o = tf.reshape(self.o,
                            shape=[args.batch_size, args.seq_length, -1])
        self.learning_loss_summary = tf.summary.scalar('learning_loss',
                                                       self.learning_loss)

        with tf.variable_scope('optimizer'):
            self.optimizer = tf.train.AdamOptimizer(
                learning_rate=args.learning_rate)
            # self.optimizer = tf.train.RMSPropOptimizer(
            #     learning_rate=args.learning_rate, momentum=0.9, decay=0.95
            # )
            # gvs = self.optimizer.compute_gradients(self.learning_loss)
            # capped_gvs = [(tf.clip_by_value(grad, -10., 10.), var) for grad, var in gvs]
            # self.train_op = self.optimizer.apply_gradients(gvs)
            self.train_op = self.optimizer.minimize(self.learning_loss)
Esempio n. 7
0
    def __init__(self, args):
        if args.label_type == 'one_hot':
            args.output_dim = args.n_classes
        elif args.label_type == 'five_hot':
            args.output_dim = 25

        self.x_image = tf.placeholder(dtype=tf.float32,
                                      shape=[
                                          args.batch_size, args.seq_length,
                                          args.image_width * args.image_height
                                      ])
        self.x_label = tf.placeholder(
            dtype=tf.float32,
            shape=[args.batch_size, args.seq_length, args.output_dim])
        self.y = tf.placeholder(
            dtype=tf.float32,
            shape=[args.batch_size, args.seq_length, args.output_dim])

        if args.model == 'LSTM':

            def rnn_cell(rnn_size):
                return tf.nn.rnn_cell.BasicLSTMCell(rnn_size)

            cell = tf.nn.rnn_cell.MultiRNNCell(
                [rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)])
        elif args.model == 'NTM':
            import ntm.ntm_cell as ntm_cell
            cell = ntm_cell.NTMCell(args.rnn_size,
                                    args.memory_size,
                                    args.memory_vector_dim,
                                    read_head_num=args.read_head_num,
                                    write_head_num=args.write_head_num,
                                    addressing_mode='content_and_location',
                                    output_dim=args.output_dim)
        elif args.model == 'MANN':  #this is the standard MANN cell
            import ntm.mann_cell as mann_cell  #here no seperate write heads
            cell = mann_cell.MANNCell(
                args.rnn_size,
                args.memory_size,
                args.memory_vector_dim,  #initalizing the memory cell.
                head_num=args.read_head_num)
        elif args.model == 'MANN2':
            import ntm.mann_cell_2 as mann_cell
            cell = mann_cell.MANNCell(args.rnn_size,
                                      args.memory_size,
                                      args.memory_vector_dim,
                                      head_num=args.read_head_num)

        state = cell.zero_state(
            args.batch_size,
            tf.float32)  #Get the zero state or initialize the state
        self.state_list = [state]  # For debugging keep the zero state
        self.o = []  #for the output
        for t in range(args.seq_length):  #till the end of the sequence
            #here the x label should be time shifted, one step shifted
            output, state = cell(
                tf.concat([self.x_image[:, t, :], self.x_label[:, t, :]],
                          axis=1), state
            )  #call function in the MANN cell. get the output and the state dictionar
            # output, state = cell(self.y[:, t, :], state)
            with tf.variable_scope("o2o", reuse=(
                    t > 0
            )):  #sending the output via a fully connected to get real output
                o2o_w = tf.get_variable(
                    'o2o_w', [output.get_shape()[1], args.output_dim],
                    initializer=tf.random_uniform_initializer(minval=-0.1,
                                                              maxval=0.1))
                # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                o2o_b = tf.get_variable(
                    'o2o_b', [args.output_dim],
                    initializer=tf.random_uniform_initializer(minval=-0.1,
                                                              maxval=0.1))
                # initializer=tf.random_normal_initializer(mean=0.0, stddev=0.1))
                output = tf.nn.xw_plus_b(
                    output, o2o_w, o2o_b
                )  #output is the cell output for each time step them we use fully connectd layers
            if args.label_type == 'one_hot':
                output = tf.nn.softmax(output, dim=1)  #softmax
            elif args.label_type == 'five_hot':
                output = tf.stack(
                    [tf.nn.softmax(o) for o in tf.split(output, 5, axis=1)],
                    axis=1)
            self.o.append(output)  #stacking the outputs
            self.state_list.append(
                state)  #keeping the states in the state list
        self.o = tf.stack(self.o, axis=1)
        self.state_list.append(state)  #get the final state and stack it

        eps = 1e-8
        if args.label_type == 'one_hot':
            self.learning_loss = -tf.reduce_mean(  # cross entropy function
                tf.reduce_sum(self.y * tf.log(self.o + eps), axis=[1, 2]))
        elif args.label_type == 'five_hot':
            self.learning_loss = -tf.reduce_mean(  # cross entropy function
                tf.reduce_sum(tf.stack(tf.split(self.y, 5, axis=2), axis=2) *
                              tf.log(self.o + eps),
                              axis=[1, 2, 3]))
        self.o = tf.reshape(self.o,
                            shape=[args.batch_size, args.seq_length, -1])
        self.learning_loss_summary = tf.summary.scalar('learning_loss',
                                                       self.learning_loss)

        with tf.variable_scope('optimizer'):
            self.optimizer = tf.train.AdamOptimizer(
                learning_rate=args.learning_rate)  #optimizing the loss
            # self.optimizer = tf.train.RMSPropOptimizer(
            #     learning_rate=args.learning_rate, momentum=0.9, decay=0.95
            # )
            # gvs = self.optimizer.compute_gradients(self.learning_loss)
            # capped_gvs = [(tf.clip_by_value(grad, -10., 10.), var) for grad, var in gvs]
            # self.train_op = self.optimizer.apply_gradients(gvs)
            self.train_op = self.optimizer.minimize(
                self.learning_loss)  #optimizing the loss
Esempio n. 8
0
    def __init__(self, args, seq_length, reuse=False):
        self.x = tf.placeholder(
            name='x',
            dtype=tf.float32,
            shape=[args.batch_size, seq_length,
                   args.vector_dim])  #input placeh
        self.y = self.x  #output placeholder here this is 3 dimentional
        eof = np.zeros([args.batch_size,
                        args.vector_dim + 1])  #this is end of sequence

        eof[:, args.vector_dim] = np.ones([args.batch_size])
        eof = tf.constant(eof, dtype=tf.float32)
        zero = tf.constant(np.zeros([args.batch_size, args.vector_dim + 1]),
                           dtype=tf.float32)

        if args.model == 'LSTM':
            # single_cell = tf.nn.rnn_cell.BasicLSTMCell(args.rnn_size)
            # cannot use [single_cell] * 3 in tensorflow 1.2
            def rnn_cell(rnn_size):
                return tf.nn.rnn_cell.BasicLSTMCell(
                    rnn_size, reuse=reuse)  #rnn cell structure

            cell = tf.nn.rnn_cell.MultiRNNCell([
                rnn_cell(args.rnn_size) for _ in range(args.rnn_num_layers)
            ])  #multi layer rnn
        elif args.model == 'NTM':  #This cell is a neural machine translation cell. Which can output a controller output(LSTM cell) then
            import ntm.ntm_cell as ntm_cell
            cell = ntm_cell.NTMCell(
                args.rnn_size,
                args.memory_size,
                args.memory_vector_dim,
                1,
                1,  #output a cell object
                addressing_mode='content_and_location',
                reuse=reuse,
                output_dim=args.vector_dim)


#this is like the encoder
        state = cell.zero_state(
            args.batch_size, tf.float32
        )  #This is like the initialized state which contains a dictionalY of controller state , memmory vecotr , read head and write head
        self.state_list = [state]  #this is the zero state
        for t in range(seq_length):  #for each time step we send these things
            output, state = cell(
                tf.concat([self.x[:, t, :],
                           np.zeros([args.batch_size, 1])],
                          axis=1), state)  #feeding forward throug the RNN
            self.state_list.append(state)
        output, state = cell(eof, state)  #last state
        self.state_list.append(
            state)  #state is a big dictiory . Each time step we keep it

        #this is the decoder
        self.o = []  #outpts at each time steps should save here
        for t in range(seq_length):  #output at each time step

            output, state = cell(
                zero, state)  #here in the decoder there is no input to feed

            self.o.append(output[:, 0:args.vector_dim])
            self.state_list.append(
                state)  #also keep the hidden state of the decorder
        self.o = tf.sigmoid(
            tf.transpose(self.o, perm=[1, 0, 2])
        )  #getting the sigmoid since they should give binary values  #binary output at each time teps are stored this is 8 bit

        # self.copy_loss = tf.reduce_mean(tf.reduce_sum(tf.square(self.y - self.o), reduction_indices=[1, 2]))
        eps = 1e-8
        self.copy_loss = -tf.reduce_mean(  # cross entropy function
            self.y * tf.log(self.o + eps) + (1 - self.y) *
            tf.log(1 - self.o +
                   eps)  #the loss is logistic log loss element wise loss
        )
        with tf.variable_scope('optimizer', reuse=reuse):
            self.optimizer = tf.train.RMSPropOptimizer(
                learning_rate=args.learning_rate, momentum=0.9,
                decay=0.95)  #optimizing
            gvs = self.optimizer.compute_gradients(self.copy_loss)
            capped_gvs = [
                (tf.clip_by_value(grad, -10., 10.), var) for grad, var in gvs
            ]  #need to clip gradients because normaly
            self.train_op = self.optimizer.apply_gradients(capped_gvs)
        self.copy_loss_summary = tf.summary.scalar(
            'copy_loss_%d' % seq_length,
            self.copy_loss)  #for different sequence length