Beispiel #1
0
def cifar10_shufflenet(x, n_groups=2, n_filters=200, ratio=1.0, seed=42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    conv = conv2d_bn_relu(x,
                          size=3,
                          n_filters=24,
                          is_training=training,
                          kernel_init=He_normal(seed=seed + 1),
                          name="initial_conv")
    layers.append(("initial_conv", conv))

    slayer1 = shufflenet.shufflenet_layer(conv,
                                          n_filters=n_filters,
                                          n_repeat=3,
                                          n_groups=n_groups,
                                          reduction_ratio=ratio,
                                          is_training=training,
                                          kernel_init=He_normal(seed=seed + 2),
                                          name="shufflenet_layer_1")
    layers.append(("shufflenet_layer_1", slayer1))

    slayer2 = shufflenet.shufflenet_layer(slayer1,
                                          n_filters=n_filters * 2,
                                          n_repeat=7,
                                          n_groups=n_groups,
                                          reduction_ratio=ratio,
                                          is_training=training,
                                          kernel_init=He_normal(seed=seed + 3),
                                          name="shufflenet_layer_2")
    layers.append(("shufflenet_layer_2", slayer2))

    slayer3 = shufflenet.shufflenet_layer(slayer2,
                                          n_filters=n_filters * 4,
                                          n_repeat=3,
                                          n_groups=n_groups,
                                          reduction_ratio=ratio,
                                          is_training=training,
                                          kernel_init=He_normal(seed=seed + 4),
                                          name="shufflenet_layer_3")
    layers.append(("shufflenet_layer_3", slayer3))

    pool = global_avg_pool2d(slayer3)
    layers.append(("pool", pool))

    dense1 = dense(pool,
                   n_units=10,
                   kernel_init=Kumar_normal(activation=None,
                                            mode="FAN_IN",
                                            seed=seed + 5),
                   name="dense_1")
    layers.append(("logit", dense1))

    prob = tf.nn.softmax(dense1, name="prob")
    layers.append(("prob", prob))

    return layers, variables
def cifar10_resnext(x, n_blocks, cardinality = 8, group_width = 16, seed = 42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    conv = conv2d_bn_act(
            x, size = 3, n_filters = 32,
            activation = tf.nn.relu,
            is_training = training,
            kernel_init = He_normal(seed = seed+1),
            name = "initial_conv")
    layers.append(("initial_conv", conv))
            
    res1 = resnext.residual_layer(
            conv, n_blocks = n_blocks, stride = 1,
            cardinality = cardinality,
            group_width = group_width,
            block_function = resnext.bottleneck_block,
            is_training = training,
            kernel_init = He_normal(seed = seed+2),
            name = "residual_1")
    layers.append(("residual_1", res1))

    res2 = resnext.residual_layer(
            res1, n_blocks = n_blocks, stride = 2,
            cardinality = cardinality,
            group_width = group_width*2,
            block_function = resnext.bottleneck_block,
            is_training = training,
            kernel_init = He_normal(seed = seed+3),            
            name="residual_2")
    layers.append(("residual_2", res2))

    res3 = resnext.residual_layer(
            res2, n_blocks = n_blocks, stride = 2,
            cardinality = cardinality,
            group_width = group_width*4,
            block_function = resnext.bottleneck_block,
            is_training = training,
            kernel_init = He_normal(seed = seed+4),
            name = "residual_3")
    layers.append(("residual_3", res3))

    pool = global_avg_pool2d(res3)
    layers.append(("pool", pool))
    
    dense1 = dense(
            pool, n_units = 10,
            kernel_init = Kumar_normal(activation = None, mode = "FAN_IN", seed = seed+5),
            name = "dense_1")
    layers.append(("logit", dense1))
    
    prob = tf.nn.softmax(dense1, name = "prob")
    layers.append(("prob", prob))
    
    return layers, variables
Beispiel #3
0
def final_layer(inputs,
                activation=tf.nn.relu,
                is_training=False,
                name="final_layer"):
    with tf.variable_scope(name):
        x = tf.layers.batch_normalization(inputs,
                                          training=is_training,
                                          name="batch_norm")
        x = activation(x, name="activation")
        x = global_avg_pool2d(x)
    return x
Beispiel #4
0
def auxiliary_classifier(inputs,
                         classes,
                         is_training=False,
                         regularizer=None,
                         activation=tf.nn.relu,
                         conv_kernel_init=He_normal(seed=42),
                         conv_bias_init=tf.zeros_initializer(),
                         dense_kernel_init=Kumar_normal(activation=None,
                                                        mode="FAN_IN",
                                                        seed=42),
                         name="nasnet_auxiliary_classifier"):

    with tf.variable_scope(name):
        x = inputs
        if activation is not None:
            x = activation(x, name="activation_1")
        x = avg_pool2d(x, size=5, stride=3, padding="VALID", name="avg_pool")

        x = conv2d_bn(x,
                      n_filters=128,
                      size=1,
                      is_training=is_training,
                      regularizer=regularizer,
                      kernel_init=conv_kernel_init,
                      bias_init=conv_bias_init,
                      name="conv_projection")

        if activation is not None:
            x = activation(x, name="activation_2")

        x = conv2d_bn(x,
                      n_filters=768,
                      size=[x.get_shape()[1].value,
                            x.get_shape()[2].value],
                      padding="VALID",
                      is_training=is_training,
                      regularizer=regularizer,
                      kernel_init=conv_kernel_init,
                      bias_init=conv_bias_init,
                      name="conv_reduction")

        if activation is not None:
            x = activation(x, name="activation_3")

        x = global_avg_pool2d(x, name="global_avg_pool")

        x = dense(x,
                  n_units=classes,
                  regularizer=regularizer,
                  kernel_init=dense_kernel_init,
                  name="dense")

    return x
Beispiel #5
0
def squeeze_and_excite(inputs,
                       ratio=16,
                       regularizer=None,
                       kernel_init_1=He_normal(seed=42),
                       kernel_init_2=Kumar_normal(activation="sigmoid",
                                                  mode="FAN_AVG",
                                                  seed=42),
                       name="squeeze_excite"):
    in_filt = inputs.shape[3].value
    with tf.variable_scope(name):
        x = global_avg_pool2d(inputs)
        x = dense_relu(x,
                       n_units=in_filt // ratio,
                       regularizer=regularizer,
                       kernel_init=kernel_init_1,
                       name="dense_1")
        x = dense_sigmoid(x,
                          n_units=in_filt,
                          regularizer=regularizer,
                          kernel_init=kernel_init_2,
                          name="dense_2")
        x = tf.reshape(x, [-1, 1, 1, in_filt])
        outputs = tf.multiply(inputs, x)
    return outputs
Beispiel #6
0
def cifar10_resnet_bottleneck_wd(x, n_blocks=3, weight_decay=0.0001, seed=42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    conv = conv2d_bn_act(
        x,
        size=3,
        n_filters=16,
        activation=tf.nn.relu,
        is_training=training,
        regularizer=tf.contrib.layers.l2_regularizer(weight_decay),
        kernel_init=He_normal(seed=seed + 1),
        name="initial_conv")
    layers.append(("initial_conv", conv))

    res1 = resnet.residual_layer(
        conv,
        n_filters=16,
        n_blocks=n_blocks,
        stride=1,
        block_function=partial(
            resnet.residual_block,
            regularizer=tf.contrib.layers.l2_regularizer(weight_decay)),
        is_training=training,
        kernel_init=He_normal(seed=seed + 2),
        name="residual_1")
    layers.append(("residual_1", res1))

    res2 = resnet.residual_layer(
        res1,
        n_filters=32,
        n_blocks=n_blocks,
        stride=2,
        block_function=partial(
            resnet.residual_block,
            regularizer=tf.contrib.layers.l2_regularizer(weight_decay)),
        is_training=training,
        kernel_init=He_normal(seed=seed + 3),
        name="residual_2")
    layers.append(("residual_2", res2))

    res3 = resnet.residual_layer(
        res2,
        n_filters=64,
        n_blocks=n_blocks,
        stride=2,
        block_function=partial(
            resnet.bottleneck_block,
            n_filters_reduce=16,
            regularizer=tf.contrib.layers.l2_regularizer(weight_decay)),
        is_training=training,
        kernel_init=He_normal(seed=seed + 4),
        name="residual_3")
    layers.append(("residual_3", res3))

    pool = global_avg_pool2d(res3)
    layers.append(("pool", pool))

    dense1 = dense(pool,
                   n_units=10,
                   regularizer=tf.contrib.layers.l2_regularizer(weight_decay),
                   kernel_init=Kumar_normal(activation=None,
                                            mode="FAN_IN",
                                            seed=seed + 5),
                   name="dense_1")
    layers.append(("logit", dense1))

    prob = tf.nn.softmax(dense1, name="prob")
    layers.append(("prob", prob))

    return layers, variables
def cifar10_mobilenet_v2(x, expand_ratio=6, seed=42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    conv1 = conv2d_bn_relu(x,
                           size=3,
                           n_filters=32,
                           kernel_init=He_normal(seed=seed + 1),
                           is_training=training,
                           name="initial_conv")
    layers.append(("initial_conv", conv1))

    # 32x32x32 -> 32x32x16
    invres1 = mobilenet_v2.inverted_residual_block(
        conv1,
        n_repeat=1,
        n_filters=16,
        stride=1,
        expand_ratio=1,
        kernel_init=He_normal(seed=seed + 2),
        is_training=training,
        name="inverted_residual_block_1")
    layers.append(("inverted_residual_block_1", invres1))

    # 32x32x16 -> 32x32x24
    invres2 = mobilenet_v2.inverted_residual_block(
        invres1,
        n_repeat=2,
        n_filters=24,
        stride=1,
        expand_ratio=expand_ratio,
        kernel_init=He_normal(seed=seed + 3),
        is_training=training,
        name="inverted_residual_block_2")
    layers.append(("inverted_residual_block_2", invres2))

    #32x32x24 -> 16x16x32
    invres3 = mobilenet_v2.inverted_residual_block(
        invres2,
        n_repeat=3,
        n_filters=32,
        stride=2,
        expand_ratio=expand_ratio,
        kernel_init=He_normal(seed=seed + 4),
        is_training=training,
        name="inverted_residual_block_3")
    layers.append(("inverted_residual_block_3", invres3))

    #16x16x32 -> 8x8x64
    invres4 = mobilenet_v2.inverted_residual_block(
        invres3,
        n_repeat=4,
        n_filters=64,
        stride=2,
        expand_ratio=expand_ratio,
        kernel_init=He_normal(seed=seed + 5),
        is_training=training,
        name="inverted_residual_block_4")
    layers.append(("inverted_residual_block_4", invres4))

    #8x8x64 -> 8x8x96
    invres5 = mobilenet_v2.inverted_residual_block(
        invres4,
        n_repeat=3,
        n_filters=96,
        stride=1,
        expand_ratio=expand_ratio,
        kernel_init=He_normal(seed=seed + 6),
        is_training=training,
        name="inverted_residual_block_5")
    layers.append(("inverted_residual_block_5", invres5))

    #8x8x96 -> 4x4x160
    invres6 = mobilenet_v2.inverted_residual_block(
        invres5,
        n_repeat=3,
        n_filters=160,
        stride=2,
        expand_ratio=expand_ratio,
        kernel_init=He_normal(seed=seed + 7),
        is_training=training,
        name="inverted_residual_block_6")
    layers.append(("inverted_residual_block_6", invres6))

    #4x4x160 -> 4x4x320
    invres7 = mobilenet_v2.inverted_residual_block(
        invres6,
        n_repeat=1,
        n_filters=320,
        stride=1,
        expand_ratio=expand_ratio,
        kernel_init=He_normal(seed=seed + 8),
        is_training=training,
        name="inverted_residual_block_7")
    layers.append(("inverted_residual_block_7", invres7))

    conv2 = conv2d_bn_relu(invres7,
                           size=1,
                           n_filters=1280,
                           kernel_init=He_normal(seed=seed + 9),
                           is_training=training,
                           name="final_conv")
    layers.append(("final_conv", conv2))

    pool = global_avg_pool2d(conv2)
    layers.append(("pool", pool))

    dense1 = dense(pool,
                   n_units=10,
                   kernel_init=Kumar_normal(activation=None,
                                            mode="FAN_IN",
                                            seed=seed + 10),
                   name="dense_1")
    layers.append(("logit", dense1))

    prob = tf.nn.softmax(dense1, name="prob")
    layers.append(("prob", prob))

    return layers, variables
Beispiel #8
0
def cifar10_nasnet(x, drop_rate = 0.0, seed = 42):
    penultimate_filters = 768
    nb_blocks = 6
    stem_filters = 32
    filters_multiplier = 2

    filters = penultimate_filters // 24 # 2x2x6 -> increase two times 2x and concatenate 6 branches
    
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))
    
    
    x = conv2d_bn(
        x, n_filters = stem_filters, size = 3, stride = 1,
        is_training = training,
        kernel_init = He_normal(seed = seed+1),
        name = "initial_conv")
    layers.append(("initial_conv", x))
    
    p = None
        
    for i in range(nb_blocks):
        x, p = nasnet.Normal_A(
            x, p,
            n_filters = filters,
            is_training = training,
            kernel_init = He_normal(seed = seed+2+i),
            name = "nasnet_normal_"+str(i)
            )
        layers.append(("nasnet_normal_"+str(i), x))

    x, _ = nasnet.Reduction_A(
        x, p,
        n_filters = filters * filters_multiplier,
        is_training = training,
        kernel_init = He_normal(seed = seed+3+nb_blocks),
        name = "nasnet_reduction_0"
        )
    layers.append(("nasnet_reduction_0", x))

    for i in range(nb_blocks):
        x, p = nasnet.Normal_A(
            x, p,
            n_filters = filters * filters_multiplier,
            is_training = training,
            kernel_init = He_normal(seed = seed+4+nb_blocks+i),
            name = "nasnet_normal_"+str(nb_blocks+i)
            )
        layers.append(("nasnet_normal_"+str(nb_blocks+i), x))

    x, _ = nasnet.Reduction_A(
        x, p,
        n_filters = filters * filters_multiplier ** 2,
        is_training = training,
        kernel_init = He_normal(seed = seed+5+2*nb_blocks),
        name = "nasnet_reduction_1"
        )
    layers.append(("nasnet_reduction_1", x))

    aux = nasnet.auxiliary_classifier(
        x, classes = 10,
        is_training = training,
        conv_kernel_init = He_normal(seed = seed),
        dense_kernel_init = Kumar_normal(activation = None, mode = "FAN_IN", seed = seed),
        name = "nasnet_aux_classifier"
        )
    layers.append(("aux_logit", aux))
    aux_prob = tf.nn.softmax(aux, name = "prob")
    layers.append(("aux_prob", aux_prob)) 

    for i in range(nb_blocks):
        x, p = nasnet.Normal_A(
            x, p,
            n_filters = filters * filters_multiplier ** 2,
            is_training = training,
            kernel_init = He_normal(seed = seed+6+2*nb_blocks+i),
            name = "nasnet_normal_"+str(2*nb_blocks+i)
            )
        layers.append(("nasnet_normal_"+str(2*nb_blocks+i), x))
        
    x = tf.nn.relu(x, name = "relu")    
    layers.append(("relu", x))
    
    x = global_avg_pool2d(x, name = "pool")
    layers.append(("pool", x))
    if drop_rate > 0.0:
        x = tf.layers.dropout(
                x, rate = drop_rate, training = training,
                seed = seed+7+3*nb_blocks, name = "dropout")
        layers.append(("dropout", x))
    x = dense(
            x, n_units = 10,
            kernel_init = Kumar_normal(activation = None, mode = "FAN_IN", seed = seed+8+3*nb_blocks),
            name = "dense")
    layers.append(("logit", x))
    
    prob = tf.nn.softmax(x, name = "prob")
    layers.append(("prob", prob))
    
    return layers, variables
def cifar10_se_resnet(x, n_blocks, ratio = 8, seed = 42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    conv = conv2d_bn_act(
            x, size = 3, n_filters = 16,
            activation = tf.nn.relu,
            is_training = training,
            kernel_init = He_normal(seed = seed+1),
            name = "initial_conv")
    layers.append(("initial_conv", conv))
            
    res1 = resnet.residual_layer(
            conv, n_filters = 16, n_blocks = n_blocks, stride = 1,
            block_function = partial(
                    senet.se_resnet_residual_block,
                    ratio = ratio,
                    se_kernel_init_1 = He_normal(seed = seed+2),
                    se_kernel_init_2 = Kumar_normal(activation = "sigmoid", mode = "FAN_AVG", seed = seed+2),
                    ), 
            is_training = training,
            kernel_init = He_normal(seed = seed+2),
            name = "se_residual_1")
    layers.append(("se_residual_1", res1))

    res2 = resnet.residual_layer(
            res1, n_filters = 32, n_blocks = n_blocks, stride = 2,
            block_function = partial(
                    senet.se_resnet_residual_block,
                    ratio = ratio,
                    se_kernel_init_1 = He_normal(seed = seed+3),
                    se_kernel_init_2 = Kumar_normal(activation = "sigmoid", mode = "FAN_AVG", seed = seed+3),
                    ), 
            is_training = training,
            kernel_init = He_normal(seed = seed+3),
            name = "se_residual_2")
    layers.append(("se_residual_2", res2))

    res3 = resnet.residual_layer(
            res2, n_filters = 64, n_blocks = n_blocks, stride = 2,
            block_function = partial(
                    senet.se_resnet_residual_block,
                    ratio = ratio,
                    se_kernel_init_1 = He_normal(seed = seed+4),
                    se_kernel_init_2 = Kumar_normal(activation = "sigmoid", mode = "FAN_AVG", seed = seed+4),
                    ),
            is_training = training,
            kernel_init = He_normal(seed = seed+4),
            name = "se_residual_3")
    layers.append(("se_residual_3", res3))

    pool = global_avg_pool2d(res3)
    layers.append(("pool", pool))
    
    dense1 = dense(
            pool, n_units = 10,
            kernel_init = Kumar_normal(activation = None, mode = "FAN_IN", seed = seed+5),
            name = "dense_1")
    layers.append(("logit", dense1))
    
    prob = tf.nn.softmax(dense1, name = "prob")
    layers.append(("prob", prob))
    
    return layers, variables
def cifar10_bottleneck_densenet(x, n_repeat, drop_rate=0.25, seed=42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    conv = conv2d(x,
                  size=3,
                  n_filters=16,
                  kernel_init=He_normal(seed=seed + 1),
                  name="initial_conv")
    layers.append(("initial_conv", conv))

    dblock1 = densenet.bottleneck_block(conv,
                                        n_repeat=n_repeat,
                                        n_filters=12,
                                        reduction_ratio=4,
                                        drop_rate=drop_rate,
                                        is_training=training,
                                        kernel_init=He_normal(seed=seed + 2),
                                        seed=seed + 2,
                                        name="dense_bootleneck_block_1")
    layers.append(("dense_bootleneck_block_1", dblock1))

    tlayer1 = densenet.transition_layer(dblock1,
                                        pool_size=2,
                                        pool_stride=2,
                                        drop_rate=drop_rate,
                                        is_training=training,
                                        kernel_init=He_normal(seed=seed + 3),
                                        seed=seed + 3,
                                        name="transition_layer_1")
    layers.append(("transition_layer_1", tlayer1))

    dblock2 = densenet.bottleneck_block(tlayer1,
                                        n_repeat=n_repeat,
                                        n_filters=12,
                                        reduction_ratio=4,
                                        drop_rate=drop_rate,
                                        is_training=training,
                                        kernel_init=He_normal(seed=seed + 4),
                                        seed=seed + 4,
                                        name="dense_bootleneck_block_2")
    layers.append(("dense_bootleneck_block_2", dblock2))

    tlayer2 = densenet.transition_layer(dblock2,
                                        pool_size=2,
                                        pool_stride=2,
                                        drop_rate=drop_rate,
                                        is_training=training,
                                        kernel_init=He_normal(seed=seed + 5),
                                        seed=seed + 5,
                                        name="transition_layer_2")
    layers.append(("transition_layer_2", tlayer2))

    dblock3 = densenet.bottleneck_block(tlayer2,
                                        n_repeat=n_repeat,
                                        n_filters=12,
                                        reduction_ratio=4,
                                        drop_rate=drop_rate,
                                        is_training=training,
                                        kernel_init=He_normal(seed=seed + 6),
                                        seed=seed + 6,
                                        name="dense_bootleneck_block_3")
    layers.append(("dense_bootleneck_block_3", dblock3))

    pool = global_avg_pool2d(dblock3)
    layers.append(("pool", pool))

    dense1 = dense(pool,
                   n_units=10,
                   kernel_init=Kumar_normal(activation=None,
                                            mode="FAN_IN",
                                            seed=seed + 7),
                   name="dense_1")
    layers.append(("logit", dense1))

    prob = tf.nn.softmax(dense1, name="prob")
    layers.append(("prob", prob))

    return layers, variables
Beispiel #11
0
def cifar10_sequential_allconvC(x,
                                drop_rate=0.5,
                                drop_rate_input=0.2,
                                seed=42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    dropout_in = tf.layers.dropout(x,
                                   rate=drop_rate_input,
                                   training=training,
                                   seed=seed + 0,
                                   name="dropout_input")
    layers.append(("dropout_in", dropout_in))

    conv1 = conv2d_relu(
        dropout_in,
        size=3,
        n_filters=96,
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 1),
        name="conv_1")
    layers.append(("conv_1", conv1))

    conv2 = conv2d_relu(
        conv1,
        size=3,
        n_filters=96,
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 2),
        name="conv_2")
    layers.append(("conv_2", conv2))

    conv3 = conv2d_relu(
        conv2,
        stride=2,
        size=3,
        n_filters=96,
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 3),
        name="conv_3")
    layers.append(("conv_3", conv3))

    dropout1 = tf.layers.dropout(conv3,
                                 rate=drop_rate,
                                 training=training,
                                 seed=seed + 3,
                                 name="dropout_1")
    layers.append(("dropout_1", dropout1))

    conv4 = conv2d_relu(
        dropout1,
        size=3,
        n_filters=192,
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 4),
        name="conv_4")
    layers.append(("conv_4", conv4))

    conv5 = conv2d_relu(
        conv4,
        size=3,
        n_filters=192,
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 5),
        name="conv_5")
    layers.append(("conv_5", conv5))

    conv6 = conv2d_relu(
        conv5,
        stride=2,
        size=3,
        n_filters=192,
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 6),
        name="conv_6")
    layers.append(("conv_6", conv6))

    dropout2 = tf.layers.dropout(conv6,
                                 rate=drop_rate,
                                 training=training,
                                 seed=seed + 6,
                                 name="dropout_2")
    layers.append(("dropout_2", dropout2))

    conv7 = conv2d_relu(
        dropout2,
        size=3,
        n_filters=192,
        padding="VALID",
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 7),
        name="conv_7")
    layers.append(("conv_7", conv7))

    conv8 = conv2d_relu(
        conv7,
        size=1,
        n_filters=192,
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 8),
        name="conv_8")
    layers.append(("conv_8", conv8))

    conv9 = conv2d_relu(
        conv8,
        size=1,
        n_filters=10,
        kernel_init=tf.contrib.layers.xavier_initializer(seed=seed + 9),
        name="conv_9")
    layers.append(("conv_9", conv9))

    pool = global_avg_pool2d(conv9)
    layers.append(("logit", pool))

    prob = tf.nn.softmax(pool, name="prob")
    layers.append(("prob", prob))

    return layers, variables
Beispiel #12
0
def cifar10_xception_wd(x, drop_rate=0.5, weight_decay=0.0001, seed=42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    # 32x32
    conv = conv2d_bn_relu(
        x,
        size=3,
        n_filters=64,
        is_training=training,
        regularizer=tf.contrib.layers.l2_regularizer(weight_decay),
        kernel_init=He_normal(seed=seed + 1),
        name="initial_conv")
    layers.append(("initial_conv", conv))

    entry = xception.entry_module(
        conv,
        n_filters=[128, 256, 728],
        is_training=training,
        regularizer=tf.contrib.layers.l2_regularizer(weight_decay),
        kernel_init=He_normal(seed=seed + 2),
        name="entry")
    layers.append(("entry", entry))

    mid = xception.middle_module(
        entry,
        n_filters=728,
        n_repeat=3,
        is_training=training,
        regularizer=tf.contrib.layers.l2_regularizer(weight_decay),
        kernel_init=He_normal(seed=seed + 3),
        name="middle")
    layers.append(("middle", mid))

    exits = xception.exit_module(
        mid,
        n_filters_1=[728, 1024],
        n_filters_2=[1536, 2048],
        is_training=training,
        regularizer=tf.contrib.layers.l2_regularizer(weight_decay),
        kernel_init=He_normal(seed=seed + 4),
        name="exit")
    layers.append(("exit", exits))

    pool1 = global_avg_pool2d(exits, name="pool1")
    layers.append(("pool1", pool1))

    dropout1 = tf.layers.dropout(pool1,
                                 rate=drop_rate,
                                 training=training,
                                 seed=seed + 5,
                                 name="dropout")
    layers.append(("dropout1", dropout1))

    dense1 = dense(dropout1,
                   n_units=10,
                   regularizer=tf.contrib.layers.l2_regularizer(weight_decay),
                   kernel_init=Kumar_normal(activation=None,
                                            mode="FAN_IN",
                                            seed=seed + 6),
                   name="dense1")
    layers.append(("logit", dense1))

    prob = tf.nn.softmax(dense1, name="prob")
    layers.append(("prob", prob))

    return layers, variables
Beispiel #13
0
def cifar10_resnet_identity(x, n_blocks=3, seed=42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    conv = conv2d_bn_act(x,
                         size=3,
                         n_filters=16,
                         activation=tf.nn.relu,
                         is_training=training,
                         kernel_init=He_normal(seed=seed + 1),
                         name="initial_conv")
    layers.append(("initial_conv", conv))

    res1 = resnet.residual_layer(conv,
                                 n_filters=16,
                                 n_blocks=n_blocks,
                                 stride=1,
                                 block_function=partial(
                                     resnet_identity.identity_mapping_block,
                                     skip_first_bn_act=True),
                                 is_training=training,
                                 kernel_init=He_normal(seed=seed + 2),
                                 name="residual_1")
    layers.append(("residual_1", res1))

    res2 = resnet.residual_layer(
        res1,
        n_filters=32,
        n_blocks=n_blocks,
        stride=2,
        block_function=resnet_identity.identity_mapping_block,
        is_training=training,
        kernel_init=He_normal(seed=seed + 3),
        name="residual_2")
    layers.append(("residual_2", res2))

    res3 = resnet.residual_layer(
        res2,
        n_filters=64,
        n_blocks=n_blocks,
        stride=2,
        block_function=resnet_identity.identity_mapping_block,
        is_training=training,
        kernel_init=He_normal(seed=seed + 4),
        name="residual_3")
    layers.append(("residual_3", res3))

    bn = tf.layers.batch_normalization(res3,
                                       training=training,
                                       name="batch_norm")
    layers.append(("batch_norm", bn))
    bn_relu = tf.nn.relu(bn, name="relu")
    layers.append(("batch_norm_relu", bn_relu))

    pool = global_avg_pool2d(bn_relu)
    layers.append(("pool", pool))

    dense1 = dense(pool,
                   n_units=10,
                   kernel_init=Kumar_normal(activation=None,
                                            mode="FAN_IN",
                                            seed=seed + 5),
                   name="dense_1")
    layers.append(("logit", dense1))

    prob = tf.nn.softmax(dense1, name="prob")
    layers.append(("prob", prob))

    return layers, variables
def cifar10_mobilenet(x, seed=42):
    layers = []
    variables = []

    training = tf.placeholder(tf.bool, name="training")
    variables.append(("training", training))

    conv = conv2d_bn_relu(x,
                          size=3,
                          n_filters=32,
                          kernel_init=He_normal(seed=seed + 1),
                          is_training=training,
                          name="initial_conv")
    layers.append(("initial_conv", conv))

    mblock1 = mobilenet.mobilenet_block(conv,
                                        n_filters=64,
                                        stride=1,
                                        kernel_init=He_normal(seed=seed + 2),
                                        is_training=training,
                                        name="mobilenet_block_1")
    layers.append(("mobilenet_block_1", mblock1))

    # 16x16
    mblock2 = mobilenet.mobilenet_block(mblock1,
                                        n_filters=128,
                                        stride=2,
                                        kernel_init=He_normal(seed=seed + 3),
                                        is_training=training,
                                        name="mobilenet_block_2")
    layers.append(("mobilenet_block_2", mblock2))

    mblock3 = mobilenet.mobilenet_block(mblock2,
                                        n_filters=128,
                                        stride=1,
                                        kernel_init=He_normal(seed=seed + 4),
                                        is_training=training,
                                        name="mobilenet_block_3")
    layers.append(("mobilenet_block_3", mblock3))

    # 8x8
    mblock4 = mobilenet.mobilenet_block(mblock3,
                                        n_filters=256,
                                        stride=2,
                                        kernel_init=He_normal(seed=seed + 5),
                                        is_training=training,
                                        name="mobilenet_block_4")
    layers.append(("mobilenet_block_4", mblock4))

    mblock5 = mobilenet.mobilenet_block(mblock4,
                                        n_filters=256,
                                        stride=1,
                                        kernel_init=He_normal(seed=seed + 6),
                                        is_training=training,
                                        name="mobilenet_block_5")
    layers.append(("mobilenet_block_5", mblock5))

    # 4x4
    mblock6 = mobilenet.mobilenet_block(mblock5,
                                        n_filters=512,
                                        stride=2,
                                        kernel_init=He_normal(seed=seed + 7),
                                        is_training=training,
                                        name="mobilenet_block_6")
    layers.append(("mobilenet_block_6", mblock6))

    mblock7 = mobilenet.mobilenet_block(mblock6,
                                        n_filters=512,
                                        stride=1,
                                        kernel_init=He_normal(seed=seed + 8),
                                        is_training=training,
                                        name="mobilenet_block_7")
    layers.append(("mobilenet_block_7", mblock7))

    mblock8 = mobilenet.mobilenet_block(mblock7,
                                        n_filters=512,
                                        stride=1,
                                        kernel_init=He_normal(seed=seed + 9),
                                        is_training=training,
                                        name="mobilenet_block_8")
    layers.append(("mobilenet_block_8", mblock8))

    mblock9 = mobilenet.mobilenet_block(mblock8,
                                        n_filters=512,
                                        stride=1,
                                        kernel_init=He_normal(seed=seed + 10),
                                        is_training=training,
                                        name="mobilenet_block_9")
    layers.append(("mobilenet_block_9", mblock9))

    mblock10 = mobilenet.mobilenet_block(mblock9,
                                         n_filters=512,
                                         stride=1,
                                         kernel_init=He_normal(seed=seed + 11),
                                         is_training=training,
                                         name="mobilenet_block_10")
    layers.append(("mobilenet_block_10", mblock10))

    mblock11 = mobilenet.mobilenet_block(mblock10,
                                         n_filters=512,
                                         stride=1,
                                         kernel_init=He_normal(seed=seed + 12),
                                         is_training=training,
                                         name="mobilenet_block_11")
    layers.append(("mobilenet_block_11", mblock11))

    # 2x2
    mblock12 = mobilenet.mobilenet_block(mblock11,
                                         n_filters=1024,
                                         stride=2,
                                         kernel_init=He_normal(seed=seed + 13),
                                         is_training=training,
                                         name="mobilenet_block_12")
    layers.append(("mobilenet_block_12", mblock12))

    mblock13 = mobilenet.mobilenet_block(mblock12,
                                         n_filters=1024,
                                         stride=1,
                                         kernel_init=He_normal(seed=seed + 14),
                                         is_training=training,
                                         name="mobilenet_block_13")
    layers.append(("mobilenet_block_13", mblock13))

    pool = global_avg_pool2d(mblock13)
    layers.append(("pool", pool))

    dense1 = dense(pool,
                   n_units=10,
                   kernel_init=Kumar_normal(activation=None,
                                            mode="FAN_IN",
                                            seed=seed + 15),
                   name="dense1")
    layers.append(("logit", dense1))

    prob = tf.nn.softmax(dense1, name="prob")
    layers.append(("prob", prob))

    return layers, variables