示例#1
0
def visualize_rotation(image):
    image_tf = tf.expand_dims(image, 0)
    visualize_plots([
        image,
        Sequential([preprocessing.RandomRotation(factor=(1 / 6, 1 / 6))])(image_tf)[0],
        Sequential([preprocessing.RandomRotation(factor=(-1 / 6, -1 / 6))])(image_tf)[0]
    ], "rotation")
    def build(self, hp, inputs=None):
        input_node = nest.flatten(inputs)[0]
        output_node = input_node

        # Translate
        translation_factor = self.translation_factor
        if translation_factor is None:
            translation_factor = hp.Choice("translation_factor", [0.0, 0.1])
        if translation_factor != 0 and translation_factor != (0, 0):
            height_factor, width_factor = self._get_fraction_value(
                translation_factor)
            output_node = preprocessing.RandomTranslation(
                height_factor, width_factor)(output_node)

        # Flip
        horizontal_flip = self.horizontal_flip
        if horizontal_flip is None:
            horizontal_flip = hp.Boolean("horizontal_flip", default=True)
        vertical_flip = self.vertical_flip
        if self.vertical_flip is None:
            vertical_flip = hp.Boolean("vertical_flip", default=True)
        if not horizontal_flip and not vertical_flip:
            flip_mode = ""
        elif horizontal_flip and vertical_flip:
            flip_mode = "horizontal_and_vertical"
        elif horizontal_flip and not vertical_flip:
            flip_mode = "horizontal"
        elif not horizontal_flip and vertical_flip:
            flip_mode = "vertical"
        if flip_mode != "":
            output_node = preprocessing.RandomFlip(mode=flip_mode)(output_node)

        # Rotate
        rotation_factor = self.rotation_factor
        if rotation_factor is None:
            rotation_factor = hp.Choice("rotation_factor", [0.0, 0.1])
        if rotation_factor != 0:
            output_node = preprocessing.RandomRotation(rotation_factor)(
                output_node)

        # Zoom
        zoom_factor = self.zoom_factor
        if zoom_factor is None:
            zoom_factor = hp.Choice("zoom_factor", [0.0, 0.1])
        if zoom_factor != 0 and zoom_factor != (0, 0):
            height_factor, width_factor = self._get_fraction_value(zoom_factor)
            # TODO: Add back RandomZoom when it is ready.
            # output_node = preprocessing.RandomZoom(
            # height_factor, width_factor)(output_node)

        # Contrast
        contrast_factor = self.contrast_factor
        if contrast_factor is None:
            contrast_factor = hp.Choice("contrast_factor", [0.0, 0.1])
        if contrast_factor != 0 and contrast_factor != (0, 0):
            output_node = preprocessing.RandomContrast(contrast_factor)(
                output_node)

        return output_node
 def _get_data_augment_layer(self):
     data_augment_layer = tf.keras.models.Sequential([
         KerasPreprocessing.RandomFlip('horizontal'),
         KerasPreprocessing.RandomRotation(0.2),
         KerasPreprocessing.RandomZoom(0.2),
         KerasPreprocessing.RandomHeight(0.2),
         KerasPreprocessing.Resizing(height=self.IMAGE_DIM[0], width=self.IMAGE_DIM[1])
     ])
     
     return data_augment_layer
示例#4
0
文件: krutz-lab6.py 项目: jkrutz/Lab6
def create_data_augmentation_layer():
    # When you don't have a large image dataset,
    # it's a good practice to artificially introduce sample diversity
    # by applying random yet realistic transformations to the training images.
    data_augmentation = keras.Sequential(
        [
            preprocessing.RandomFlip("horizontal"),
            preprocessing.RandomRotation(0.1),
        ]
    )
    return data_augmentation
示例#5
0
    def __init__(self):
        super(ClassifierHybrid, self).__init__()
        self.global_step = 0
        self.backbone = self.get_backbone()
        self.backbone.trainable = False
        trainable_count = np.sum(
            [K.count_params(w) for w in self.backbone.trainable_weights])
        non_trainable_count = np.sum(
            [K.count_params(w) for w in self.backbone.non_trainable_weights])

        print('Total params: {:,}'.format(trainable_count +
                                          non_trainable_count))
        print('Trainable params: {:,}'.format(trainable_count))
        print('Non-trainable params: {:,}'.format(non_trainable_count))
        # self.head = tf.keras.Sequential([
        #     layers.Flatten(),
        #     layers.Dense(256, activation='relu'),
        #     layers.Dense(196)
        # ])

        # self.vision_transformer = ViT(img_size=9, channels=1408, patch_size=1, num_layers=8,
        #                  num_classes=196, d_model=512, num_heads=8, d_mlp=512)

        self.vision_transformer = ViT(img_size=args.num_patches,
                                      channels=args.num_channels,
                                      patch_size=args.patch_size,
                                      num_layers=args.num_layers,
                                      num_classes=args.num_classes,
                                      d_model=args.d_model,
                                      num_heads=args.num_heads,
                                      d_mlp=args.d_mlp)
        self.prepare_datasets()
        self.flag = True
        self.augmentation = tf.keras.Sequential(
            [
                tf.keras.Input(shape=(260, 260, 3)),
                preprocessing.RandomRotation(factor=0.15),
                preprocessing.RandomTranslation(height_factor=0.1,
                                                width_factor=0.1),
                preprocessing.RandomFlip(),
                preprocessing.RandomContrast(factor=0.1),
            ],
            name="augmentation",
        )
示例#6
0
    def generic_builder(self, name, net, lr=1e-2, dropout_rate=0.2):
        cfg = self.cfg
        inputs = layers.Input(shape=cfg['img_shape'])

        img_augmentation = Sequential(
            [
                preprocessing.RandomRotation(factor=0.15),
                #preprocessing.RandomTranslation(height_factor=0.1, width_factor=0.1),
                # preprocessing.RandomFlip(),
                preprocessing.RandomContrast(factor=0.1),
            ],
            name="img_augmentation",
        )
        x = img_augmentation(inputs)
        if cfg['transfer_learning']:
            model = net(include_top=False, input_tensor=x, weights='imagenet')

            # Freeze the pretrained weights
            model.trainable = False

            # Rebuild top
            x = layers.GlobalAveragePooling2D(name="avg_pool")(model.output)
            x = layers.BatchNormalization()(x)
            top_dropout_rate = dropout_rate
            x = layers.Dropout(top_dropout_rate, name="top_dropout")(x)
            outputs = layers.Dense(cfg['num_classes'],
                                   activation="softmax",
                                   name="pred")(x)
        else:
            model = net(include_top=False, input_tensor=x, weights=None)
            model.trainable = True
            top_dropout_rate = dropout_rate
            x = layers.Dropout(top_dropout_rate, name="top_dropout")(x)
            outputs = layers.Dense(cfg['num_classes'],
                                   activation="softmax",
                                   name="pred")(x)

        # Compile
        model = tf.keras.Model(inputs, outputs, name=name)
        optimizer = tf.keras.optimizers.SGD(learning_rate=lr)
        model.compile(optimizer=optimizer,
                      loss="categorical_crossentropy",
                      metrics=["accuracy", "top_k_categorical_accuracy"])
        return model
示例#7
0
    def build(self, hp, inputs=None):
        input_node = nest.flatten(inputs)[0]
        output_node = input_node

        if self.translation_factor != 0 and self.translation_factor != (0, 0):
            height_factor, width_factor = self._get_fraction_value(
                self.translation_factor)
            output_node = preprocessing.RandomTranslation(
                height_factor, width_factor)(output_node)

        horizontal_flip = self.horizontal_flip
        if horizontal_flip is None:
            horizontal_flip = hp.Boolean('horizontal_flip', default=True)
        vertical_flip = self.vertical_flip
        if self.vertical_flip is None:
            vertical_flip = hp.Boolean('vertical_flip', default=True)
        if not horizontal_flip and not vertical_flip:
            flip_mode = ''
        elif horizontal_flip and vertical_flip:
            flip_mode = 'horizontal_and_vertical'
        elif horizontal_flip and not vertical_flip:
            flip_mode = 'horizontal'
        elif not horizontal_flip and vertical_flip:
            flip_mode = 'vertical'
        if flip_mode != '':
            output_node = preprocessing.RandomFlip(mode=flip_mode)(output_node)

        if self.rotation_factor != 0:
            output_node = preprocessing.RandomRotation(
                self.rotation_factor)(output_node)

        if self.zoom_factor != 0 and self.zoom_factor != (0, 0):
            height_factor, width_factor = self._get_fraction_value(
                self.zoom_factor)
            # TODO: Add back RandomZoom when it is ready.
            # output_node = preprocessing.RandomZoom(
            # height_factor, width_factor)(output_node)

        if self.contrast_factor != 0 and self.contrast_factor != (0, 0):
            output_node = preprocessing.RandomContrast(
                self.contrast_factor)(output_node)

        return output_node
示例#8
0
文件: model.py 项目: kotzyi/factory4
    def __init__(self, model_name, learning_rate, pre_trained_model_path):
        self.model_name = model_name
        self.lr = float(learning_rate)
        self.base_model, self.image_size = base_model[model_name]
        if model_name.startswith('b'):
            self.weight_path = os.path.join(pre_trained_model_path,
                                            self.model_name + ".h5")
        else:
            self.weight_path = 'imagenet'

        self.img_augmentation = Sequential(
            [
                preprocessing.RandomRotation(factor=0.15),
                preprocessing.RandomTranslation(height_factor=0.1,
                                                width_factor=0.1),
                preprocessing.RandomFlip(),
                preprocessing.RandomContrast(factor=0.1),
            ],
            name="img_augmentation",
        )
示例#9
0
def get_data_augmentation_layers(rotation: bool = False, flip: bool = False,
                                 zoom: bool = False, contrast: bool = False) -> List[PreprocessingLayer]:
    """
    Creates a list of augmentation layers which can be applied later.

    :param rotation: Data Augmentation: Whether to apply random rotation to the images.
    :param flip: Data Augmentation: Whether to apply random horizontal flip to the images.
    :param zoom: Data Augmentation:  Whether to apply random zoom to the images.
    :param contrast: Data Augmentation: Whether to apply random contrast enhancement to the images.
    :return: The list of data augmentation layers.
    """
    data_augmentation = []
    if rotation:
        data_augmentation.append(preprocessing.RandomRotation(factor=(1 / 6)))  # Between +/- 30deg
    if flip:
        data_augmentation.append(preprocessing.RandomFlip("horizontal"))
    if zoom:
        data_augmentation.append(preprocessing.RandomZoom(height_factor=0.2))  # Zoom +/- 20%
    if contrast:
        data_augmentation.append(preprocessing.RandomContrast(factor=0.1))

    return data_augmentation
示例#10
0
def build_model(num_classes, config):
    img_augmentation = Sequential(
        [
            preprocessing.RandomRotation(factor=0.1),
            preprocessing.RandomTranslation(height_factor=0.1,
                                            width_factor=0.1),
            preprocessing.RandomContrast(factor=0.1),
        ],
        name="img_augmentation",
    )

    inputs = layers.Input(shape=(config.img_size, config.img_size, 3))
    x = img_augmentation(inputs)
    model = EfficientNetB0(include_top=False,
                           input_tensor=x,
                           weights="imagenet")

    # Freeze the pretrained weights
    model.trainable = False

    # Rebuild top
    x = layers.GlobalAveragePooling2D(name="avg_pool")(model.output)
    x = layers.BatchNormalization()(x)

    top_dropout_rate = 0.3
    x = layers.Dropout(top_dropout_rate, name="top_dropout")(x)
    outputs = layers.Dense(num_classes, activation="softmax", name="pred")(x)

    # Compile
    model = tf.keras.Model(inputs, outputs, name="EfficientNet")
    optimizer = tf.keras.optimizers.Adam(learning_rate=config.learning_rate)
    model.compile(optimizer=optimizer,
                  loss=config.loss_function,
                  metrics=["accuracy", metrics.top_k_categorical_accuracy])

    return model
示例#11
0
def build_model(NUM_CLASSES, IMG_SIZE):
    """モデルの構築

    Args:
        NUM_CLASSES (int): 種類数
        IMG_SIZE (int): サイズ

    Returns:
        tf.keras.Model: モデル
        int: Adamのハイパーパラメータ
    """
    img_augmentation = Sequential([
        preprocessing.RandomRotation(factor=0.15),
        preprocessing.RandomTranslation(height_factor=0.1, width_factor=0.1),
        preprocessing.RandomFlip(),
        preprocessing.RandomContrast(factor=0.1)
    ],
                                  name='img_augmentation')
    inputs = Input(shape=(IMG_SIZE, IMG_SIZE, 3))
    x = img_augmentation(inputs)
    model = EfficientNetB3(include_top=False,
                           input_tensor=x,
                           weights='imagenet')
    model.trainable = False
    x = GlobalAveragePooling2D(name='avg_pool')(model.output)
    x = BatchNormalization(trainable=True)(x)
    top_dropout_rate = 0.2
    x = Dropout(top_dropout_rate, name='top_dropout')(x)
    outputs = Dense(NUM_CLASSES, activation='softmax', name='pred')(x)
    model = Model(inputs, outputs, name='EfficientNet')
    lr = 1e-4
    optimizer = Adam(learning_rate=lr)
    model.compile(optimizer=optimizer,
                  loss='categorical_crossentropy',
                  metrics=['accuracy'])
    return (model, lr)
示例#12
0

def create_dataset(filenames, batch_size):
    """Create dataset from tfrecords file
  :tfrecords_files: Mask to collect tfrecords file of dataset
  :returns: tf.data.Dataset
  """
    return tf.data.TFRecordDataset(filenames)\
      .map(parse_proto_example, num_parallel_calls=tf.data.AUTOTUNE)\
      .batch(batch_size)\
      .prefetch(tf.data.AUTOTUNE)


data_augmentation = tf.keras.Sequential([
    preprocessing.RandomRotation(0.1,
                                 fill_mode='constant',
                                 interpolation='nearest'),
    preprocessing.RandomCrop(RESIZE_TO, RESIZE_TO),
    preprocessing.RandomFlip(mode="horizontal")
])


def build_model():
    inputs = tf.keras.Input(shape=(SCALE_W, SCALE_H, 3))
    x = data_augmentation(inputs)
    x = EfficientNetB0(include_top=False, weights='imagenet', input_tensor=x)
    x.trainable = False
    x = tf.keras.layers.GlobalAveragePooling2D()(x.output)
    outputs = tf.keras.layers.Dense(NUM_CLASSES,
                                    activation=tf.keras.activations.softmax)(x)
    return tf.keras.Model(inputs=inputs, outputs=outputs)
示例#13
0
def main():
    # Loading data.
    # Load data from tensorflow_dataset (hereafter TFDS). Stanford Dogs
    # dataset is provided in TFDS as stanford_dogs. It features 20,580
    # images that belong to 120 classes of dog breeds (12,000 for
    # training and 8,580 for testing).
    # By simply changing dataset_name below, you may also try this
    # notebook for other datasets in TFDS such as cifar10, cifar100,
    # food101, etc. When the images are much smaller than the size of
    # Efficientnet input, we can simply upsample the input images. It
    # has been shown in Tan and Le, 2019 that transfer learning result
    # is better for increased resolution even if input images remain
    # small.
    # For TPU: if using TFDS datasets, a GCS bucket location is
    # required to save the datasets. For example:
    # tfds.load(dataset_name, data_dir="gs://example-bucket/datapath")
    # Also, both the current environment and the TPU service account
    # have proper access to the bucket. Alternatively, for small
    # datasets you may try loading data into the memory and use
    # tf.data.Dataset.from_tensor_slices().
    batch_size = 64

    dataset_name = "stanford_dogs"
    (ds_train, ds_test), ds_info = tfds.load(dataset_name,
                                             split=["train", "test"],
                                             with_info=True,
                                             as_supervised=True)
    num_classes = ds_info.features["label"].num_classes

    # When the dataset include images with various size, need to resize
    # them into a shared size. The Stanford Dogs dataset includes only
    # images at least 200x200 pixels in size. Here, resize the images
    # to the input size needed for EfficientNet.
    size = (IMG_SIZE, IMG_SIZE)
    ds_train = ds_train.map(lambda image, label:
                            (tf.image.resize(image, size), label))
    ds_test = ds_test.map(lambda image, label:
                          (tf.image.resize(image, size), label))

    # Visualize the data.
    # The following code shows the first 9 images with their labels.
    #'''
    def format_label(label):
        string_label = label_info.int2str(label)
        return string_label.split("-")[1]

    label_info = ds_info.features["labels"]
    for i, (image, label) in enumerate(ds_train.take(9)):
        ax = plt.subplot(3, 3, i + 1)
        plt.imshow(image.numpy().astype("uint8"))
        plt.title("{}".format(format_label(label)))
        plt.axis("off")
    #'''

    # Data augmentation.
    # Use preprocessing layers APIs for image augmentation.
    img_augmentation = Sequential(
        [
            preprocessing.RandomRotation(factor - 0.15),
            preprocessing.RandomTranslation(height_factor=0.1,
                                            width_factor=0.1),
            preprocessing.RandomFlip(),
            preprocessing.RandomContrast(factor=0.1),
        ],
        name="img_augmentation",
    )

    # This sequential model object can be used both as part of the
    # model built later, and as a function to preprocess data before
    # feeding into the model. Using them as a function makes it easy to
    # visualize the augmented images. Here, plot 9 examples of
    # augmentation result of a given figure.
    #'''
    for image, label in ds_train.take(1):
        for i in range(9):
            ax = plt.subplot(3, 3, i + 1)
            aug_img = img_augmentation(tf.expand_dims(image, axis=0))
            plt.imshow(image[0].numpy().astype("uint8"))
            plt.title("{}".format(format_label(label)))
            plt.axis("off")
    #'''

    # Prepare inputs.
    # Once verified the input data and augmentation are working
    # correctly, prepared dataset for training. The input data is
    # resized to uniform IMG_SIZE. The labels are put into a one-hot
    # (a.k.a categorical) encoding. The dataset is batched.
    # Note: prefetch and AUTOTUNE may in some situation improve
    # performance, but depends on environment and the specific dataset
    # used. See this guide for more information on data pipeline
    # performance.
    # One-hot / categorical encoding.
    def input_preprocess(image, label):
        label = tf.one_hot(label, num_classes)
        return image, label

    ds_train = ds_train.map(input_preprocess,
                            num_parallel_calls=tf.data.experimental.AUTOTUNE)
    ds_train = ds_train.batch(batch_size=batch_size, drop_remainder=True)
    ds_train = ds_train.prefetch(tf.data.experimental.AUTOTUNE)

    ds_test = ds_test.map(input_preprocess)
    ds_test = ds_test.batch(batch_size=batch_size, drop_remainder=True)

    # Train a model from scratch.
    # Build an EfficientNetB0 with 120 output classes, that is
    # initialized from scratch.
    # Note: the accuracy will increase very slowly and may overfit.
    with strategy.scope():
        inputs = la

    # Exit the program.
    exit(0)
示例#14
0
    x_train, y_train = prepare_all_data(config,
                                        x_train,
                                        y_train,
                                        input_shape,
                                        random_labels=config.random_labels)
    x_test, y_test = prepare_all_data(config, x_test, y_test, input_shape)
if input_shape[-1] > 1:
    means = x_train.mean(axis=(0, 1, 2))
    std = x_train.std(axis=(0, 1, 2))
    x_train = (x_train - means) / std
    x_test = (x_test - means) / std
print("LOADING OVER", flush=True)

augmenter = tf.keras.Sequential([
    preprocessing.RandomContrast(0.2),
    preprocessing.RandomRotation(0.05),  # = 12° rotation maximum
    preprocessing.RandomTranslation(0.13, 0.13,
                                    fill_mode='reflect'),  # nearest better
    preprocessing.RandomFlip('horizontal')
])
augmenter.build((None, ) + input_shape)


class CustomSequential(Sequential):
    def __init__(self, *args, **kwargs):
        super().__init__(*args, **kwargs)

    def train_step(self, data):
        x, y = data
        x = augmenter(x, training=True)
        with tf.GradientTape() as tape:
示例#15
0
    image = tf.image.convert_image_dtype(image, dtype=tf.float32)
    return image, label


AUTOTUNE = tf.data.experimental.AUTOTUNE
ds_train = (ds_train_.map(convert_to_float).cache().prefetch(
    buffer_size=AUTOTUNE))

# Data Augmentation

augment = keras.Sequential([
    preprocessing.RandomContrast(factor=0.5),
    preprocessing.RandomFlip(mode='horizontal'),  # meaning, left-to-right
    preprocessing.RandomFlip(mode='vertical'),  # meaning, top-to-bottom
    preprocessing.RandomWidth(factor=0.15),  # horizontal stretch
    preprocessing.RandomRotation(factor=0.20),
    preprocessing.RandomTranslation(height_factor=0.1, width_factor=0.1),
    preprocessing.RandomContrast(factor=0.10),
    preprocessing.RandomFlip(mode='horizontal'),
    preprocessing.RandomRotation(factor=0.10),
])

ex = next(iter(ds_train.unbatch().map(lambda x, y: x).batch(1)))

plt.figure(figsize=(10, 10))
for i in range(16):
    image = augment(ex, training=True)
    plt.subplot(4, 4, i + 1)
    plt.imshow(tf.squeeze(image))
    plt.axis('off')
plt.show()
示例#16
0
plt.figure(figsize=(16, 12))
for n in range(30):
    ax = plt.subplot(5, 6, n + 1)
    plt.imshow(img_test[n].astype("uint8"))
    plt.title(np.array(class_names)[label_test[n] == True][0])
    plt.axis("off")
"""
## Augmentation

Define image augmentation using keras preprocessing layers and apply them to the training set.
"""

# Define image augmentation model
image_augmentation = keras.Sequential([
    preprocessing.RandomFlip(mode="horizontal"),
    preprocessing.RandomRotation(factor=0.1),
    preprocessing.RandomZoom(height_factor=(-0.1, -0)),
    preprocessing.RandomContrast(factor=0.1),
], )

# Apply the augmentations to the training images and plot a few examples
img_train = image_augmentation(img_train).numpy()

plt.figure(figsize=(16, 12))
for n in range(30):
    ax = plt.subplot(5, 6, n + 1)
    plt.imshow(img_train[n].astype("uint8"))
    plt.title(np.array(class_names)[label_train[n] == True][0])
    plt.axis("off")
"""
## Define model building & training functions
示例#17
0
#                  )
hist = model.fit(batch_train_dataset,
                 epochs=epochs,
                 validation_data = batch_test_dataset,
                 verbose=2)

plot_hist(hist)

from tensorflow.keras.layers.experimental import preprocessing
from tensorflow.keras.models import Sequential
from tensorflow.keras import layers

img_augmentation = Sequential(
  [
    # representing lower and upper bound for rotating clockwise and counter-clockwise. 
    preprocessing.RandomRotation(factor=0.15), # a float represented as fraction of 2pi, ex :0.15 (= 54 degree!) 
    
    preprocessing.RandomTranslation(height_factor=0.1, # lower and upper bound for shifting vertically
                                    width_factor=0.1 #lower and upper bound for shifting horizontally.
                                    ),
    preprocessing.RandomFlip(), # Randomly flip each image horizontally and vertically.
    preprocessing.RandomContrast(factor=0.1),
  ],
  name="img_augmentation",
) # 각각의 layer가 순서대로? 실행된다고 생각

for image, label in train_dataset.take(1):
  for i in range(9):
    ax = plt.subplot(3, 3, i + 1)
    aug_img = img_augmentation(tf.expand_dims(image, axis=0))
    plt.imshow(aug_img[0].numpy().astype("uint8")) # 이동에 대한 ~가 줄어들고, 일반화 잘됨. 데이터는 많아지겠지. 학습시간도
check_point_path = '10_class_10_ptc_checkpoint'

check_point_dir = tf.keras.callbacks.ModelCheckpoint(check_point_path, monitor="val_accuracy", save_best_only=True,
                                                     save_weights_only=True)

from tensorflow.keras import layers
from tensorflow.keras.layers.experimental import preprocessing
from tensorflow.keras.models import Sequential

from tensorflow.keras import layers
from tensorflow.keras.layers.experimental import preprocessing
from tensorflow.keras.models import Sequential

data_augmentation = Sequential([
      preprocessing.RandomFlip("horizontal"),
      preprocessing.RandomRotation(0.2),
      preprocessing.RandomHeight(0.2),
      preprocessing.RandomZoom(0.2)                          
])

base_model = tf.keras.applications.EfficientNetB0(include_top=False)
base_model.trainable = False

input_layer = tf.keras.layers.Input(shape=(224, 224, 3))

x = data_augmentation(input_layer)

x = base_model(x)

x = tf.keras.layers.GlobalAveragePooling2D()(x)
outputs = tf.keras.layers.Dense(10, activation="softmax")(x)
def build_augmenters(imgaug=True):
    if imgaug:
        ia.seed(SEED)
        sometimes = lambda aug: iaa.Sometimes(0.8, aug, seed=SEED)

        augmenters = iaa.Sequential(
            [
                iaa.SomeOf(
                    (1, 3) if EXTRACT_IMAGES else (2, 4),
                    [
                        iaa.Add((-10, 10), per_channel=0.5, seed=SEED),
                        iaa.AdditiveGaussianNoise(loc=0,
                                                  scale=(0.0, 0.05 * 255),
                                                  per_channel=0.5,
                                                  seed=SEED),
                        iaa.OneOf(
                            [
                                iaa.GaussianBlur(sigma=(0, 0.5), seed=SEED),
                                iaa.AverageBlur(k=1, seed=SEED),
                                iaa.MedianBlur(k=1, seed=SEED),
                            ],
                            seed=SEED,
                        ),
                        iaa.LinearContrast(
                            (0.8, 1.2), per_channel=0.5, seed=SEED),
                    ],
                ),
                sometimes(
                    iaa.OneOf(
                        [
                            iaa.Fliplr(0.5, seed=SEED),
                            iaa.Flipud(0.2, seed=SEED),
                        ],
                        seed=SEED,
                    ), ),
                sometimes(
                    iaa.Affine(
                        scale={
                            "x": (0.5, 1) if EXTRACT_IMAGES else (0.8, 1.2),
                            "y": (0.5, 1) if EXTRACT_IMAGES else (0.8, 1.2)
                        },
                        translate_percent={
                            "x": (-0.01, 0.01) if EXTRACT_IMAGES else
                            (-0.2, 0.2),
                            "y": (-0.01, 0.01) if EXTRACT_IMAGES else
                            (-0.2, 0.2),
                        },
                        rotate=(-25, 25) if EXTRACT_IMAGES else (-45, 45),
                        shear=(-16, 16),
                        order=[0, 1],
                        cval=(0, 255),
                        mode=ia.ALL,
                        seed=SEED,
                    )),
            ],
            random_order=True,
            seed=SEED,
        )
        return augmenters
    else:
        img_augmentation = Sequential(
            [
                preprocessing.RandomRotation(factor=0.3, seed=SEED),
                preprocessing.RandomTranslation(
                    height_factor=0.01 if EXTRACT_IMAGES else 0.2,
                    width_factor=0.01 if EXTRACT_IMAGES else 0.2,
                    seed=SEED),
                preprocessing.RandomFlip(seed=SEED),
                preprocessing.RandomContrast(
                    factor=0.1 if EXTRACT_IMAGES else 0.2, seed=SEED),
            ],
            name="img_augmentation",
        )
        return img_augmentation
示例#20
0
class model:
    inputShape = [128, 128, 3]  #Unknown atm
    model = keras.Sequential([
        #Input
        layers.InputLayer(input_shape=inputShape),
        #Augment
        preprocessing.RandomRotation(factor=0.1),
        #Layer 1
        layers.BatchNormalization(renorm=True),
        layers.Conv2D(filters=10,
                      kernel_size=3,
                      activation='relu',
                      padding='same'),
        #Layer 2
        layers.BatchNormalization(renorm=True),
        layers.Conv2D(filters=20,
                      kernel_size=3,
                      activation='relu',
                      padding='same'),
        #Head
        layers.BatchNormalization(renorm=True),
        layers.Flatten(),
        layers.Dense(8, activation='relu'),
        layers.Dense(1, activation='sigmoid'),
    ])
    optimiser = keras.optimizers.SGD(lr=0.01, nesterov=True)
    model.compile(
        optimizer=optimiser,
        loss='binary_crossentropy',
        metrics=['binary_accuracy'],
    )
    early_stopping = keras.callbacks.EarlyStopping(
        patience=10,
        min_delta=0.001,
        restore_best_weights=True,
    )

    def __init__(self, **kwargs):
        overallDir = kwargs["dir"]
        ds_train_ = image_dataset_from_directory(
            overallDir + '/train',
            labels='inferred',
            label_mode='binary',
            image_size=[128, 128],
            interpolation='nearest',
            batch_size=64,
            shuffle=True,
        )
        ds_valid_ = image_dataset_from_directory(
            overallDir + '/valid',
            labels='inferred',
            label_mode='binary',
            image_size=[128, 128],
            interpolation='nearest',
            batch_size=64,
            shuffle=False,
        )

        def convert_to_float(image, label):
            image = tf.image.convert_image_dtype(image, dtype=tf.float32)
            return image, label

        AUTOTUNE = tf.data.experimental.AUTOTUNE
        ds_train = (ds_train_.map(convert_to_float).cache().prefetch(
            buffer_size=AUTOTUNE))
        ds_valid = (ds_valid_.map(convert_to_float).cache().prefetch(
            buffer_size=AUTOTUNE))
        self.his = self.model.fit(
            ds_train,
            validation_data=ds_valid,
            epochs=50,
            callbacks=[self.early_stopping],
        )
示例#21
0
"""
## Quick recipes

### Image data augmentation (on-device)

Note that image data augmentation layers are only active during training (similarly to
the `Dropout` layer).
"""

from tensorflow import keras
from tensorflow.keras import layers

# Create a data augmentation stage with horizontal flipping, rotations, zooms
data_augmentation = keras.Sequential([
    preprocessing.RandomFlip("horizontal"),
    preprocessing.RandomRotation(0.1),
    preprocessing.RandomZoom(0.1),
])

# Create a model that includes the augmentation stage
input_shape = (32, 32, 3)
classes = 10
inputs = keras.Input(shape=input_shape)
# Augment images
x = data_augmentation(inputs)
# Rescale image values to [0, 1]
x = preprocessing.Rescaling(1.0 / 255)(x)
# Add the rest of the model
outputs = keras.applications.ResNet50(weights=None,
                                      input_shape=input_shape,
                                      classes=classes)(x)
示例#22
0
        num_repeats = 50000 // n_train_images
        n_total = num_repeats * n_train_images
        train_dataset = train_dataset.repeat(num_repeats)

        # apply random rotation
        if category in textures:  # for textures: rotate and crop without edges

            def augmentation(x):  # slow computation, we will cache this
                return tf.py_function(random_rotation_crop_no_edges,
                                      inp=[x, rotation_range, patch_size],
                                      Tout=tf.float32)
        else:  # for objects: rotate (reflect edges), no crop
            factor = (rotation_range[0] / 360., rotation_range[1] / 360.)
            # fill_mode='nearest' is available only in tf-nightly (2.4.0)
            augmentation = P.RandomRotation(factor,
                                            fill_mode='nearest',
                                            seed=seed)

        train_dataset = train_dataset.shuffle(10000).batch(32)
        train_dataset = train_dataset.map(augmentation,
                                          num_parallel_calls=AUTOTUNE)
        train_dataset = train_dataset.unbatch()

        if category in textures:  # for textures: cache the augmented dataset
            print('Creating cache for dataset:', cache_file)
            train_dataset = tqdm(train_dataset.as_numpy_iterator(),
                                 total=n_total)
            train_dataset = np.stack(list(train_dataset))
            np.save(cache_file, train_dataset)
            train_dataset = tf.data.Dataset.from_tensor_slices(train_dataset)
    else:
示例#23
0
  example['image'] = tf.image.convert_image_dtype(example['image'], dtype=tf.uint8)
  example['image'] = tf.image.resize(example['image'], tf.constant([RESIZE_TO, RESIZE_TO]))
  return example['image'], tf.one_hot(example['image/label'], depth=NUM_CLASSES)  
  
def create_dataset(filenames, batch_size):
  """Create dataset from tfrecords file
  :tfrecords_files: Mask to collect tfrecords file of dataset
  :returns: tf.data.Dataset
  """
  return tf.data.TFRecordDataset(filenames)\
    .map(parse_proto_example, num_parallel_calls=tf.data.AUTOTUNE)\
    .batch(batch_size)\
    .prefetch(tf.data.AUTOTUNE)

data_augmentation = tf.keras.Sequential([
        preprocessing.RandomRotation(0.5, fill_mode='wrap', interpolation='nearest', seed=1)
    ])

def build_model():
  inputs = tf.keras.Input(shape=(RESIZE_TO, RESIZE_TO, 3))
  x = data_augmentation(inputs)
  x = EfficientNetB0(include_top=False, weights='imagenet', input_tensor = x)
  x.trainable = False
  x = tf.keras.layers.GlobalAveragePooling2D()(x.output)
  outputs = tf.keras.layers.Dense(NUM_CLASSES, activation=tf.keras.activations.softmax)(x)
  return tf.keras.Model(inputs=inputs, outputs=outputs)


def main():
  args = argparse.ArgumentParser()
  args.add_argument('--train', type=str, help='Glob pattern to collect train tfrecord files, use single quote to escape *')
示例#24
0
import tensorflow as tf
import math
import tensorflow.keras.backend as K

from tensorflow.keras import layers
from tensorflow.keras.layers.experimental import preprocessing
from tensorflow.keras.models import Sequential

from config import WIDTH, HEIGHT, CHANNELS, SEED

img_augmentation_sequential = Sequential(
    [
        preprocessing.RandomRotation(factor=0.15, seed=SEED),
        preprocessing.RandomTranslation(height_factor=0.1, width_factor=0.1),
        preprocessing.RandomFlip(),
        preprocessing.RandomContrast(factor=0.1),
        preprocessing.CenterCrop(height=HEIGHT, width=WIDTH),
        preprocessing.RandomZoom(
            height_factor=(0.2, 0.5),
            width_factor=(0.1, 0.3),  # None to preserve ratio
            fill_mode="constant",
            seed=SEED,
        ),
        preprocessing.Resizing(WIDTH, HEIGHT),
    ],
    name="img_augmentation",
)


def image_augmentation_functional(image, label):