示例#1
0
    def __init__(self):
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""
        self.verbose = 0

        self._whether_apply_bias = False
        self._current_dimension = 0
        self._cost_layer = "Undefined"

        self._logs = {}
        self._timings = {}
        self._metrics, self._metric_names = [], []

        self._x, self._y = None, None
        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

        self._layer_factory = LayerFactory()
        self._optimizer_factory = OptFactory()

        self._available_metrics = {
            "acc": NNDist._acc, "_acc": NNDist._acc,
            "f1": NNDist._f1_score, "_f1_score": NNDist._f1_score
        }
示例#2
0
文件: Networks.py 项目: F0nnys/ML
    def __init__(self, **kwargs):
        super(NNDist, self).__init__(**kwargs)
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""
        self.verbose = 1

        self._apply_bias = False
        self._current_dimension = 0

        self._logs = {}
        self._metrics, self._metric_names = [], []

        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

        self._layer_factory = LayerFactory()
        self._optimizer_factory = OptFactory()

        self._available_metrics = {
            "acc": self.acc,
            "_acc": self.acc,
            "f1": self.f1_score,
            "_f1_score": self.f1_score
        }
示例#3
0
    def fit(self,
            x,
            y,
            sample_weight=None,
            c=None,
            lr=None,
            optimizer=None,
            batch_size=None,
            epoch=None,
            tol=None,
            animation_params=None):
        if sample_weight is None:
            sample_weight = self._params["sample_weight"]
        if c is None:
            c = self._params["c"]
        if lr is None:
            lr = self._params["lr"]
        if batch_size is None:
            batch_size = self._params["batch_size"]
        if epoch is None:
            epoch = self._params["epoch"]
        if tol is None:
            tol = self._params["tol"]
        if optimizer is None:
            optimizer = self._params["optimizer"]
        *animation_properties, animation_params = self._get_animation_params(
            animation_params)
        x, y = np.atleast_2d(x), np.asarray(y, dtype=np.float32)
        if sample_weight is None:
            sample_weight = np.ones(len(y))
        else:
            sample_weight = np.asarray(sample_weight) * len(y)

        self._w = np.zeros(x.shape[1], dtype=np.float32)
        self._b = np.zeros(1, dtype=np.float32)
        self._model_parameters = [self._w, self._b]
        self._optimizer = OptFactory().get_optimizer_by_name(
            optimizer, self._model_parameters, lr, epoch)

        bar = ProgressBar(max_value=epoch, name="LinearSVM")
        ims = []
        train_repeat = self._get_train_repeat(x, batch_size)
        for i in range(epoch):
            self._optimizer.update()
            l = self._batch_training(x, y, batch_size, train_repeat,
                                     sample_weight, c)
            if l < tol:
                bar.terminate()
                break
            self._handle_animation(i, x, y, ims, animation_params,
                                   *animation_properties)
            bar.update()
        self._handle_mp4(ims, animation_properties)
示例#4
0
 def _prepare(self, sample_weight, **kwargs):
     lr = kwargs.get("lr", self._params["lr"])
     self._alpha = np.random.random(len(self._x)).astype(np.float32)
     self._b = np.random.random(1).astype(np.float32)
     self._model_parameters = [self._alpha, self._b]
     self._optimizer = OptFactory().get_optimizer_by_name(
         self._optimizer, self._model_parameters, lr, self._params["epoch"])
示例#5
0
    def fit(self, x, y, sample_weight=None, c=None, lr=None, optimizer=None,
            batch_size=None, epoch=None, tol=None, animation_params=None):
        if sample_weight is None:
            sample_weight = self._params["sample_weight"]
        if c is None:
            c = self._params["c"]
        if lr is None:
            lr = self._params["lr"]
        if batch_size is None:
            batch_size = self._params["batch_size"]
        if epoch is None:
            epoch = self._params["epoch"]
        if tol is None:
            tol = self._params["tol"]
        if optimizer is None:
            optimizer = self._params["optimizer"]
        *animation_properties, animation_params = self._get_animation_params(animation_params)
        x, y = np.atleast_2d(x), np.asarray(y, dtype=np.float32)
        if sample_weight is None:
            sample_weight = np.ones(len(y))
        else:
            sample_weight = np.asarray(sample_weight) * len(y)

        self._w = np.zeros(x.shape[1], dtype=np.float32)
        self._b = np.zeros(1, dtype=np.float32)
        self._model_parameters = [self._w, self._b]
        self._optimizer = OptFactory().get_optimizer_by_name(
            optimizer, self._model_parameters, lr, epoch
        )

        bar = ProgressBar(max_value=epoch, name="LinearSVM")
        ims = []
        train_repeat = self._get_train_repeat(x, batch_size)
        for i in range(epoch):
            self._optimizer.update()
            l = self._batch_training(
                x, y, batch_size, train_repeat, sample_weight, c
            )
            if l < tol:
                bar.terminate()
                break
            self._handle_animation(i, x, y, ims, animation_params, *animation_properties)
            bar.update()
        self._handle_mp4(ims, animation_properties)
示例#6
0
    def __init__(self, **kwargs):
        super(NNDist, self).__init__(**kwargs)
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""
        self.verbose = 1

        self._apply_bias = False
        self._current_dimension = 0

        self._logs = {}
        self._metrics, self._metric_names = [], []

        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

        self._layer_factory = LayerFactory()
        self._optimizer_factory = OptFactory()

        self._available_metrics = {
            "acc": self.acc, "_acc": self.acc,
            "f1": self.f1_score, "_f1_score": self.f1_score
        }
示例#7
0
class NNDist(ClassifierBase):
    NNTiming = Timing()

    def __init__(self, **kwargs):
        super(NNDist, self).__init__(**kwargs)
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""
        self.verbose = 1

        self._apply_bias = False
        self._current_dimension = 0

        self._logs = {}
        self._metrics, self._metric_names = [], []

        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

        self._layer_factory = LayerFactory()
        self._optimizer_factory = OptFactory()

        self._available_metrics = {
            "acc": self.acc, "_acc": self.acc,
            "f1": self.f1_score, "_f1_score": self.f1_score
        }

    @NNTiming.timeit(level=4, prefix="[Initialize] ")
    def initialize(self):
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""
        self.verbose = 1

        self._apply_bias = False
        self._current_dimension = 0

        self._logs = []
        self._metrics, self._metric_names = [], []

        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

    # Property

    @property
    def name(self):
        return (
            "-".join([str(_layer.shape[1]) for _layer in self._layers]) +
            " at {}".format(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime()))
        )

    @property
    def layer_names(self):
        return [layer.name for layer in self._layers]

    @layer_names.setter
    def layer_names(self, value):
        self._layer_names = value

    @property
    def layer_shapes(self):
        return [layer.shape for layer in self._layers]

    @layer_shapes.setter
    def layer_shapes(self, value):
        self._layer_shapes = value

    @property
    def layer_params(self):
        return self._layer_params

    @layer_params.setter
    def layer_params(self, value):
        self.layer_params = value

    @property
    def layer_special_params(self):
        return [layer.special_params for layer in self._layers]

    @layer_special_params.setter
    def layer_special_params(self, value):
        for layer, sp_param in zip(self._layers, value):
            if sp_param is not None:
                layer.set_special_params(sp_param)

    @property
    def optimizer(self):
        return self._optimizer_name

    @optimizer.setter
    def optimizer(self, value):
        try:
            self._optimizer_name = value
        except KeyError:
            raise BuildNetworkError("Invalid Optimizer '{}' provided".format(value))

    # Utils

    @NNTiming.timeit(level=4)
    def _get_min_max(self, x, y):
        self._x_min, self._x_max = np.min(x), np.max(x)
        self._y_min, self._y_max = np.min(y), np.max(y)

    @NNTiming.timeit(level=4)
    def _split_data(self, x, y, x_test, y_test,
                    train_only, training_scale=NNConfig.TRAINING_SCALE):
        x, y = np.asarray(x, dtype=np.float32), np.asarray(y, dtype=np.float32)
        if x_test is not None and y_test is not None:
            x_test, y_test = np.asarray(x_test, dtype=np.float32), np.asarray(y_test, dtype=np.float32)
        if train_only:
            if x_test is not None and y_test is not None:
                x, y = np.vstack((x, x_test)), np.vstack((y, y_test))
            x_train, y_train, x_test, y_test = x, y, x, y
        else:
            shuffle_suffix = np.random.permutation(len(x))
            x, y = x[shuffle_suffix], y[shuffle_suffix]
            if x_test is None or y_test is None:
                train_len = int(len(x) * training_scale)
                x_train, y_train = x[:train_len], y[:train_len]
                x_test, y_test = x[train_len:], y[train_len:]
            elif x_test is None or y_test is None:
                raise BuildNetworkError("Please provide test sets if you want to split data on your own")
            else:
                x_train, y_train = x, y
        if NNConfig.BOOST_LESS_SAMPLES:
            if y_train.shape[1] != 2:
                raise BuildNetworkError("It is not permitted to boost less samples in multiple classification")
            y_train_arg = np.argmax(y_train, axis=1)
            y0 = y_train_arg == 0
            y1 = ~y0
            y_len, y0_len = len(y_train), np.sum(y0)  # type: int
            if y0_len > int(0.5 * y_len):
                y0, y1 = y1, y0
                y0_len = y_len - y0_len
            boost_suffix = np.random.randint(y0_len, size=y_len - y0_len)
            x_train = np.vstack((x_train[y1], x_train[y0][boost_suffix]))
            y_train = np.vstack((y_train[y1], y_train[y0][boost_suffix]))
            shuffle_suffix = np.random.permutation(len(x_train))
            x_train, y_train = x_train[shuffle_suffix], y_train[shuffle_suffix]
        return (x_train, x_test), (y_train, y_test)

    @NNTiming.timeit(level=4)
    def _add_params(self, shape, conv_channel=None, fc_shape=None):
        if fc_shape is not None:
            self._weights.append(np.random.randn(fc_shape, shape[1]).astype(np.float32))
            self._bias.append(np.zeros((1, shape[1]), dtype=np.float32))
        elif conv_channel is not None:
            if len(shape[1]) <= 2:
                self._weights.append(np.random.randn(
                    conv_channel, conv_channel, shape[1][0], shape[1][1]).astype(np.float32)
                )
            else:
                self._weights.append(np.random.randn(
                    shape[1][0], conv_channel, shape[1][1], shape[1][2]).astype(np.float32)
                )
            self._bias.append(np.zeros((1, shape[1][0]), dtype=np.float32))
        else:
            self._weights.append(np.random.randn(*shape).astype(np.float32))
            self._bias.append(np.zeros((1, shape[1]), dtype=np.float32))

    @NNTiming.timeit(level=4)
    def _add_layer(self, layer, *args, **kwargs):
        if not self._layers and isinstance(layer, str):
            layer = self._layer_factory.get_root_layer_by_name(layer, *args, **kwargs)
            if layer:
                self.add(layer)
                return
        parent = self._layers[-1]
        if isinstance(parent, CostLayer):
            raise BuildLayerError("Adding layer after CostLayer is not permitted")
        if isinstance(layer, str):
            layer, shape = self._layer_factory.get_layer_by_name(
                layer, parent, self._current_dimension, *args, **kwargs
            )
            if shape is None:
                self.add(layer)
                return
            _current, _next = shape
        else:
            _current, _next = args
        if isinstance(layer, SubLayer):
            parent.child = layer
            layer.is_sub_layer = True
            layer.root = layer.root
            layer.root.last_sub_layer = layer
            self.parent = parent
            self._layers.append(layer)
            self._weights.append(np.array([.0]))
            self._bias.append(np.array([.0]))
            self._current_dimension = _next
        else:
            fc_shape, conv_channel, last_layer = None, None, self._layers[-1]
            if isinstance(last_layer, ConvLayer):
                if isinstance(layer, ConvLayer):
                    conv_channel = last_layer.n_filters
                    _current = (conv_channel, last_layer.out_h, last_layer.out_w)
                    layer.feed_shape((_current, _next))
                else:
                    layer.is_fc = True
                    last_layer.is_fc_base = True
                    fc_shape = last_layer.out_h * last_layer.out_w * last_layer.n_filters
            self._layers.append(layer)
            self._add_params((_current, _next), conv_channel, fc_shape)
            self._current_dimension = _next
        self._update_layer_information(layer)

    @NNTiming.timeit(level=4)
    def _update_layer_information(self, layer):
        self._layer_params.append(layer.params)
        if len(self._layer_params) > 1 and not layer.is_sub_layer:
            self._layer_params[-1] = ((self._layer_params[-1][0][1],), *self._layer_params[-1][1:])

    @NNTiming.timeit(level=1)
    def _get_prediction(self, x, name=None, batch_size=1e6, verbose=None):
        if verbose is None:
            verbose = self.verbose
        fc_shape = np.prod(x.shape[1:])  # type: int
        single_batch = int(batch_size / fc_shape)
        if not single_batch:
            single_batch = 1
        if single_batch >= len(x):
            return self._get_activations(x, predict=True).pop()
        epoch = int(len(x) / single_batch)
        if not len(x) % single_batch:
            epoch += 1
        name = "Prediction" if name is None else "Prediction ({})".format(name)
        sub_bar = ProgressBar(max_value=epoch, name=name, start=False)
        if verbose >= NNVerbose.METRICS:
            sub_bar.start()
        rs, count = [self._get_activations(x[:single_batch], predict=True).pop()], single_batch
        if verbose >= NNVerbose.METRICS:
            sub_bar.update()
        while count < len(x):
            count += single_batch
            if count >= len(x):
                rs.append(self._get_activations(x[count-single_batch:], predict=True).pop())
            else:
                rs.append(self._get_activations(x[count-single_batch:count], predict=True).pop())
            if verbose >= NNVerbose.METRICS:
                sub_bar.update()
        return np.vstack(rs)

    @NNTiming.timeit(level=1)
    def _get_activations(self, x, predict=False):
        activations = [self._layers[0].activate(x, self._weights[0], self._bias[0], predict)]
        for i, layer in enumerate(self._layers[1:]):
            activations.append(layer.activate(
                activations[-1], self._weights[i + 1], self._bias[i + 1], predict))
        return activations

    @NNTiming.timeit(level=3)
    def _append_log(self, x, y, name, get_loss=True):
        y_pred = self._get_prediction(x, name)
        for i, metric in enumerate(self._metrics):
            self._logs[name][i].append(metric(
                np.argmax(y, axis=1), np.argmax(y_pred, axis=1)
            ))
        if get_loss:
            self._logs[name][-1].append(self._layers[-1].calculate(y, y_pred) / len(y))

    @NNTiming.timeit(level=3)
    def _print_metric_logs(self, show_loss, data_type):
        print()
        print("=" * 47)
        for i, name in enumerate(self._metric_names):
            print("{:<16s} {:<16s}: {:12.8}".format(
                data_type, name, self._logs[data_type][i][-1]))
        if show_loss:
            print("{:<16s} {:<16s}: {:12.8}".format(
                data_type, "loss", self._logs[data_type][-1][-1]))
        print("=" * 47)

    @NNTiming.timeit(level=1)
    def _draw_2d_network(self, radius=6, width=1200, height=800, padding=0.2,
                         plot_scale=2, plot_precision=0.03,
                         sub_layer_height_scale=0, **kwargs):
        if not kwargs["show"] and not kwargs["mp4"]:
            return
        layers = len(self._layers) + 1
        units = [layer.shape[0] for layer in self._layers] + [self._layers[-1].shape[1]]
        whether_sub_layers = np.array([False] + [isinstance(layer, SubLayer) for layer in self._layers])
        n_sub_layers = np.sum(whether_sub_layers)  # type: int

        plot_num = int(1 / plot_precision)
        if plot_num % 2 == 1:
            plot_num += 1
        half_plot_num = int(plot_num * 0.5)
        xf = np.linspace(self._x_min * plot_scale, self._x_max * plot_scale, plot_num)
        yf = np.linspace(self._x_min * plot_scale, self._x_max * plot_scale, plot_num) * -1
        input_x, input_y = np.meshgrid(xf, yf)
        input_xs = np.c_[input_x.ravel(), input_y.ravel()]

        activations = [activation.T.reshape(units[i + 1], plot_num, plot_num)
                       for i, activation in enumerate(self._get_activations(input_xs, predict=True))]
        graphs = []
        for j, activation in enumerate(activations):
            graph_group = []
            if j == len(activations) - 1:
                classes = np.argmax(activation, axis=0)
            else:
                classes = None
            for k, ac in enumerate(activation):
                data = np.zeros((plot_num, plot_num, 3), np.uint8)
                if j != len(activations) - 1:
                    mask = ac >= np.average(ac)
                else:
                    mask = classes == k
                data[mask], data[~mask] = [0, 165, 255], [255, 165, 0]
                graph_group.append(data)
            graphs.append(graph_group)

        img = np.full([height, width, 3], 255, dtype=np.uint8)
        axis0_padding = int(height / (layers - 1 + 2 * padding)) * padding + plot_num
        axis0_step = (height - 2 * axis0_padding) / layers
        sub_layer_decrease = int((1 - sub_layer_height_scale) * axis0_step)
        axis0 = np.linspace(
            axis0_padding,
            height + n_sub_layers * sub_layer_decrease - axis0_padding,
            layers, dtype=np.int)
        axis0 -= sub_layer_decrease * np.cumsum(whether_sub_layers)
        axis1_padding = plot_num
        axis1 = [np.linspace(axis1_padding, width - axis1_padding, unit + 2, dtype=np.int)
                 for unit in units]
        axis1 = [axis[1:-1] for axis in axis1]

        colors, thicknesses, masks = [], [], []
        for weight in self._weights:
            line_info = VisUtil.get_line_info(weight.copy())
            colors.append(line_info[0])
            thicknesses.append(line_info[1])
            masks.append(line_info[2])

        for i, (y, xs) in enumerate(zip(axis0, axis1)):
            for j, x in enumerate(xs):
                if i == 0:
                    cv2.circle(img, (x, y), radius, (20, 215, 20), int(radius / 2))
                else:
                    graph = graphs[i - 1][j]
                    img[y - half_plot_num:y + half_plot_num, x - half_plot_num:x + half_plot_num] = graph
            if i > 0:
                cv2.putText(img, self._layers[i - 1].name, (12, y - 36), cv2.LINE_AA, 0.6, (0, 0, 0), 1)
        for i, y in enumerate(axis0):
            if i == len(axis0) - 1:
                break
            for j, x in enumerate(axis1[i]):
                new_y = axis0[i + 1]
                whether_sub_layer = isinstance(self._layers[i], SubLayer)
                for k, new_x in enumerate(axis1[i + 1]):
                    if whether_sub_layer and j != k:
                        continue
                    if masks[i][j][k]:
                        cv2.line(img, (x, y + half_plot_num), (new_x, new_y - half_plot_num),
                                 colors[i][j][k], thicknesses[i][j][k])

        return img

    # Optimizing Process

    @NNTiming.timeit(level=4)
    def _init_optimizer(self):
        if not isinstance(self._w_optimizer, Optimizer):
            self._w_optimizer = self._optimizer_factory.get_optimizer_by_name(
                self._w_optimizer, self._weights, self._lr, self._epoch)
        if not isinstance(self._b_optimizer, Optimizer):
            self._b_optimizer = self._optimizer_factory.get_optimizer_by_name(
                self._b_optimizer, self._bias, self._lr, self._epoch)
        if self._w_optimizer.name != self._b_optimizer.name:
            self._optimizer_name = None
        else:
            self._optimizer_name = self._w_optimizer.name

    @NNTiming.timeit(level=1)
    def _opt(self, i, activation, delta):
        if not isinstance(self._layers[i], ConvLayer):
            self._weights[i] *= self._regularization_param
            self._weights[i] += self._w_optimizer.run(
                i, activation.reshape(activation.shape[0], -1).T.dot(delta)
            )
            if self._apply_bias:
                self._bias[i] += self._b_optimizer.run(
                    i, np.sum(delta, axis=0, keepdims=True)
                )
        else:
            self._weights[i] *= self._regularization_param
            if delta[1] is not None:
                self._weights[i] += self._w_optimizer.run(i, delta[1])
            if self._apply_bias and delta[2] is not None:
                self._bias[i] += self._b_optimizer.run(i, delta[2])

    # API

    @NNTiming.timeit(level=4, prefix="[API] ")
    def add(self, layer, *args, **kwargs):
        if isinstance(layer, str):
            # noinspection PyTypeChecker
            self._add_layer(layer, *args, **kwargs)
        else:
            if not isinstance(layer, Layer):
                raise BuildLayerError("Invalid Layer provided (should be subclass of Layer)")
            if not self._layers:
                if isinstance(layer, SubLayer):
                    raise BuildLayerError("Invalid Layer provided (first layer should not be subclass of SubLayer)")
                if len(layer.shape) != 2:
                    raise BuildLayerError("Invalid input Layer provided (shape should be {}, {} found)".format(
                        2, len(layer.shape)
                    ))
                self._layers, self._current_dimension = [layer], layer.shape[1]
                self._update_layer_information(layer)
                if isinstance(layer, ConvLayer):
                    self._add_params(layer.shape, layer.n_channels)
                else:
                    self._add_params(layer.shape)
            else:
                if len(layer.shape) > 2:
                    raise BuildLayerError("Invalid Layer provided (shape should be {}, {} found)".format(
                        2, len(layer.shape)
                    ))
                if len(layer.shape) == 2:
                    _current, _next = layer.shape
                    if isinstance(layer, SubLayer):
                        if _next != self._current_dimension:
                            raise BuildLayerError("Invalid SubLayer provided (shape[1] should be {}, {} found)".format(
                                self._current_dimension, _next
                            ))
                    elif not isinstance(layer, ConvLayer) and _current != self._current_dimension:
                        raise BuildLayerError("Invalid Layer provided (shape[0] should be {}, {} found)".format(
                            self._current_dimension, _current
                        ))
                    self._add_layer(layer, _current, _next)

                elif len(layer.shape) == 1:
                    _next = layer.shape[0]
                    layer.shape = (self._current_dimension, _next)
                    self._add_layer(layer, self._current_dimension, _next)
                else:
                    raise LayerError("Invalid Layer provided (invalid shape '{}' found)".format(layer.shape))

    @NNTiming.timeit(level=4, prefix="[API] ")
    def build(self, units="build"):
        if isinstance(units, str):
            if units == "build":
                for name, param in zip(self._layer_names, self._layer_params):
                    self.add(name, *param)
            else:
                raise NotImplementedError("Invalid param '{}' provided to 'build' method".format(units))
        else:
            try:
                units = np.asarray(units).flatten().astype(np.int)
            except ValueError as err:
                raise BuildLayerError(err)
            if len(units) < 2:
                raise BuildLayerError("At least 2 layers are needed")
            _input_shape = (units[0], units[1])
            self.initialize()
            self.add(ReLU(_input_shape))
            for unit_num in units[2:-1]:
                self.add(ReLU((unit_num,)))
            self.add("CrossEntropy", (units[-1],))

    @NNTiming.timeit(level=4, prefix="[API] ")
    def preview(self):
        if not self._layers:
            rs = "None"
        else:
            rs = (
                "Input  :  {:<10s} - {}\n".format("Dimension", self._layers[0].shape[0]) +
                "\n".join([
                    "Layer  :  {:<10s} - {} {}".format(
                        layer.name, layer.shape[1], layer.description
                    ) if isinstance(layer, SubLayer) else
                    "Layer  :  {:<10s} - {:<14s} - strides: {:2d} - padding: {:2d} - out: {}".format(
                        layer.name, str(layer.shape[1]), layer.stride, layer.padding,
                        (layer.n_filters, layer.out_h, layer.out_w)
                    ) if isinstance(layer, ConvLayer) else "Layer  :  {:<10s} - {}".format(
                        layer.name, layer.shape[1]
                    ) for layer in self._layers[:-1]
                ]) + "\nCost   :  {:<16s}".format(str(self._layers[-1]))
            )
        print("=" * 30 + "\n" + "Structure\n" + "-" * 30 + "\n" + rs + "\n" + "-" * 30 + "\n")

    @NNTiming.timeit(level=1, prefix="[API] ")
    def fit(self,
            x=None, y=None, x_test=None, y_test=None,
            batch_size=128, record_period=1, train_only=False,
            optimizer=None, w_optimizer=None, b_optimizer=None,
            lr=0.001, lb=0.001, epoch=20, weight_scale=1, apply_bias=True,
            show_loss=True, metrics=None, do_log=True, verbose=None,
            visualize=False, visualize_setting=None,
            draw_weights=False, animation_params=None):
        self._lr, self._epoch = lr, epoch
        for weight in self._weights:
            weight *= weight_scale
        if not self._w_optimizer or not self._b_optimizer:
            if not self._optimizer_name:
                if optimizer is None:
                    optimizer = "Adam"
                self._w_optimizer = optimizer if w_optimizer is None else w_optimizer
                self._b_optimizer = optimizer if b_optimizer is None else b_optimizer
            else:
                if not self._w_optimizer:
                    self._w_optimizer = self._optimizer_name
                if not self._b_optimizer:
                    self._b_optimizer = self._optimizer_name
        self._init_optimizer()
        assert isinstance(self._w_optimizer, Optimizer) and isinstance(self._b_optimizer, Optimizer)
        print()
        print("=" * 30)
        print("Optimizers")
        print("-" * 30)
        print("w: {}\nb: {}".format(self._w_optimizer, self._b_optimizer))
        print("-" * 30)
        if not self._layers:
            raise BuildNetworkError("Please provide layers before fitting data")
        if y.shape[1] != self._current_dimension:
            raise BuildNetworkError("Output layer's shape should be {}, {} found".format(
                self._current_dimension, y.shape[1]))

        (x_train, x_test), (y_train, y_test) = self._split_data(
            x, y, x_test, y_test, train_only)
        train_len = len(x_train)
        batch_size = min(batch_size, train_len)
        do_random_batch = train_len > batch_size
        train_repeat = 1 if not do_random_batch else int(train_len / batch_size) + 1
        self._regularization_param = 1 - lb * lr / batch_size
        self._get_min_max(x_train, y_train)

        self._metrics = ["acc"] if metrics is None else metrics
        for i, metric in enumerate(self._metrics):
            if isinstance(metric, str):
                if metric not in self._available_metrics:
                    raise BuildNetworkError("Metric '{}' is not implemented".format(metric))
                self._metrics[i] = self._available_metrics[metric]
        self._metric_names = [_m.__name__ for _m in self._metrics]

        self._logs = {
            name: [[] for _ in range(len(self._metrics) + 1)] for name in ("train", "cv", "test")
        }
        if verbose is not None:
            self.verbose = verbose

        layer_width = len(self._layers)
        self._apply_bias = apply_bias

        bar = ProgressBar(max_value=max(1, epoch // record_period), name="Epoch", start=False)
        if self.verbose >= NNVerbose.EPOCH:
            bar.start()
        img, ims = None, []

        if draw_weights:
            weight_trace = [[[org] for org in weight] for weight in self._weights]
        else:
            weight_trace = []

        *animation_properties, animation_params = self._get_animation_params(animation_params)
        sub_bar = ProgressBar(max_value=train_repeat * record_period - 1, name="Iteration", start=False)
        for counter in range(epoch):
            self._w_optimizer.update()
            self._b_optimizer.update()
            if self.verbose >= NNVerbose.ITER and counter % record_period == 0:
                sub_bar.start()
            for _ in range(train_repeat):
                if do_random_batch:
                    batch = np.random.choice(train_len, batch_size)
                    x_batch, y_batch = x_train[batch], y_train[batch]
                else:
                    x_batch, y_batch = x_train, y_train
                activations = self._get_activations(x_batch)

                deltas = [self._layers[-1].bp_first(y_batch, activations[-1])]
                for i in range(-1, -len(activations), -1):
                    deltas.append(self._layers[i - 1].bp(activations[i - 1], self._weights[i], deltas[-1]))

                for i in range(layer_width - 1, 0, -1):
                    if not isinstance(self._layers[i], SubLayer):
                        self._opt(i, activations[i - 1], deltas[layer_width - i - 1])
                self._opt(0, x_batch, deltas[-1])

                if draw_weights:
                    for i, weight in enumerate(self._weights):
                        for j, new_weight in enumerate(weight.copy()):
                            weight_trace[i][j].append(new_weight)
                if self.verbose >= NNVerbose.DEBUG:
                    pass
                if self.verbose >= NNVerbose.ITER:
                    if sub_bar.update() and self.verbose >= NNVerbose.METRICS_DETAIL:
                        self._append_log(x, y, "train", get_loss=show_loss)
                        self._append_log(x_test, y_test, "cv", get_loss=show_loss)
                        self._print_metric_logs(show_loss, "train")
                        self._print_metric_logs(show_loss, "cv")
            if self.verbose >= NNVerbose.ITER:
                sub_bar.update()
            self._handle_animation(
                counter, x, y, ims, animation_params, *animation_properties,
                img=self._draw_2d_network(**animation_params), name="Neural Network"
            )
            if do_log:
                self._append_log(x, y, "train", get_loss=show_loss)
                self._append_log(x_test, y_test, "cv", get_loss=show_loss)
            if (counter + 1) % record_period == 0:
                if do_log and self.verbose >= NNVerbose.METRICS:
                    self._print_metric_logs(show_loss, "train")
                    self._print_metric_logs(show_loss, "cv")
                if visualize:
                    if visualize_setting is None:
                        self.visualize2d(x_test, y_test)
                    else:
                        self.visualize2d(x_test, y_test, *visualize_setting)
                if self.verbose >= NNVerbose.EPOCH:
                    bar.update(counter // record_period + 1)
                    if self.verbose >= NNVerbose.ITER:
                        sub_bar = ProgressBar(max_value=train_repeat * record_period - 1, name="Iteration", start=False)

        if do_log:
            self._append_log(x_test, y_test, "test", get_loss=show_loss)
        if img is not None:
            cv2.waitKey(0)
            cv2.destroyAllWindows()
        if draw_weights:
            ts = np.arange(epoch * train_repeat + 1)
            for i, weight in enumerate(self._weights):
                plt.figure()
                for j in range(len(weight)):
                    plt.plot(ts, weight_trace[i][j])
                plt.title("Weights toward layer {} ({})".format(i + 1, self._layers[i].name))
                plt.show()
        self._handle_mp4(ims, animation_properties, "NN")
        return self._logs

    @NNTiming.timeit(level=2, prefix="[API] ")
    def save(self, path=None, name=None, overwrite=True):
        path = os.path.join("Models", "Cache") if path is None else os.path.join("Models", path)
        name = "Model.nn" if name is None else name
        if not os.path.exists(path):
            os.makedirs(path)
        _dir = os.path.join(path, name)
        if not overwrite and os.path.isfile(_dir):
            _count = 1
            _new_dir = _dir + "({})".format(_count)
            while os.path.isfile(_new_dir):
                _count += 1
                _new_dir = _dir + "({})".format(_count)
            _dir = _new_dir
        print()
        print("=" * 60)
        print("Saving Model to {}...".format(_dir))
        print("-" * 60)
        with open(_dir, "wb") as file:
            pickle.dump({
                "structures": {
                    "_layer_names": self.layer_names,
                    "_layer_params": self._layer_params,
                    "_cost_layer": self._layers[-1].name,
                    "_next_dimension": self._current_dimension
                },
                "params": {
                    "_logs": self._logs,
                    "_metric_names": self._metric_names,
                    "_weights": self._weights,
                    "_bias": self._bias,
                    "_optimizer_name": self._optimizer_name,
                    "_w_optimizer": self._w_optimizer,
                    "_b_optimizer": self._b_optimizer,
                    "layer_special_params": self.layer_special_params,
                }
            }, file)
        print("Done")
        print("=" * 60)

    @NNTiming.timeit(level=2, prefix="[API] ")
    def load(self, path=os.path.join("Models", "Cache", "Model.nn")):
        self.initialize()
        try:
            with open(path, "rb") as file:
                dic = pickle.load(file)
                for key, value in dic["structures"].items():
                    setattr(self, key, value)
                self.build()
                for key, value in dic["params"].items():
                    setattr(self, key, value)
                self._init_optimizer()
                for i in range(len(self._metric_names) - 1, -1, -1):
                    name = self._metric_names[i]
                    if name not in self._available_metrics:
                        self._metric_names.pop(i)
                    else:
                        self._metrics.insert(0, self._available_metrics[name])
                print()
                print("=" * 30)
                print("Model restored")
                print("=" * 30)
                return dic
        except Exception as err:
            raise BuildNetworkError("Failed to load Network ({}), structure initialized".format(err))

    @NNTiming.timeit(level=4, prefix="[API] ")
    def predict(self, x, get_raw_results=False, **kwargs):
        x = np.asarray(x)
        if len(x.shape) == 1:
            x = x.reshape(1, -1)
        y_pred = self._get_prediction(x)
        return y_pred if get_raw_results else np.argmax(y_pred, axis=1)

    def draw_results(self):
        metrics_log, loss_log = {}, {}
        for key, value in sorted(self._logs.items()):
            metrics_log[key], loss_log[key] = value[:-1], value[-1]

        for i, name in enumerate(sorted(self._metric_names)):
            plt.figure()
            plt.title("Metric Type: {}".format(name))
            for key, log in sorted(metrics_log.items()):
                if key == "test":
                    continue
                xs = np.arange(len(log[i])) + 1
                plt.plot(xs, log[i], label="Data Type: {}".format(key))
            plt.legend(loc=4)
            plt.show()
            plt.close()

        plt.figure()
        plt.title("Cost")
        for key, loss in sorted(loss_log.items()):
            if key == "test":
                continue
            xs = np.arange(len(loss)) + 1
            plt.plot(xs, loss, label="Data Type: {}".format(key))
        plt.legend()
        plt.show()

    def draw_conv_weights(self):
        for i, (name, weight) in enumerate(zip(self.layer_names, self._weights)):
            if len(weight.shape) != 4:
                return
            for j, _w in enumerate(weight):
                for k, _ww in enumerate(_w):
                    VisUtil.show_img(_ww, "{} {} filter {} channel {}".format(name, i+1, j+1, k+1))

    def draw_conv_series(self, x, shape=None):
        for xx in x:
            VisUtil.show_img(VisUtil.trans_img(xx, shape), "Original")
            activations = self._get_activations(np.array([xx]), predict=True)
            for i, (layer, ac) in enumerate(zip(self._layers, activations)):
                if len(ac.shape) == 4:
                    for n in ac:
                        _n, height, width = n.shape
                        a = int(ceil(sqrt(_n)))
                        g = np.ones((a * height + a, a * width + a), n.dtype)
                        g *= np.min(n)
                        _i = 0
                        for y in range(a):
                            for x in range(a):
                                if _i < _n:
                                    g[y * height + y:(y + 1) * height + y, x * width + x:(x + 1) * width + x] = n[
                                        _i, :, :]
                                    _i += 1
                        # normalize to [0,1]
                        max_g = g.max()
                        min_g = g.min()
                        g = (g - min_g) / (max_g - min_g)
                        VisUtil.show_img(g, "Layer {} ({})".format(i + 1, layer.name))
                else:
                    ac = ac[0]
                    length = sqrt(np.prod(ac.shape))
                    if length < 10:
                        continue
                    (height, width) = xx.shape[1:] if shape is None else shape[1:]
                    sqrt_shape = sqrt(height * width)
                    oh, ow = int(length * height / sqrt_shape), int(length * width / sqrt_shape)
                    VisUtil.show_img(ac[:oh*ow].reshape(oh, ow), "Layer {} ({})".format(i + 1, layer.name))
示例#8
0
class NNDist:
    NNTiming = Timing()

    def __init__(self):
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""
        self.verbose = 0

        self._whether_apply_bias = False
        self._current_dimension = 0
        self._cost_layer = "Undefined"

        self._logs = {}
        self._timings = {}
        self._metrics, self._metric_names = [], []

        self._x, self._y = None, None
        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

        self._layer_factory = LayerFactory()
        self._optimizer_factory = OptFactory()

        self._available_metrics = {
            "acc": NNDist._acc, "_acc": NNDist._acc,
            "f1": NNDist._f1_score, "_f1_score": NNDist._f1_score
        }

    def __getitem__(self, item):
        if isinstance(item, int):
            if item < 0 or item >= len(self._layers):
                return
            bias = self._bias[item]
            return {
                "name": self._layers[item].name,
                "weight": self._weights[item],
                "bias": bias
            }
        if isinstance(item, str):
            return getattr(self, "_" + item)
        return

    def __str__(self):
        return "Neural Network"

    __repr__ = __str__

    @NNTiming.timeit(level=4, prefix="[Initialize] ")
    def initialize(self):
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""

        self._whether_apply_bias = False
        self._current_dimension = 0
        self._cost_layer = "Undefined"

        self._logs = []
        self._timings = {}
        self._metrics, self._metric_names = [], []

        self._x, self._y = None, None
        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

    @NNTiming.timeit(level=4, prefix="[API] ")
    def feed_timing(self, timing):
        if isinstance(timing, Timing):
            self.NNTiming = timing
            for layer in self._layers:
                layer.feed_timing(timing)

    # Property

    @property
    def name(self):
        return (
            "-".join([str(_layer.shape[1]) for _layer in self._layers]) +
            " at {}".format(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime()))
        )

    @property
    def layer_names(self):
        return [layer.name for layer in self._layers]

    @layer_names.setter
    def layer_names(self, value):
        self._layer_names = value

    @property
    def layer_shapes(self):
        return [layer.shape for layer in self._layers]

    @layer_shapes.setter
    def layer_shapes(self, value):
        self._layer_shapes = value

    @property
    def layer_params(self):
        return self._layer_params

    @layer_params.setter
    def layer_params(self, value):
        self.layer_params = value

    @property
    def layer_special_params(self):
        return [layer.special_params for layer in self._layers]

    @layer_special_params.setter
    def layer_special_params(self, value):
        for layer, sp_param in zip(self._layers, value):
            if sp_param is not None:
                layer.set_special_params(sp_param)

    @property
    def optimizer(self):
        return self._optimizer_name

    @optimizer.setter
    def optimizer(self, value):
        try:
            self._optimizer_name = value
        except KeyError:
            raise BuildNetworkError("Invalid Optimizer '{}' provided".format(value))

    # Utils

    @NNTiming.timeit(level=4)
    def _feed_data(self, x, y):
        if x is None:
            if self._x is None:
                raise BuildNetworkError("Please provide input matrix")
            x = self._x
        if y is None:
            if self._y is None:
                raise BuildNetworkError("Please provide input matrix")
            y = self._y
        if len(x) != len(y):
            raise BuildNetworkError("Data fed to network should be identical in length, x: {} and y: {} found".format(
                len(x), len(y)
            ))
        self._x, self._y = x, y
        self._x_min, self._x_max = np.min(x), np.max(x)
        self._y_min, self._y_max = np.min(y), np.max(y)
        return x, y

    @NNTiming.timeit(level=4)
    def _add_weight(self, shape, conv_channel=None, fc_shape=None):
        if fc_shape is not None:
            self._weights.append(np.random.randn(fc_shape, shape[1]))
            self._bias.append(np.zeros((1, shape[1])))
        elif conv_channel is not None:
            if len(shape[1]) <= 2:
                self._weights.append(np.random.randn(conv_channel, conv_channel, shape[1][0], shape[1][1]))
            else:
                self._weights.append(np.random.randn(shape[1][0], conv_channel, shape[1][1], shape[1][2]))
            self._bias.append(np.zeros((1, shape[1][0])))
        else:
            self._weights.append(np.random.randn(*shape))
            self._bias.append(np.zeros((1, shape[1])))

    @NNTiming.timeit(level=4)
    def _add_layer(self, layer, *args, **kwargs):
        if not self._layers and isinstance(layer, str):
            _layer = self._layer_factory.handle_str_main_layers(layer, *args, **kwargs)
            if _layer:
                self.add(_layer); return
        _parent = self._layers[-1]
        if isinstance(_parent, CostLayer):
            raise BuildLayerError("Adding layer after CostLayer is not permitted")
        if isinstance(layer, str):
            layer, shape = self._layer_factory.get_layer_by_name(
                layer, _parent, self._current_dimension, *args, **kwargs
            )
            if shape is None:
                self.add(layer); return
            _current, _next = shape
        else:
            _current, _next = args
        if isinstance(layer, SubLayer):
            if not isinstance(layer, CostLayer) and _current != _parent.shape[1]:
                raise BuildLayerError("Output shape should be identical with input shape "
                                      "if chosen SubLayer is not a CostLayer")
            _parent.child = layer
            layer.is_sub_layer = True
            layer.root = layer.root
            layer.root.last_sub_layer = layer
            if isinstance(layer, CostLayer):
                layer.root.is_last_root = True
            self.parent = _parent
            self._layers.append(layer)
            if not isinstance(layer, ConvLayer):
                self._weights.append(np.eye(_current))
                self._bias.append(np.zeros((1, _current)))
            else:
                self._weights.append(np.array([.0]))
                self._bias.append(np.array([.0]))
            self._current_dimension = _next
        else:
            fc_shape, conv_channel, last_layer = None, None, self._layers[-1]
            if isinstance(last_layer, ConvLayer):
                if isinstance(layer, ConvLayer):
                    conv_channel = last_layer.n_filters
                    _current = (conv_channel, last_layer.out_h, last_layer.out_w)
                    layer.feed_shape((_current, _next))
                else:
                    layer.is_fc = True
                    last_layer.is_fc_base = True
                    fc_shape = last_layer.out_h * last_layer.out_w * last_layer.n_filters
            self._layers.append(layer)
            self._add_weight((_current, _next), conv_channel, fc_shape)
            self._current_dimension = _next
        self._update_layer_information(layer)

    @NNTiming.timeit(level=4)
    def _add_cost_layer(self):
        _last_layer = self._layers[-1]
        _last_layer_root = _last_layer.root
        if not isinstance(_last_layer, CostLayer):
            if _last_layer_root.name == "Sigmoid":
                self._cost_layer = "Cross Entropy"
            elif _last_layer_root.name == "Softmax":
                self._cost_layer = "Log Likelihood"
            else:
                self._cost_layer = "MSE"
            self.add(self._cost_layer)
        else:
            self._cost_layer = _last_layer.cost_function

    @NNTiming.timeit(level=4)
    def _update_layer_information(self, layer):
        self._layer_params.append(layer.params)

    @NNTiming.timeit(level=1)
    def _get_prediction(self, x, name=None, batch_size=1e6, verbose=None):
        if verbose is None:
            verbose = self.verbose
        single_batch = int(batch_size / np.prod(x.shape[1:]))
        if not single_batch:
            single_batch = 1
        if single_batch >= len(x):
            return self._get_activations(x, predict=True).pop()
        epoch = int(len(x) / single_batch)
        if not len(x) % single_batch:
            epoch += 1
        name = "Prediction" if name is None else "Prediction ({})".format(name)
        sub_bar = ProgressBar(min_value=0, max_value=epoch, name=name)
        if verbose >= NNVerbose.METRICS:
            sub_bar.start()
        rs, count = [self._get_activations(x[:single_batch], predict=True).pop()], single_batch
        if verbose >= NNVerbose.METRICS:
            sub_bar.update()
        while count < len(x):
            count += single_batch
            if count >= len(x):
                rs.append(self._get_activations(x[count-single_batch:], predict=True).pop())
            else:
                rs.append(self._get_activations(x[count-single_batch:count], predict=True).pop())
            if verbose >= NNVerbose.METRICS:
                sub_bar.update()
        return np.vstack(rs)

    @NNTiming.timeit(level=1)
    def _get_activations(self, x, predict=False):
        _activations = [self._layers[0].activate(x, self._weights[0], self._bias[0], predict)]
        for i, layer in enumerate(self._layers[1:]):
            _activations.append(layer.activate(
                _activations[-1], self._weights[i + 1], self._bias[i + 1], predict))
        return _activations

    @NNTiming.timeit(level=3)
    def _append_log(self, x, y, name, get_loss=True):
        y_pred = self._get_prediction(x, name)
        for i, metric in enumerate(self._metrics):
            self._logs[name][i].append(metric(y, y_pred))
        if get_loss:
            self._logs[name][-1].append(self._layers[-1].calculate(y, y_pred) / len(y))

    @NNTiming.timeit(level=3)
    def _print_metric_logs(self, show_loss, data_type):
        print()
        print("=" * 47)
        for i, name in enumerate(self._metric_names):
            print("{:<16s} {:<16s}: {:12.8}".format(
                data_type, name, self._logs[data_type][i][-1]))
        if show_loss:
            print("{:<16s} {:<16s}: {:12.8}".format(
                data_type, "loss", self._logs[data_type][-1][-1]))
        print("=" * 47)

    @NNTiming.timeit(level=1)
    def _draw_network(self, radius=6, width=1200, height=800, padding=0.2, sub_layer_height_scale=0, delay=1,
                      weight_average=None, activations=None):

        layers = len(self._layers) + 1
        units = [layer.shape[0] for layer in self._layers] + [self._layers[-1].shape[1]]
        whether_sub_layers = np.array([False] + [isinstance(layer, SubLayer) for layer in self._layers])
        n_sub_layers = int(np.sum(whether_sub_layers))

        img = np.zeros((height, width, 3), np.uint8)
        axis0_padding = int(height / (layers - 1 + 2 * padding)) * padding
        axis0_step = (height - 2 * axis0_padding) / layers
        sub_layer_decrease = int((1 - sub_layer_height_scale) * axis0_step)
        axis0 = np.linspace(
            axis0_padding,
            height + n_sub_layers * sub_layer_decrease - axis0_padding,
            layers, dtype=np.int)
        axis0 -= sub_layer_decrease * np.cumsum(whether_sub_layers)
        axis1_divide = [int(width / (unit + 1)) for unit in units]
        axis1 = [np.linspace(divide, width - divide, units[i], dtype=np.int)
                 for i, divide in enumerate(axis1_divide)]

        colors, thicknesses = [], []
        color_weights = [weight.copy() for weight in self._weights]
        color_min = [np.min(weight) for weight in color_weights]
        color_max = [np.max(weight) for weight in color_weights]
        color_average = [np.average(weight) for weight in color_weights] if weight_average is None else weight_average
        for weight, weight_min, weight_max, weight_average in zip(
                color_weights, color_min, color_max, color_average
        ):
            line_info = VisUtil.get_line_info(weight, weight_min, weight_max, weight_average)
            colors.append(line_info[0])
            thicknesses.append(line_info[1])

        activations = [np.average(np.abs(activation), axis=0) for activation in activations]
        activations = [activation / np.max(activation) for activation in activations]
        for i, (y, xs) in enumerate(zip(axis0, axis1)):
            for j, x in enumerate(xs):
                if i == 0:
                    cv2.circle(img, (x, y), radius, (20, 215, 20), int(radius / 2))
                else:
                    activation = activations[i - 1][j]
                    try:
                        cv2.circle(img, (x, y), radius, (
                            int(255 * activation), int(255 * activation), int(255 * activation)), int(radius / 2))
                    except ValueError:
                        cv2.circle(img, (x, y), radius, (0, 0, 255), int(radius / 2))
            if i > 0:
                cv2.putText(img, self._layers[i - 1].name, (12, y - 36), cv2.LINE_AA, 0.6, (255, 255, 255), 2)

        for i, y in enumerate(axis0):
            if i == len(axis0) - 1:
                break
            for j, x in enumerate(axis1[i]):
                new_y = axis0[i + 1]
                whether_sub_layer = isinstance(self._layers[i], SubLayer)
                for k, new_x in enumerate(axis1[i + 1]):
                    if whether_sub_layer and j != k:
                        continue
                    cv2.line(img, (x, y), (new_x, new_y), colors[i][j][k], thicknesses[i][j][k])

        cv2.imshow("Neural Network", img)
        cv2.waitKey(delay)
        return img

    @NNTiming.timeit(level=1)
    def _draw_detailed_network(self, radius=6, width=1200, height=800, padding=0.2,
                               plot_scale=2, plot_precision=0.03,
                               sub_layer_height_scale=0, delay=1,
                               weight_average=None):

        layers = len(self._layers) + 1
        units = [layer.shape[0] for layer in self._layers] + [self._layers[-1].shape[1]]
        whether_sub_layers = np.array([False] + [isinstance(layer, SubLayer) for layer in self._layers])
        n_sub_layers = int(np.sum(whether_sub_layers))

        plot_num = int(1 / plot_precision)
        if plot_num % 2 == 1:
            plot_num += 1
        half_plot_num = int(plot_num * 0.5)
        xf = np.linspace(self._x_min * plot_scale, self._x_max * plot_scale, plot_num)
        yf = np.linspace(self._x_min * plot_scale, self._x_max * plot_scale, plot_num) * -1
        input_x, input_y = np.meshgrid(xf, yf)
        input_xs = np.c_[input_x.ravel(), input_y.ravel()]

        _activations = [activation.T.reshape(units[i + 1], plot_num, plot_num)
                        for i, activation in enumerate(self._get_activations(input_xs, predict=True))]
        _graphs = []
        for j, activation in enumerate(_activations):
            _graph_group = []
            for ac in activation:
                data = np.zeros((plot_num, plot_num, 3), np.uint8)
                mask = ac >= np.average(ac)
                data[mask], data[~mask] = [0, 125, 255], [255, 125, 0]
                _graph_group.append(data)
            _graphs.append(_graph_group)

        img = np.ones((height, width, 3), np.uint8) * 255
        axis0_padding = int(height / (layers - 1 + 2 * padding)) * padding + plot_num
        axis0_step = (height - 2 * axis0_padding) / layers
        sub_layer_decrease = int((1 - sub_layer_height_scale) * axis0_step)
        axis0 = np.linspace(
            axis0_padding,
            height + n_sub_layers * sub_layer_decrease - axis0_padding,
            layers, dtype=np.int)
        axis0 -= sub_layer_decrease * np.cumsum(whether_sub_layers)
        axis1_padding = plot_num
        axis1 = [np.linspace(axis1_padding, width - axis1_padding, unit + 2, dtype=np.int)
                 for unit in units]
        axis1 = [axis[1:-1] for axis in axis1]

        colors, thicknesses = [], []
        color_weights = [weight.copy() for weight in self._weights]
        color_min = [np.min(weight) for weight in color_weights]
        color_max = [np.max(weight) for weight in color_weights]
        color_average = [np.average(weight) for weight in color_weights] if weight_average is None else weight_average
        for weight, weight_min, weight_max, weight_average in zip(
                color_weights, color_min, color_max, color_average
        ):
            line_info = VisUtil.get_line_info(weight, weight_min, weight_max, weight_average)
            colors.append(line_info[0])
            thicknesses.append(line_info[1])

        for i, (y, xs) in enumerate(zip(axis0, axis1)):
            for j, x in enumerate(xs):
                if i == 0:
                    cv2.circle(img, (x, y), radius, (20, 215, 20), int(radius / 2))
                else:
                    graph = _graphs[i - 1][j]
                    img[y - half_plot_num:y + half_plot_num, x - half_plot_num:x + half_plot_num] = graph
            if i > 0:
                cv2.putText(img, self._layers[i - 1].name, (12, y - 36), cv2.LINE_AA, 0.6, (0, 0, 0), 1)

        for i, y in enumerate(axis0):
            if i == len(axis0) - 1:
                break
            for j, x in enumerate(axis1[i]):
                new_y = axis0[i + 1]
                whether_sub_layer = isinstance(self._layers[i], SubLayer)
                for k, new_x in enumerate(axis1[i + 1]):
                    if whether_sub_layer and j != k:
                        continue
                    cv2.line(img, (x, y + half_plot_num), (new_x, new_y - half_plot_num),
                             colors[i][j][k], thicknesses[i][j][k])

        cv2.imshow("Neural Network", img)
        cv2.waitKey(delay)
        return img

    @NNTiming.timeit(level=1)
    def _draw_img_network(self, img_shape, width=1200, height=800, padding=0.2,
                          sub_layer_height_scale=0, delay=1,
                          weight_average=None):

        img_width, img_height = img_shape
        half_width = int(img_width * 0.5) if img_width % 2 == 0 else int(img_width * 0.5) + 1
        half_height = int(img_height * 0.5) if img_height % 2 == 0 else int(img_height * 0.5) + 1

        layers = len(self._layers)
        units = [layer.shape[1] for layer in self._layers]
        whether_sub_layers = np.array([isinstance(layer, SubLayer) for layer in self._layers])
        n_sub_layers = int(np.sum(whether_sub_layers))

        _activations = [self._weights[0].copy().T]
        for weight in self._weights[1:]:
            _activations.append(weight.T.dot(_activations[-1]))
        _graphs = []
        for j, activation in enumerate(_activations):
            _graph_group = []
            for ac in activation:
                ac = ac.reshape(img_width, img_height)
                ac -= np.average(ac)
                data = np.zeros((img_width, img_height, 3), np.uint8)
                mask = ac >= 0.25
                data[mask], data[~mask] = [0, 130, 255], [255, 130, 0]
                _graph_group.append(data)
            _graphs.append(_graph_group)

        img = np.zeros((height, width, 3), np.uint8)
        axis0_padding = int(height / (layers - 1 + 2 * padding)) * padding + img_height
        axis0_step = (height - 2 * axis0_padding) / layers
        sub_layer_decrease = int((1 - sub_layer_height_scale) * axis0_step)
        axis0 = np.linspace(
            axis0_padding,
            height + n_sub_layers * sub_layer_decrease - axis0_padding,
            layers, dtype=np.int)
        axis0 -= sub_layer_decrease * np.cumsum(whether_sub_layers)
        axis1_padding = img_width
        axis1 = [np.linspace(axis1_padding, width - axis1_padding, unit + 2, dtype=np.int)
                 for unit in units]
        axis1 = [axis[1:-1] for axis in axis1]

        colors, thicknesses = [], []
        color_weights = [weight.copy() for weight in self._weights]
        color_min = [np.min(weight) for weight in color_weights]
        color_max = [np.max(weight) for weight in color_weights]
        color_average = [np.average(weight) for weight in color_weights] if weight_average is None else weight_average
        for weight, weight_min, weight_max, weight_average in zip(
                color_weights, color_min, color_max, color_average
        ):
            line_info = VisUtil.get_line_info(weight, weight_min, weight_max, weight_average)
            colors.append(line_info[0])
            thicknesses.append(line_info[1])

        for i, (y, xs) in enumerate(zip(axis0, axis1)):
            for j, x in enumerate(xs):
                graph = _graphs[i][j]
                img[y - half_height:y + half_height, x - half_width:x + half_width] = graph
            cv2.putText(img, self._layers[i].name, (12, y - 36), cv2.LINE_AA, 0.6, (255, 255, 255), 2)

        for i, y in enumerate(axis0):
            if i == len(axis0) - 1:
                break
            for j, x in enumerate(axis1[i]):
                new_y = axis0[i + 1]
                whether_sub_layer = isinstance(self._layers[i + 1], SubLayer)
                for k, new_x in enumerate(axis1[i + 1]):
                    if whether_sub_layer and j != k:
                        continue
                    cv2.line(img, (x, y + half_height), (new_x, new_y - half_height),
                             colors[i + 1][j][k], thicknesses[i + 1][j][k])

        cv2.imshow("Neural Network", img)
        cv2.waitKey(delay)
        return img

    # Metrics

    @staticmethod
    @NNTiming.timeit(level=2, prefix="[Private StaticMethod] ")
    def _acc(y, y_pred):
        y_arg, y_pred_arg = np.argmax(y, axis=1), np.argmax(y_pred, axis=1)
        return np.sum(y_arg == y_pred_arg) / len(y_arg)

    @staticmethod
    @NNTiming.timeit(level=2, prefix="[Private StaticMethod] ")
    def _f1_score(y, y_pred):
        y_true, y_pred = np.argmax(y, axis=1), np.argmax(y_pred, axis=1)
        tp = np.sum(y_true * y_pred)
        if tp == 0:
            return .0
        fp = np.sum((1 - y_true) * y_pred)
        fn = np.sum(y_true * (1 - y_pred))
        return 2 * tp / (2 * tp + fn + fp)

    # Optimizing Process

    @NNTiming.timeit(level=4)
    def _init_optimizer(self):
        if not isinstance(self._w_optimizer, Optimizers):
            self._w_optimizer = self._optimizer_factory.get_optimizer_by_name(
                self._w_optimizer, self._weights, self.NNTiming, self._lr, self._epoch)
        if not isinstance(self._b_optimizer, Optimizers):
            self._b_optimizer = self._optimizer_factory.get_optimizer_by_name(
                self._b_optimizer, self._bias, self.NNTiming, self._lr, self._epoch)
        if self._w_optimizer.name != self._b_optimizer.name:
            self._optimizer_name = None
        else:
            self._optimizer_name = self._w_optimizer.name

    @NNTiming.timeit(level=1)
    def _opt(self, i, _activation, _delta):
        if not isinstance(self._layers[i], ConvLayer):
            self._weights[i] *= self._regularization_param
            self._weights[i] += self._w_optimizer.run(
                i, _activation.reshape(_activation.shape[0], -1).T.dot(_delta)
            )
            if self._whether_apply_bias:
                self._bias[i] += self._b_optimizer.run(
                    i, np.sum(_delta, axis=0, keepdims=True)
                )
        else:
            self._weights[i] *= self._regularization_param
            if _delta[1] is not None:
                self._weights[i] += self._w_optimizer.run(i, _delta[1])
            if self._whether_apply_bias and _delta[2] is not None:
                self._bias[i] += self._b_optimizer.run(i, _delta[2])

    # API

    @NNTiming.timeit(level=4, prefix="[API] ")
    def feed(self, x, y):
        self._feed_data(x, y)

    @NNTiming.timeit(level=4, prefix="[API] ")
    def add(self, layer, *args, **kwargs):
        if isinstance(layer, str):
            self._add_layer(layer, *args, **kwargs)
        else:
            if not isinstance(layer, Layer):
                raise BuildLayerError("Invalid Layer provided (should be subclass of Layer)")
            if not self._layers:
                if isinstance(layer, SubLayer):
                    raise BuildLayerError("Invalid Layer provided (first layer should not be subclass of SubLayer)")
                if len(layer.shape) != 2:
                    raise BuildLayerError("Invalid input Layer provided (shape should be {}, {} found)".format(
                        2, len(layer.shape)
                    ))
                self._layers, self._current_dimension = [layer], layer.shape[1]
                self._update_layer_information(layer)
                if isinstance(layer, ConvLayer):
                    self._add_weight(layer.shape, layer.n_channels)
                else:
                    self._add_weight(layer.shape)
            else:
                if len(layer.shape) > 2:
                    raise BuildLayerError("Invalid Layer provided (shape should be {}, {} found)".format(
                        2, len(layer.shape)
                    ))
                if len(layer.shape) == 2:
                    _current, _next = layer.shape
                    if isinstance(layer, SubLayer):
                        if _next != self._current_dimension:
                            raise BuildLayerError("Invalid SubLayer provided (shape[1] should be {}, {} found)".format(
                                self._current_dimension, _next
                            ))
                    elif not isinstance(layer, ConvLayer) and _current != self._current_dimension:
                        raise BuildLayerError("Invalid Layer provided (shape[0] should be {}, {} found)".format(
                            self._current_dimension, _current
                        ))
                    self._add_layer(layer, _current, _next)

                elif len(layer.shape) == 1:
                    _next = layer.shape[0]
                    layer.shape = (self._current_dimension, _next)
                    self._add_layer(layer, self._current_dimension, _next)
                else:
                    raise LayerError("Invalid Layer provided (invalid shape '{}' found)".format(layer.shape))

    @NNTiming.timeit(level=4, prefix="[API] ")
    def build(self, units="build"):
        if isinstance(units, str):
            if units == "build":
                for name, param in zip(self._layer_names, self._layer_params):
                    self._add_layer(name, *param)
                self._add_cost_layer()
            else:
                raise NotImplementedError("Invalid param '{}' provided to 'build' method".format(units))
        else:
            try:
                units = np.array(units).flatten().astype(np.int)
            except ValueError as err:
                raise BuildLayerError(err)
            if len(units) < 2:
                raise BuildLayerError("At least 2 layers are needed")
            _input_shape = (units[0], units[1])
            self.initialize()
            self.add(Sigmoid(_input_shape))
            for unit_num in units[2:]:
                self.add(Sigmoid((unit_num,)))
            self._add_cost_layer()

    @NNTiming.timeit(level=4, prefix="[API] ")
    def preview(self, add_cost=True):
        if not self._layers:
            rs = "None"
        else:
            if add_cost:
                self._add_cost_layer()
            rs = (
                "Input  :  {:<10s} - {}\n".format("Dimension", self._layers[0].shape[0]) +
                "\n".join([
                    "Layer  :  {:<10s} - {} {}".format(
                        _layer.name, _layer.shape[1], _layer.description
                    ) if isinstance(_layer, SubLayer) else
                    "Layer  :  {:<10s} - {:<14s} - strides: {:2d} - padding: {:2d} - out: {}".format(
                        _layer.name, str(_layer.shape[1]), _layer.stride, _layer.padding,
                        (_layer.n_filters, _layer.out_h, _layer.out_w)
                    ) if isinstance(_layer, ConvLayer) else "Layer  :  {:<10s} - {}".format(
                        _layer.name, _layer.shape[1]
                    ) for _layer in self._layers[:-1]
                ]) + "\nCost   :  {:<10s}".format(self._cost_layer)
            )
        print("=" * 30 + "\n" + "Structure\n" + "-" * 30 + "\n" + rs + "\n" + "-" * 30 + "\n")

    @NNTiming.timeit(level=4, prefix="[API] ")
    def split_data(self, x, y, x_test, y_test,
                   train_only, training_scale=NNConfig.TRAINING_SCALE):
        if train_only:
            if x_test is not None and y_test is not None:
                x, y = np.vstack((x, x_test)), np.vstack((y, y_test))
            x_train, y_train = np.array(x), np.array(y)
            x_test, y_test = x_train, y_train
        else:
            shuffle_suffix = np.random.permutation(len(x))
            x, y = x[shuffle_suffix], y[shuffle_suffix]
            if x_test is None or y_test is None:
                train_len = int(len(x) * training_scale)
                x_train, y_train = np.array(x[:train_len]), np.array(y[:train_len])
                x_test, y_test = np.array(x[train_len:]), np.array(y[train_len:])
            elif x_test is None or y_test is None:
                raise BuildNetworkError("Please provide test sets if you want to split data on your own")
            else:
                x_train, y_train = np.array(x), np.array(y)
                x_test, y_test = np.array(x_test), np.array(y_test)
        if NNConfig.BOOST_LESS_SAMPLES:
            if y_train.shape[1] != 2:
                raise BuildNetworkError("It is not permitted to boost less samples in multiple classification")
            y_train_arg = np.argmax(y_train, axis=1)
            y0 = y_train_arg == 0
            y1 = ~y0
            y_len, y0_len = len(y_train), int(np.sum(y0))
            if y0_len > 0.5 * y_len:
                y0, y1 = y1, y0
                y0_len = y_len - y0_len
            boost_suffix = np.random.randint(y0_len, size=y_len - y0_len)
            x_train = np.vstack((x_train[y1], x_train[y0][boost_suffix]))
            y_train = np.vstack((y_train[y1], y_train[y0][boost_suffix]))
            shuffle_suffix = np.random.permutation(len(x_train))
            x_train, y_train = x_train[shuffle_suffix], y_train[shuffle_suffix]
        return (x_train, x_test), (y_train, y_test)

    @NNTiming.timeit(level=1, prefix="[API] ")
    def fit(self,
            x=None, y=None, x_test=None, y_test=None,
            batch_size=512, record_period=1, train_only=True,
            optimizer=None, w_optimizer=None, b_optimizer=None,
            lr=0.01, lb=0.01, epoch=20, weight_scale=1, apply_bias=True,
            show_loss=True, metrics=None, do_log=True, verbose=None,
            visualize=False, visualize_setting=None,
            draw_weights=False, draw_network=False, draw_detailed_network=False,
            draw_img_network=False, img_shape=None, weight_average=None):

        if draw_img_network and img_shape is None:
            raise BuildNetworkError("Please provide image's shape to draw_img_network")

        x, y = self._feed_data(x, y)
        self._lr, self._epoch = lr, epoch
        for weight in self._weights:
            weight *= weight_scale
        if not self._w_optimizer or not self._b_optimizer:
            if not self._optimizer_name:
                if optimizer is None:
                    optimizer = "Adam"
                self._w_optimizer = optimizer if w_optimizer is None else w_optimizer
                self._b_optimizer = optimizer if b_optimizer is None else b_optimizer
            else:
                if not self._w_optimizer:
                    self._w_optimizer = self._optimizer_name
                if not self._b_optimizer:
                    self._b_optimizer = self._optimizer_name
        self._init_optimizer()
        assert isinstance(self._w_optimizer, Optimizers) and isinstance(self._b_optimizer, Optimizers)
        print()
        print("=" * 30)
        print("Optimizers")
        print("-" * 30)
        print("w: {}\nb: {}".format(self._w_optimizer, self._b_optimizer))
        print("-" * 30)

        if not self._layers:
            raise BuildNetworkError("Please provide layers before fitting data")
        self._add_cost_layer()

        if y.shape[1] != self._current_dimension:
            raise BuildNetworkError("Output layer's shape should be {}, {} found".format(
                self._current_dimension, y.shape[1]))

        (x_train, x_test), (y_train, y_test) = self.split_data(
            x, y, x_test, y_test, train_only)
        train_len = len(x_train)
        batch_size = min(batch_size, train_len)
        do_random_batch = train_len >= batch_size
        train_repeat = int(train_len / batch_size) + 1
        self._regularization_param = 1 - lb * lr / batch_size
        self._feed_data(x_train, y_train)

        self._metrics = ["acc"] if metrics is None else metrics
        for i, metric in enumerate(self._metrics):
            if isinstance(metric, str):
                if metric not in self._available_metrics:
                    raise BuildNetworkError("Metric '{}' is not implemented".format(metric))
                self._metrics[i] = self._available_metrics[metric]
        self._metric_names = [_m.__name__ for _m in self._metrics]

        self._logs = {
            name: [[] for _ in range(len(self._metrics) + 1)] for name in ("train", "cv", "test")
        }
        if verbose is not None:
            self.verbose = verbose

        layer_width = len(self._layers)
        self._whether_apply_bias = apply_bias

        bar = ProgressBar(min_value=0, max_value=max(1, epoch // record_period), name="Epoch")
        if self.verbose >= NNVerbose.EPOCH:
            bar.start()
        img = None

        weight_trace = [[[org] for org in weight] for weight in self._weights]
        sub_bar = ProgressBar(min_value=0, max_value=train_repeat * record_period - 1, name="Iteration")
        for counter in range(epoch):
            self._w_optimizer.update(); self._b_optimizer.update()
            _xs, _activations = [], []
            if self.verbose >= NNVerbose.EPOCH and counter % record_period == 0:
                sub_bar.start()

            for _i in range(train_repeat):
                if do_random_batch:
                    batch = np.random.choice(train_len, batch_size)
                    x_batch, y_batch = x_train[batch], y_train[batch]
                else:
                    x_batch, y_batch = x_train, y_train

                _activations = self._get_activations(x_batch)
                if self.verbose >= NNVerbose.DEBUG:
                    _xs = [x_batch.dot(self._weights[0])]
                    for i, weight in enumerate(self._weights[1:]):
                        _xs.append(_activations[i].dot(weight))

                _deltas = [self._layers[-1].bp_first(y_batch, _activations[-1])]
                for i in range(-1, -len(_activations), -1):
                    _deltas.append(self._layers[i - 1].bp(_activations[i - 1], self._weights[i], _deltas[-1]))

                for i in range(layer_width - 1, 0, -1):
                    if not isinstance(self._layers[i], SubLayer):
                        self._opt(i, _activations[i - 1], _deltas[layer_width - i - 1])
                self._opt(0, x_batch, _deltas[-1])

                if draw_weights:
                    for i, weight in enumerate(self._weights):
                        for j, new_weight in enumerate(weight.copy()):
                            weight_trace[i][j].append(new_weight)
                if self.verbose >= NNVerbose.DEBUG:

                    print("")
                    print("## Activations ##")
                    for i, ac in enumerate(_activations):
                        print("-- Layer {} ({}) --".format(i + 1, self._layers[i].name))
                        print(_xs[i])
                        print(ac)

                    print("")
                    print("## Deltas ##")
                    for i, delta in zip(range(len(_deltas) - 1, -1, -1), _deltas):
                        print("-- Layer {} ({}) --".format(i + 1, self._layers[i].name))
                        print(delta)

                    _ = input("Press any key to continue...")
                if self.verbose >= NNVerbose.EPOCH:
                    if sub_bar.update() and self.verbose >= NNVerbose.METRICS_DETAIL:
                        self._append_log(x, y, "train", get_loss=show_loss)
                        self._append_log(x_test, y_test, "cv", get_loss=show_loss)
                        self._print_metric_logs(show_loss, "train")
                        self._print_metric_logs(show_loss, "cv")

            if self.verbose >= NNVerbose.EPOCH:
                sub_bar.update()
            if do_log:
                self._append_log(x, y, "train", get_loss=show_loss)
                self._append_log(x_test, y_test, "cv", get_loss=show_loss)
            if (counter + 1) % record_period == 0:
                if do_log and self.verbose >= NNVerbose.METRICS:
                    self._print_metric_logs(show_loss, "train")
                    self._print_metric_logs(show_loss, "cv")
                if visualize:
                    if visualize_setting is None:
                        self.visualize2d(x_test, y_test)
                    else:
                        self.visualize2d(x_test, y_test, *visualize_setting)
                if x_test.shape[1] == 2:
                    if draw_network:
                        img = self._draw_network(weight_average=weight_average, activations=_activations)
                    if draw_detailed_network:
                        img = self._draw_detailed_network(weight_average=weight_average)
                elif draw_img_network:
                    img = self._draw_img_network(img_shape, weight_average=weight_average)
                if self.verbose >= NNVerbose.EPOCH:
                    bar.update(counter // record_period + 1)
                    sub_bar = ProgressBar(min_value=0, max_value=train_repeat * record_period - 1, name="Iteration")

        if do_log:
            self._append_log(x_test, y_test, "test", get_loss=show_loss)
        if img is not None:
            cv2.waitKey(0)
            cv2.destroyAllWindows()

        if draw_weights:
            ts = np.arange(epoch * train_repeat + 1)
            for i, weight in enumerate(self._weights):
                plt.figure()
                for j in range(len(weight)):
                    plt.plot(ts, weight_trace[i][j])
                plt.title("Weights toward layer {} ({})".format(i + 1, self._layers[i].name))
                plt.show()

        return self._logs

    @NNTiming.timeit(level=2, prefix="[API] ")
    def save(self, path=None, name=None, overwrite=True):

        path = "Models" if path is None else path
        name = "Model.nn" if name is None else name
        if not os.path.exists(path):
            os.mkdir(path)
        slash = "\\" if platform.system() == "Windows" else "/"

        _dir = path + slash + name
        if not overwrite and os.path.isfile(_dir):
            _count = 1
            _new_dir = _dir + "({})".format(_count)
            while os.path.isfile(_new_dir):
                _count += 1
                _new_dir = _dir + "({})".format(_count)
            _dir = _new_dir

        with open(_dir, "wb") as file:
            pickle.dump({
                "structures": {
                    "_layer_names": self.layer_names,
                    "_layer_params": self._layer_params,
                    "_cost_layer": self._layers[-1].name,
                    "_next_dimension": self._current_dimension
                },
                "params": {
                    "_logs": self._logs,
                    "_metric_names": self._metric_names,
                    "_weights": self._weights,
                    "_bias": self._bias,
                    "_optimizer_name": self._optimizer_name,
                    "_w_optimizer": self._w_optimizer,
                    "_b_optimizer": self._b_optimizer,
                    "layer_special_params": self.layer_special_params,
                }
            }, file)

    @NNTiming.timeit(level=2, prefix="[API] ")
    def load(self, path):
        self.initialize()
        try:
            with open(path, "rb") as file:
                _dic = pickle.load(file)
                for key, value in _dic["structures"].items():
                    setattr(self, key, value)
                self.build()
                for key, value in _dic["params"].items():
                    setattr(self, key, value)
                self._init_optimizer()
                for i in range(len(self._metric_names) - 1, -1, -1):
                    name = self._metric_names[i]
                    if name not in self._available_metrics:
                        self._metric_names.pop(i)
                    else:
                        self._metrics.insert(0, self._available_metrics[name])
                return _dic
        except Exception as err:
            raise BuildNetworkError("Failed to load Network ({}), structure initialized".format(err))

    @NNTiming.timeit(level=4, prefix="[API] ")
    def predict(self, x):
        x = np.array(x)
        if len(x.shape) == 1:
            x = x.reshape(1, -1)
        return self._get_prediction(x)

    @NNTiming.timeit(level=4, prefix="[API] ")
    def predict_classes(self, x, flatten=True):
        x = np.array(x)
        if len(x.shape) == 1:
            x = x.reshape(1, -1)
        if flatten:
            return np.argmax(self._get_prediction(x), axis=1)
        return np.argmax([self._get_prediction(x)], axis=2).T

    @NNTiming.timeit(level=4, prefix="[API] ")
    def estimate(self, x, y, metrics=None):
        if metrics is None:
            metrics = self._metrics
        else:
            for i in range(len(metrics) - 1, -1, -1):
                metric = metrics[i]
                if isinstance(metric, str):
                    if metric not in self._available_metrics:
                        metrics.pop(i)
                    else:
                        metrics[i] = self._available_metrics[metric]
        logs, y_pred = [], self._get_prediction(x, verbose=2)
        for metric in metrics:
            logs.append(metric(y, y_pred))
        return logs

    @NNTiming.timeit(level=5, prefix="[API] ")
    def visualize2d(self, x=None, y=None, plot_scale=2, plot_precision=0.01):

        x = self._x if x is None else x
        y = self._y if y is None else y

        plot_num = int(1 / plot_precision)

        xf = np.linspace(self._x_min * plot_scale, self._x_max * plot_scale, plot_num)
        yf = np.linspace(self._x_min * plot_scale, self._x_max * plot_scale, plot_num)
        input_x, input_y = np.meshgrid(xf, yf)
        input_xs = np.c_[input_x.ravel(), input_y.ravel()]

        if self._x.shape[1] != 2:
            return
        output_ys_2d = np.argmax(self.predict(input_xs), axis=1).reshape(len(xf), len(yf))
        output_ys_3d = self.predict(input_xs)[:, 0].reshape(len(xf), len(yf))

        xf, yf = np.meshgrid(xf, yf, sparse=True)

        plt.contourf(input_x, input_y, output_ys_2d, cmap=cm.Spectral)
        plt.scatter(x[:, 0], x[:, 1], c=np.argmax(y, axis=1), s=40, cmap=cm.Spectral)
        plt.axis("off")
        plt.show()

        if self._y.shape[1] == 2:
            fig = plt.figure()
            ax = fig.add_subplot(111, projection='3d')

            ax.plot_surface(xf, yf, output_ys_3d, cmap=cm.coolwarm, )
            ax.set_xlabel("x")
            ax.set_ylabel("y")
            ax.set_zlabel("z")
            plt.show()

    def draw_results(self):
        metrics_log, loss_log = {}, {}
        for key, value in sorted(self._logs.items()):
            metrics_log[key], loss_log[key] = value[:-1], value[-1]

        for i, name in enumerate(sorted(self._metric_names)):
            plt.figure()
            plt.title("Metric Type: {}".format(name))
            for key, log in sorted(metrics_log.items()):
                if key == "test":
                    continue
                xs = np.arange(len(log[i])) + 1
                plt.plot(xs, log[i], label="Data Type: {}".format(key))
            plt.legend(loc=4)
            plt.show()
            plt.close()

        plt.figure()
        plt.title("Loss")
        for key, loss in sorted(loss_log.items()):
            if key == "test":
                continue
            xs = np.arange(len(loss)) + 1
            plt.plot(xs, loss, label="Data Type: {}".format(key))
        plt.legend()
        plt.show()

    def draw_conv_weights(self):
        for i, (name, weight) in enumerate(zip(self.layer_names, self._weights)):
            if len(weight.shape) != 4:
                return
            for j, _w in enumerate(weight):
                for k, _ww in enumerate(_w):
                    VisUtil.show_img(_ww, "{} {} filter {} channel {}".format(name, i+1, j+1, k+1))

    def draw_conv_series(self, x, shape=None):
        for xx in x:
            VisUtil.show_img(VisUtil.trans_img(xx, shape), "Original")
            activations = self._get_activations(np.array([xx]), predict=True)
            for i, (layer, ac) in enumerate(zip(self._layers, activations)):
                if len(ac.shape) == 4:
                    for n in ac:
                        _n, height, width = n.shape
                        a = int(ceil(sqrt(_n)))
                        g = np.ones((a * height + a, a * width + a), n.dtype)
                        g *= np.min(n)
                        _i = 0
                        for y in range(a):
                            for x in range(a):
                                if _i < _n:
                                    g[y * height + y:(y + 1) * height + y, x * width + x:(x + 1) * width + x] = n[
                                        _i, :, :]
                                    _i += 1
                        # normalize to [0,1]
                        max_g = g.max()
                        min_g = g.min()
                        g = (g - min_g) / (max_g - min_g)
                        VisUtil.show_img(g, "Layer {} ({})".format(i + 1, layer.name))
                else:
                    ac = ac[0]
                    length = sqrt(np.prod(ac.shape))
                    if length < 10:
                        continue
                    (height, width) = xx.shape[1:] if shape is None else shape[1:]
                    sqrt_shape = sqrt(height * width)
                    oh, ow = int(length * height / sqrt_shape), int(length * width / sqrt_shape)
                    VisUtil.show_img(ac[:oh*ow].reshape(oh, ow), "Layer {} ({})".format(i + 1, layer.name))

    @staticmethod
    def fuck_pycharm_warning():
        print(Axes3D.acorr)
示例#9
0
class LinearSVM(GDBase):
    LinearSVMTiming = Timing()

    def __init__(self, **kwargs):
        super(LinearSVM, self).__init__(**kwargs)
        self._w = self._b = None

        self._params["c"] = kwargs.get("c", 1)
        self._params["lr"] = kwargs.get("lr", 0.001)
        self._params["batch_size"] = kwargs.get("batch_size", 128)
        self._params["epoch"] = kwargs.get("epoch", 10 ** 4)
        self._params["tol"] = kwargs.get("tol", 1e-3)
        self._params["optimizer"] = kwargs.get("optimizer", "Adam")

    @LinearSVMTiming.timeit(level=1, prefix="[Core] ")
    def _get_grads(self, x_batch, y_batch, y_pred, sample_weight_batch, *args):
        c = args[0]
        err = (1 - y_pred * y_batch) * sample_weight_batch
        mask = err > 0  # type: np.ndarray
        if not np.any(mask):
            self._model_grads = [None, None]
        else:
            delta = -c * y_batch[mask] * sample_weight_batch[mask]
            self._model_grads = [
                np.sum(delta[..., None] * x_batch[mask], axis=0),
                np.sum(delta)
            ]
        return np.sum(err[mask]) + c * np.linalg.norm(self._w)

    @LinearSVMTiming.timeit(level=1, prefix="[API] ")
    def fit(self, x, y, sample_weight=None, c=None, lr=None, optimizer=None,
            batch_size=None, epoch=None, tol=None, animation_params=None):
        if sample_weight is None:
            sample_weight = self._params["sample_weight"]
        if c is None:
            c = self._params["c"]
        if lr is None:
            lr = self._params["lr"]
        if batch_size is None:
            batch_size = self._params["batch_size"]
        if epoch is None:
            epoch = self._params["epoch"]
        if tol is None:
            tol = self._params["tol"]
        if optimizer is None:
            optimizer = self._params["optimizer"]
        *animation_properties, animation_params = self._get_animation_params(animation_params)
        x, y = np.atleast_2d(x), np.asarray(y, dtype=np.float32)
        if sample_weight is None:
            sample_weight = np.ones(len(y))
        else:
            sample_weight = np.asarray(sample_weight) * len(y)

        self._w = np.zeros(x.shape[1], dtype=np.float32)
        self._b = np.zeros(1, dtype=np.float32)
        self._model_parameters = [self._w, self._b]
        self._optimizer = OptFactory().get_optimizer_by_name(
            optimizer, self._model_parameters, lr, epoch
        )

        bar = ProgressBar(max_value=epoch, name="LinearSVM")
        ims = []
        train_repeat = self._get_train_repeat(x, batch_size)
        for i in range(epoch):
            self._optimizer.update()
            l = self._batch_training(
                x, y, batch_size, train_repeat, sample_weight, c
            )
            if l < tol:
                bar.terminate()
                break
            self._handle_animation(i, x, y, ims, animation_params, *animation_properties)
            bar.update()
        self._handle_mp4(ims, animation_properties)

    @LinearSVMTiming.timeit(level=1, prefix="[API] ")
    def predict(self, x, get_raw_results=False, **kwargs):
        rs = np.sum(self._w * x, axis=1) + self._b
        if get_raw_results:
            return rs
        return np.sign(rs)
示例#10
0
文件: Networks.py 项目: F0nnys/ML
class NNDist(ClassifierBase):
    NNTiming = Timing()

    def __init__(self, **kwargs):
        super(NNDist, self).__init__(**kwargs)
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""
        self.verbose = 1

        self._apply_bias = False
        self._current_dimension = 0

        self._logs = {}
        self._metrics, self._metric_names = [], []

        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

        self._layer_factory = LayerFactory()
        self._optimizer_factory = OptFactory()

        self._available_metrics = {
            "acc": self.acc,
            "_acc": self.acc,
            "f1": self.f1_score,
            "_f1_score": self.f1_score
        }

    @NNTiming.timeit(level=4, prefix="[Initialize] ")
    def initialize(self):
        self._layers, self._weights, self._bias = [], [], []
        self._layer_names, self._layer_shapes, self._layer_params = [], [], []
        self._lr, self._epoch, self._regularization_param = 0, 0, 0
        self._w_optimizer, self._b_optimizer, self._optimizer_name = None, None, ""
        self.verbose = 1

        self._apply_bias = False
        self._current_dimension = 0

        self._logs = []
        self._metrics, self._metric_names = [], []

        self._x_min, self._x_max = 0, 0
        self._y_min, self._y_max = 0, 0

    # Property

    @property
    def name(self):
        return ("-".join([str(_layer.shape[1]) for _layer in self._layers]) +
                " at {}".format(
                    time.strftime("%Y-%m-%d %H:%M:%S", time.localtime())))

    @property
    def layer_names(self):
        return [layer.name for layer in self._layers]

    @layer_names.setter
    def layer_names(self, value):
        self._layer_names = value

    @property
    def layer_shapes(self):
        return [layer.shape for layer in self._layers]

    @layer_shapes.setter
    def layer_shapes(self, value):
        self._layer_shapes = value

    @property
    def layer_params(self):
        return self._layer_params

    @layer_params.setter
    def layer_params(self, value):
        self.layer_params = value

    @property
    def layer_special_params(self):
        return [layer.special_params for layer in self._layers]

    @layer_special_params.setter
    def layer_special_params(self, value):
        for layer, sp_param in zip(self._layers, value):
            if sp_param is not None:
                layer.set_special_params(sp_param)

    @property
    def optimizer(self):
        return self._optimizer_name

    @optimizer.setter
    def optimizer(self, value):
        try:
            self._optimizer_name = value
        except KeyError:
            raise BuildNetworkError(
                "Invalid Optimizer '{}' provided".format(value))

    # Utils

    @NNTiming.timeit(level=4)
    def _get_min_max(self, x, y):
        self._x_min, self._x_max = np.min(x), np.max(x)
        self._y_min, self._y_max = np.min(y), np.max(y)

    @NNTiming.timeit(level=4)
    def _split_data(self,
                    x,
                    y,
                    x_test,
                    y_test,
                    train_only,
                    training_scale=NNConfig.TRAINING_SCALE):
        x, y = np.asarray(x, dtype=np.float32), np.asarray(y, dtype=np.float32)
        if x_test is not None and y_test is not None:
            x_test, y_test = np.asarray(x_test, dtype=np.float32), np.asarray(
                y_test, dtype=np.float32)
        if train_only:
            if x_test is not None and y_test is not None:
                x, y = np.vstack((x, x_test)), np.vstack((y, y_test))
            x_train, y_train, x_test, y_test = x, y, x, y
        else:
            shuffle_suffix = np.random.permutation(len(x))
            x, y = x[shuffle_suffix], y[shuffle_suffix]
            if x_test is None or y_test is None:
                train_len = int(len(x) * training_scale)
                x_train, y_train = x[:train_len], y[:train_len]
                x_test, y_test = x[train_len:], y[train_len:]
            elif x_test is None or y_test is None:
                raise BuildNetworkError(
                    "Please provide test sets if you want to split data on your own"
                )
            else:
                x_train, y_train = x, y
        if NNConfig.BOOST_LESS_SAMPLES:
            if y_train.shape[1] != 2:
                raise BuildNetworkError(
                    "It is not permitted to boost less samples in multiple classification"
                )
            y_train_arg = np.argmax(y_train, axis=1)
            y0 = y_train_arg == 0
            y1 = ~y0
            y_len, y0_len = len(y_train), np.sum(y0)  # type: int
            if y0_len > int(0.5 * y_len):
                y0, y1 = y1, y0
                y0_len = y_len - y0_len
            boost_suffix = np.random.randint(y0_len, size=y_len - y0_len)
            x_train = np.vstack((x_train[y1], x_train[y0][boost_suffix]))
            y_train = np.vstack((y_train[y1], y_train[y0][boost_suffix]))
            shuffle_suffix = np.random.permutation(len(x_train))
            x_train, y_train = x_train[shuffle_suffix], y_train[shuffle_suffix]
        return (x_train, x_test), (y_train, y_test)

    @NNTiming.timeit(level=4)
    def _add_params(self, shape, conv_channel=None, fc_shape=None):
        if fc_shape is not None:
            self._weights.append(
                np.random.randn(fc_shape, shape[1]).astype(np.float32))
            self._bias.append(np.zeros((1, shape[1]), dtype=np.float32))
        elif conv_channel is not None:
            if len(shape[1]) <= 2:
                self._weights.append(
                    np.random.randn(conv_channel, conv_channel, shape[1][0],
                                    shape[1][1]).astype(np.float32))
            else:
                self._weights.append(
                    np.random.randn(shape[1][0], conv_channel, shape[1][1],
                                    shape[1][2]).astype(np.float32))
            self._bias.append(np.zeros((1, shape[1][0]), dtype=np.float32))
        else:
            self._weights.append(np.random.randn(*shape).astype(np.float32))
            self._bias.append(np.zeros((1, shape[1]), dtype=np.float32))

    @NNTiming.timeit(level=4)
    def _add_layer(self, layer, *args, **kwargs):
        if not self._layers and isinstance(layer, str):
            layer = self._layer_factory.get_root_layer_by_name(
                layer, *args, **kwargs)
            if layer:
                self.add(layer)
                return
        parent = self._layers[-1]
        if isinstance(parent, CostLayer):
            raise BuildLayerError(
                "Adding layer after CostLayer is not permitted")
        if isinstance(layer, str):
            layer, shape = self._layer_factory.get_layer_by_name(
                layer, parent, self._current_dimension, *args, **kwargs)
            if shape is None:
                self.add(layer)
                return
            _current, _next = shape
        else:
            _current, _next = args
        if isinstance(layer, SubLayer):
            parent.child = layer
            layer.is_sub_layer = True
            layer.root = layer.root
            layer.root.last_sub_layer = layer
            self.parent = parent
            self._layers.append(layer)
            self._weights.append(np.array([.0]))
            self._bias.append(np.array([.0]))
            self._current_dimension = _next
        else:
            fc_shape, conv_channel, last_layer = None, None, self._layers[-1]
            if isinstance(last_layer, ConvLayer):
                if isinstance(layer, ConvLayer):
                    conv_channel = last_layer.n_filters
                    _current = (conv_channel, last_layer.out_h,
                                last_layer.out_w)
                    layer.feed_shape((_current, _next))
                else:
                    layer.is_fc = True
                    last_layer.is_fc_base = True
                    fc_shape = last_layer.out_h * last_layer.out_w * last_layer.n_filters
            self._layers.append(layer)
            self._add_params((_current, _next), conv_channel, fc_shape)
            self._current_dimension = _next
        self._update_layer_information(layer)

    @NNTiming.timeit(level=4)
    def _update_layer_information(self, layer):
        self._layer_params.append(layer.params)
        if len(self._layer_params) > 1 and not layer.is_sub_layer:
            self._layer_params[-1] = ((self._layer_params[-1][0][1], ),
                                      *self._layer_params[-1][1:])

    @NNTiming.timeit(level=1)
    def _get_prediction(self, x, name=None, batch_size=1e6, verbose=None):
        if verbose is None:
            verbose = self.verbose
        fc_shape = np.prod(x.shape[1:])  # type: int
        single_batch = int(batch_size / fc_shape)
        if not single_batch:
            single_batch = 1
        if single_batch >= len(x):
            return self._get_activations(x, predict=True).pop()
        epoch = int(len(x) / single_batch)
        if not len(x) % single_batch:
            epoch += 1
        name = "Prediction" if name is None else "Prediction ({})".format(name)
        sub_bar = ProgressBar(max_value=epoch, name=name, start=False)
        if verbose >= NNVerbose.METRICS:
            sub_bar.start()
        rs, count = [
            self._get_activations(x[:single_batch], predict=True).pop()
        ], single_batch
        if verbose >= NNVerbose.METRICS:
            sub_bar.update()
        while count < len(x):
            count += single_batch
            if count >= len(x):
                rs.append(
                    self._get_activations(x[count - single_batch:],
                                          predict=True).pop())
            else:
                rs.append(
                    self._get_activations(x[count - single_batch:count],
                                          predict=True).pop())
            if verbose >= NNVerbose.METRICS:
                sub_bar.update()
        return np.vstack(rs)

    @NNTiming.timeit(level=1)
    def _get_activations(self, x, predict=False):
        activations = [
            self._layers[0].activate(x, self._weights[0], self._bias[0],
                                     predict)
        ]
        for i, layer in enumerate(self._layers[1:]):
            activations.append(
                layer.activate(activations[-1], self._weights[i + 1],
                               self._bias[i + 1], predict))
        return activations

    @NNTiming.timeit(level=3)
    def _append_log(self, x, y, name, get_loss=True):
        y_pred = self._get_prediction(x, name)
        for i, metric in enumerate(self._metrics):
            self._logs[name][i].append(
                metric(np.argmax(y, axis=1), np.argmax(y_pred, axis=1)))
        if get_loss:
            self._logs[name][-1].append(self._layers[-1].calculate(y, y_pred) /
                                        len(y))

    @NNTiming.timeit(level=3)
    def _print_metric_logs(self, show_loss, data_type):
        print()
        print("=" * 47)
        for i, name in enumerate(self._metric_names):
            print("{:<16s} {:<16s}: {:12.8}".format(
                data_type, name, self._logs[data_type][i][-1]))
        if show_loss:
            print("{:<16s} {:<16s}: {:12.8}".format(
                data_type, "loss", self._logs[data_type][-1][-1]))
        print("=" * 47)

    @NNTiming.timeit(level=1)
    def _draw_2d_network(self,
                         radius=6,
                         width=1200,
                         height=800,
                         padding=0.2,
                         plot_scale=2,
                         plot_precision=0.03,
                         sub_layer_height_scale=0,
                         **kwargs):
        if not kwargs["show"] and not kwargs["mp4"]:
            return
        layers = len(self._layers) + 1
        units = [layer.shape[0]
                 for layer in self._layers] + [self._layers[-1].shape[1]]
        whether_sub_layers = np.array(
            [False] + [isinstance(layer, SubLayer) for layer in self._layers])
        n_sub_layers = np.sum(whether_sub_layers)  # type: int

        plot_num = int(1 / plot_precision)
        if plot_num % 2 == 1:
            plot_num += 1
        half_plot_num = int(plot_num * 0.5)
        xf = np.linspace(self._x_min * plot_scale, self._x_max * plot_scale,
                         plot_num)
        yf = np.linspace(self._x_min * plot_scale, self._x_max * plot_scale,
                         plot_num) * -1
        input_x, input_y = np.meshgrid(xf, yf)
        input_xs = np.c_[input_x.ravel(), input_y.ravel()]

        activations = [
            activation.T.reshape(units[i + 1], plot_num, plot_num)
            for i, activation in enumerate(
                self._get_activations(input_xs, predict=True))
        ]
        graphs = []
        for j, activation in enumerate(activations):
            graph_group = []
            if j == len(activations) - 1:
                classes = np.argmax(activation, axis=0)
            else:
                classes = None
            for k, ac in enumerate(activation):
                data = np.zeros((plot_num, plot_num, 3), np.uint8)
                if j != len(activations) - 1:
                    mask = ac >= np.average(ac)
                else:
                    mask = classes == k
                data[mask], data[~mask] = [0, 165, 255], [255, 165, 0]
                graph_group.append(data)
            graphs.append(graph_group)

        img = np.full([height, width, 3], 255, dtype=np.uint8)
        axis0_padding = int(height /
                            (layers - 1 + 2 * padding)) * padding + plot_num
        axis0_step = (height - 2 * axis0_padding) / layers
        sub_layer_decrease = int((1 - sub_layer_height_scale) * axis0_step)
        axis0 = np.linspace(axis0_padding,
                            height + n_sub_layers * sub_layer_decrease -
                            axis0_padding,
                            layers,
                            dtype=np.int)
        axis0 -= sub_layer_decrease * np.cumsum(whether_sub_layers)
        axis1_padding = plot_num
        axis1 = [
            np.linspace(axis1_padding,
                        width - axis1_padding,
                        unit + 2,
                        dtype=np.int) for unit in units
        ]
        axis1 = [axis[1:-1] for axis in axis1]

        colors, thicknesses, masks = [], [], []
        for weight in self._weights:
            line_info = VisUtil.get_line_info(weight.copy())
            colors.append(line_info[0])
            thicknesses.append(line_info[1])
            masks.append(line_info[2])

        for i, (y, xs) in enumerate(zip(axis0, axis1)):
            for j, x in enumerate(xs):
                if i == 0:
                    cv2.circle(img, (x, y), radius, (20, 215, 20),
                               int(radius / 2))
                else:
                    graph = graphs[i - 1][j]
                    img[y - half_plot_num:y + half_plot_num,
                        x - half_plot_num:x + half_plot_num] = graph
            if i > 0:
                cv2.putText(img, self._layers[i - 1].name, (12, y - 36),
                            cv2.LINE_AA, 0.6, (0, 0, 0), 1)
        for i, y in enumerate(axis0):
            if i == len(axis0) - 1:
                break
            for j, x in enumerate(axis1[i]):
                new_y = axis0[i + 1]
                whether_sub_layer = isinstance(self._layers[i], SubLayer)
                for k, new_x in enumerate(axis1[i + 1]):
                    if whether_sub_layer and j != k:
                        continue
                    if masks[i][j][k]:
                        cv2.line(img, (x, y + half_plot_num),
                                 (new_x, new_y - half_plot_num),
                                 colors[i][j][k], thicknesses[i][j][k])

        return img

    # Optimizing Process

    @NNTiming.timeit(level=4)
    def _init_optimizer(self):
        if not isinstance(self._w_optimizer, Optimizer):
            self._w_optimizer = self._optimizer_factory.get_optimizer_by_name(
                self._w_optimizer, self._weights, self._lr, self._epoch)
        if not isinstance(self._b_optimizer, Optimizer):
            self._b_optimizer = self._optimizer_factory.get_optimizer_by_name(
                self._b_optimizer, self._bias, self._lr, self._epoch)
        if self._w_optimizer.name != self._b_optimizer.name:
            self._optimizer_name = None
        else:
            self._optimizer_name = self._w_optimizer.name

    @NNTiming.timeit(level=1)
    def _opt(self, i, activation, delta):
        if not isinstance(self._layers[i], ConvLayer):
            self._weights[i] *= self._regularization_param
            self._weights[i] += self._w_optimizer.run(
                i,
                activation.reshape(activation.shape[0], -1).T.dot(delta))
            if self._apply_bias:
                self._bias[i] += self._b_optimizer.run(
                    i, np.sum(delta, axis=0, keepdims=True))
        else:
            self._weights[i] *= self._regularization_param
            if delta[1] is not None:
                self._weights[i] += self._w_optimizer.run(i, delta[1])
            if self._apply_bias and delta[2] is not None:
                self._bias[i] += self._b_optimizer.run(i, delta[2])

    # API

    @NNTiming.timeit(level=4, prefix="[API] ")
    def add(self, layer, *args, **kwargs):
        if isinstance(layer, str):
            # noinspection PyTypeChecker
            self._add_layer(layer, *args, **kwargs)
        else:
            if not isinstance(layer, Layer):
                raise BuildLayerError(
                    "Invalid Layer provided (should be subclass of Layer)")
            if not self._layers:
                if isinstance(layer, SubLayer):
                    raise BuildLayerError(
                        "Invalid Layer provided (first layer should not be subclass of SubLayer)"
                    )
                if len(layer.shape) != 2:
                    raise BuildLayerError(
                        "Invalid input Layer provided (shape should be {}, {} found)"
                        .format(2, len(layer.shape)))
                self._layers, self._current_dimension = [layer], layer.shape[1]
                self._update_layer_information(layer)
                if isinstance(layer, ConvLayer):
                    self._add_params(layer.shape, layer.n_channels)
                else:
                    self._add_params(layer.shape)
            else:
                if len(layer.shape) > 2:
                    raise BuildLayerError(
                        "Invalid Layer provided (shape should be {}, {} found)"
                        .format(2, len(layer.shape)))
                if len(layer.shape) == 2:
                    _current, _next = layer.shape
                    if isinstance(layer, SubLayer):
                        if _next != self._current_dimension:
                            raise BuildLayerError(
                                "Invalid SubLayer provided (shape[1] should be {}, {} found)"
                                .format(self._current_dimension, _next))
                    elif not isinstance(
                            layer,
                            ConvLayer) and _current != self._current_dimension:
                        raise BuildLayerError(
                            "Invalid Layer provided (shape[0] should be {}, {} found)"
                            .format(self._current_dimension, _current))
                    self._add_layer(layer, _current, _next)

                elif len(layer.shape) == 1:
                    _next = layer.shape[0]
                    layer.shape = (self._current_dimension, _next)
                    self._add_layer(layer, self._current_dimension, _next)
                else:
                    raise LayerError(
                        "Invalid Layer provided (invalid shape '{}' found)".
                        format(layer.shape))

    @NNTiming.timeit(level=4, prefix="[API] ")
    def build(self, units="build"):
        if isinstance(units, str):
            if units == "build":
                for name, param in zip(self._layer_names, self._layer_params):
                    self.add(name, *param)
            else:
                raise NotImplementedError(
                    "Invalid param '{}' provided to 'build' method".format(
                        units))
        else:
            try:
                units = np.asarray(units).flatten().astype(np.int)
            except ValueError as err:
                raise BuildLayerError(err)
            if len(units) < 2:
                raise BuildLayerError("At least 2 layers are needed")
            _input_shape = (units[0], units[1])
            self.initialize()
            self.add(ReLU(_input_shape))
            for unit_num in units[2:-1]:
                self.add(ReLU((unit_num, )))
            self.add("CrossEntropy", (units[-1], ))

    @NNTiming.timeit(level=4, prefix="[API] ")
    def preview(self):
        if not self._layers:
            rs = "None"
        else:
            rs = ("Input  :  {:<10s} - {}\n".format(
                "Dimension", self._layers[0].shape[0]
            ) + "\n".join([
                "Layer  :  {:<10s} - {} {}".format(layer.name, layer.shape[1],
                                                   layer.description)
                if isinstance(layer, SubLayer)
                else
                "Layer  :  {:<10s} - {:<14s} - strides: {:2d} - padding: {:2d} - out: {}"
                .format(layer.name, str(
                    layer.shape[1]), layer.stride, layer.padding,
                        (layer.n_filters, layer.out_h,
                         layer.out_w)) if isinstance(layer, ConvLayer) else
                "Layer  :  {:<10s} - {}".format(layer.name, layer.shape[1])
                for layer in self._layers[:-1]
            ]) + "\nCost   :  {:<16s}".format(str(self._layers[-1])))
        print("=" * 30 + "\n" + "Structure\n" + "-" * 30 + "\n" + rs + "\n" +
              "-" * 30 + "\n")

    @NNTiming.timeit(level=1, prefix="[API] ")
    def fit(self,
            x=None,
            y=None,
            x_test=None,
            y_test=None,
            batch_size=128,
            record_period=1,
            train_only=False,
            optimizer=None,
            w_optimizer=None,
            b_optimizer=None,
            lr=0.001,
            lb=0.001,
            epoch=20,
            weight_scale=1,
            apply_bias=True,
            show_loss=True,
            metrics=None,
            do_log=True,
            verbose=None,
            visualize=False,
            visualize_setting=None,
            draw_weights=False,
            animation_params=None):
        self._lr, self._epoch = lr, epoch
        for weight in self._weights:
            weight *= weight_scale
        if not self._w_optimizer or not self._b_optimizer:
            if not self._optimizer_name:
                if optimizer is None:
                    optimizer = "Adam"
                self._w_optimizer = optimizer if w_optimizer is None else w_optimizer
                self._b_optimizer = optimizer if b_optimizer is None else b_optimizer
            else:
                if not self._w_optimizer:
                    self._w_optimizer = self._optimizer_name
                if not self._b_optimizer:
                    self._b_optimizer = self._optimizer_name
        self._init_optimizer()
        assert isinstance(self._w_optimizer, Optimizer) and isinstance(
            self._b_optimizer, Optimizer)
        print()
        print("=" * 30)
        print("Optimizers")
        print("-" * 30)
        print("w: {}\nb: {}".format(self._w_optimizer, self._b_optimizer))
        print("-" * 30)
        if not self._layers:
            raise BuildNetworkError(
                "Please provide layers before fitting data")
        if y.shape[1] != self._current_dimension:
            raise BuildNetworkError(
                "Output layer's shape should be {}, {} found".format(
                    self._current_dimension, y.shape[1]))

        (x_train, x_test), (y_train,
                            y_test) = self._split_data(x, y, x_test, y_test,
                                                       train_only)
        train_len = len(x_train)
        batch_size = min(batch_size, train_len)
        do_random_batch = train_len > batch_size
        train_repeat = 1 if not do_random_batch else int(train_len /
                                                         batch_size) + 1
        self._regularization_param = 1 - lb * lr / batch_size
        self._get_min_max(x_train, y_train)

        self._metrics = ["acc"] if metrics is None else metrics
        for i, metric in enumerate(self._metrics):
            if isinstance(metric, str):
                if metric not in self._available_metrics:
                    raise BuildNetworkError(
                        "Metric '{}' is not implemented".format(metric))
                self._metrics[i] = self._available_metrics[metric]
        self._metric_names = [_m.__name__ for _m in self._metrics]

        self._logs = {
            name: [[] for _ in range(len(self._metrics) + 1)]
            for name in ("train", "cv", "test")
        }
        if verbose is not None:
            self.verbose = verbose

        layer_width = len(self._layers)
        self._apply_bias = apply_bias

        bar = ProgressBar(max_value=max(1, epoch // record_period),
                          name="Epoch",
                          start=False)
        if self.verbose >= NNVerbose.EPOCH:
            bar.start()
        img, ims = None, []

        if draw_weights:
            weight_trace = [[[org] for org in weight]
                            for weight in self._weights]
        else:
            weight_trace = []

        *animation_properties, animation_params = self._get_animation_params(
            animation_params)
        sub_bar = ProgressBar(max_value=train_repeat * record_period - 1,
                              name="Iteration",
                              start=False)
        for counter in range(epoch):
            self._w_optimizer.update()
            self._b_optimizer.update()
            if self.verbose >= NNVerbose.ITER and counter % record_period == 0:
                sub_bar.start()
            for _ in range(train_repeat):
                if do_random_batch:
                    batch = np.random.choice(train_len, batch_size)
                    x_batch, y_batch = x_train[batch], y_train[batch]
                else:
                    x_batch, y_batch = x_train, y_train
                activations = self._get_activations(x_batch)

                deltas = [self._layers[-1].bp_first(y_batch, activations[-1])]
                for i in range(-1, -len(activations), -1):
                    deltas.append(self._layers[i - 1].bp(
                        activations[i - 1], self._weights[i], deltas[-1]))

                for i in range(layer_width - 1, 0, -1):
                    if not isinstance(self._layers[i], SubLayer):
                        self._opt(i, activations[i - 1],
                                  deltas[layer_width - i - 1])
                self._opt(0, x_batch, deltas[-1])

                if draw_weights:
                    for i, weight in enumerate(self._weights):
                        for j, new_weight in enumerate(weight.copy()):
                            weight_trace[i][j].append(new_weight)
                if self.verbose >= NNVerbose.DEBUG:
                    pass
                if self.verbose >= NNVerbose.ITER:
                    if sub_bar.update(
                    ) and self.verbose >= NNVerbose.METRICS_DETAIL:
                        self._append_log(x, y, "train", get_loss=show_loss)
                        self._append_log(x_test,
                                         y_test,
                                         "cv",
                                         get_loss=show_loss)
                        self._print_metric_logs(show_loss, "train")
                        self._print_metric_logs(show_loss, "cv")
            if self.verbose >= NNVerbose.ITER:
                sub_bar.update()
            self._handle_animation(
                counter,
                x,
                y,
                ims,
                animation_params,
                *animation_properties,
                img=self._draw_2d_network(**animation_params),
                name="Neural Network")
            if do_log:
                self._append_log(x, y, "train", get_loss=show_loss)
                self._append_log(x_test, y_test, "cv", get_loss=show_loss)
            if (counter + 1) % record_period == 0:
                if do_log and self.verbose >= NNVerbose.METRICS:
                    self._print_metric_logs(show_loss, "train")
                    self._print_metric_logs(show_loss, "cv")
                if visualize:
                    if visualize_setting is None:
                        self.visualize2d(x_test, y_test)
                    else:
                        self.visualize2d(x_test, y_test, *visualize_setting)
                if self.verbose >= NNVerbose.EPOCH:
                    bar.update(counter // record_period + 1)
                    if self.verbose >= NNVerbose.ITER:
                        sub_bar = ProgressBar(
                            max_value=train_repeat * record_period - 1,
                            name="Iteration",
                            start=False)

        if do_log:
            self._append_log(x_test, y_test, "test", get_loss=show_loss)
        if img is not None:
            cv2.waitKey(0)
            cv2.destroyAllWindows()
        if draw_weights:
            ts = np.arange(epoch * train_repeat + 1)
            for i, weight in enumerate(self._weights):
                plt.figure()
                for j in range(len(weight)):
                    plt.plot(ts, weight_trace[i][j])
                plt.title("Weights toward layer {} ({})".format(
                    i + 1, self._layers[i].name))
                plt.show()
        self._handle_mp4(ims, animation_properties, "NN")
        return self._logs

    @NNTiming.timeit(level=2, prefix="[API] ")
    def save(self, path=None, name=None, overwrite=True):
        path = os.path.join("Models",
                            "Cache") if path is None else os.path.join(
                                "Models", path)
        name = "Model.nn" if name is None else name
        if not os.path.exists(path):
            os.makedirs(path)
        _dir = os.path.join(path, name)
        if not overwrite and os.path.isfile(_dir):
            _count = 1
            _new_dir = _dir + "({})".format(_count)
            while os.path.isfile(_new_dir):
                _count += 1
                _new_dir = _dir + "({})".format(_count)
            _dir = _new_dir
        print()
        print("=" * 60)
        print("Saving Model to {}...".format(_dir))
        print("-" * 60)
        with open(_dir, "wb") as file:
            pickle.dump(
                {
                    "structures": {
                        "_layer_names": self.layer_names,
                        "_layer_params": self._layer_params,
                        "_cost_layer": self._layers[-1].name,
                        "_next_dimension": self._current_dimension
                    },
                    "params": {
                        "_logs": self._logs,
                        "_metric_names": self._metric_names,
                        "_weights": self._weights,
                        "_bias": self._bias,
                        "_optimizer_name": self._optimizer_name,
                        "_w_optimizer": self._w_optimizer,
                        "_b_optimizer": self._b_optimizer,
                        "layer_special_params": self.layer_special_params,
                    }
                }, file)
        print("Done")
        print("=" * 60)

    @NNTiming.timeit(level=2, prefix="[API] ")
    def load(self, path=os.path.join("Models", "Cache", "Model.nn")):
        self.initialize()
        try:
            with open(path, "rb") as file:
                dic = pickle.load(file)
                for key, value in dic["structures"].items():
                    setattr(self, key, value)
                self.build()
                for key, value in dic["params"].items():
                    setattr(self, key, value)
                self._init_optimizer()
                for i in range(len(self._metric_names) - 1, -1, -1):
                    name = self._metric_names[i]
                    if name not in self._available_metrics:
                        self._metric_names.pop(i)
                    else:
                        self._metrics.insert(0, self._available_metrics[name])
                print()
                print("=" * 30)
                print("Model restored")
                print("=" * 30)
                return dic
        except Exception as err:
            raise BuildNetworkError(
                "Failed to load Network ({}), structure initialized".format(
                    err))

    @NNTiming.timeit(level=4, prefix="[API] ")
    def predict(self, x, get_raw_results=False, **kwargs):
        x = np.asarray(x)
        if len(x.shape) == 1:
            x = x.reshape(1, -1)
        y_pred = self._get_prediction(x)
        return y_pred if get_raw_results else np.argmax(y_pred, axis=1)

    def draw_results(self):
        metrics_log, loss_log = {}, {}
        for key, value in sorted(self._logs.items()):
            metrics_log[key], loss_log[key] = value[:-1], value[-1]

        for i, name in enumerate(sorted(self._metric_names)):
            plt.figure()
            plt.title("Metric Type: {}".format(name))
            for key, log in sorted(metrics_log.items()):
                if key == "test":
                    continue
                xs = np.arange(len(log[i])) + 1
                plt.plot(xs, log[i], label="Data Type: {}".format(key))
            plt.legend(loc=4)
            plt.show()
            plt.close()

        plt.figure()
        plt.title("Cost")
        for key, loss in sorted(loss_log.items()):
            if key == "test":
                continue
            xs = np.arange(len(loss)) + 1
            plt.plot(xs, loss, label="Data Type: {}".format(key))
        plt.legend()
        plt.show()

    def draw_conv_weights(self):
        for i, (name, weight) in enumerate(zip(self.layer_names,
                                               self._weights)):
            if len(weight.shape) != 4:
                return
            for j, _w in enumerate(weight):
                for k, _ww in enumerate(_w):
                    VisUtil.show_img(
                        _ww, "{} {} filter {} channel {}".format(
                            name, i + 1, j + 1, k + 1))

    def draw_conv_series(self, x, shape=None):
        for xx in x:
            VisUtil.show_img(VisUtil.trans_img(xx, shape), "Original")
            activations = self._get_activations(np.array([xx]), predict=True)
            for i, (layer, ac) in enumerate(zip(self._layers, activations)):
                if len(ac.shape) == 4:
                    for n in ac:
                        _n, height, width = n.shape
                        a = int(ceil(sqrt(_n)))
                        g = np.ones((a * height + a, a * width + a), n.dtype)
                        g *= np.min(n)
                        _i = 0
                        for y in range(a):
                            for x in range(a):
                                if _i < _n:
                                    g[y * height + y:(y + 1) * height + y,
                                      x * width + x:(x + 1) * width +
                                      x] = n[_i, :, :]
                                    _i += 1
                        # normalize to [0,1]
                        max_g = g.max()
                        min_g = g.min()
                        g = (g - min_g) / (max_g - min_g)
                        VisUtil.show_img(
                            g, "Layer {} ({})".format(i + 1, layer.name))
                else:
                    ac = ac[0]
                    length = sqrt(np.prod(ac.shape))
                    if length < 10:
                        continue
                    (height,
                     width) = xx.shape[1:] if shape is None else shape[1:]
                    sqrt_shape = sqrt(height * width)
                    oh, ow = int(length * height / sqrt_shape), int(
                        length * width / sqrt_shape)
                    VisUtil.show_img(ac[:oh * ow].reshape(oh, ow),
                                     "Layer {} ({})".format(i + 1, layer.name))
示例#11
0
class LinearSVM(GDBase):
    LinearSVMTiming = Timing()

    def __init__(self, **kwargs):
        super(LinearSVM, self).__init__(**kwargs)
        self._w = self._b = None

        self._params["c"] = kwargs.get("c", 1)
        self._params["lr"] = kwargs.get("lr", 0.001)
        self._params["batch_size"] = kwargs.get("batch_size", 128)
        self._params["epoch"] = kwargs.get("epoch", 10 ** 4)
        self._params["tol"] = kwargs.get("tol", 1e-3)
        self._params["optimizer"] = kwargs.get("optimizer", "Adam")

    @LinearSVMTiming.timeit(level=1, prefix="[Core] ")
    def _get_grads(self, x_batch, y_batch, y_pred, sample_weight_batch, *args):
        c = args[0]
        err = (1 - y_pred * y_batch) * sample_weight_batch
        mask = err > 0  # type: np.ndarray
        if not np.any(mask):
            self._model_grads = [None, None]
        else:
            delta = -c * y_batch[mask] * sample_weight_batch[mask]
            self._model_grads = [
                np.sum(delta[..., None] * x_batch[mask], axis=0),
                np.sum(delta)
            ]
        return np.sum(err[mask]) + c * np.linalg.norm(self._w)

    @LinearSVMTiming.timeit(level=1, prefix="[API] ")
    def fit(self, x, y, sample_weight=None, c=None, lr=None, optimizer=None,
            batch_size=None, epoch=None, tol=None, animation_params=None):
        if sample_weight is None:
            sample_weight = self._params["sample_weight"]
        if c is None:
            c = self._params["c"]
        if lr is None:
            lr = self._params["lr"]
        if batch_size is None:
            batch_size = self._params["batch_size"]
        if epoch is None:
            epoch = self._params["epoch"]
        if tol is None:
            tol = self._params["tol"]
        if optimizer is None:
            optimizer = self._params["optimizer"]
        *animation_properties, animation_params = self._get_animation_params(animation_params)
        x, y = np.atleast_2d(x), np.asarray(y, dtype=np.float32)
        if sample_weight is None:
            sample_weight = np.ones(len(y))
        else:
            sample_weight = np.asarray(sample_weight) * len(y)

        self._w = np.zeros(x.shape[1], dtype=np.float32)
        self._b = np.zeros(1, dtype=np.float32)
        self._model_parameters = [self._w, self._b]
        self._optimizer = OptFactory().get_optimizer_by_name(
            optimizer, self._model_parameters, lr, epoch
        )

        bar = ProgressBar(max_value=epoch, name="LinearSVM")
        ims = []
        train_repeat = self._get_train_repeat(x, batch_size)
        for i in range(epoch):
            self._optimizer.update()
            l = self._batch_training(
                x, y, batch_size, train_repeat, sample_weight, c
            )
            if l < tol:
                bar.terminate()
                break
            self._handle_animation(i, x, y, ims, animation_params, *animation_properties)
            bar.update()
        self._handle_mp4(ims, animation_properties)

    @LinearSVMTiming.timeit(level=1, prefix="[API] ")
    def predict(self, x, get_raw_results=False, **kwargs):
        rs = np.sum(self._w * x, axis=1) + self._b
        if get_raw_results:
            return rs
        return np.sign(rs)