def ADMM(self, xlf, mu):
        model_xf = xlf
        self.l_f = np.zeros_like(model_xf)
        self.g_f = np.zeros_like(self.l_f)
        self.h_f = np.zeros_like(self.l_f)
        self.gamma = self.init_penalty_factor
        self.gamma_max = self.max_penalty_factor
        self.gamma_scale_step = self.penalty_scale_step
        T = self.feature_map_sz[0] * self.feature_map_sz[1]
        S_xx = np.sum(np.conj(model_xf) * model_xf, axis=2)
        Sg_pre_f = np.sum(np.conj(model_xf) * self.g_pre, axis=2)
        Sgx_pre_f = model_xf * Sg_pre_f[:, :, None]
        iter = 1
        while iter <= self.admm_max_iterations:
            #subproblem g
            B = S_xx + T * (self.gamma + mu)
            Slx_f = np.sum(np.conj(model_xf) * self.l_f, axis=2)
            Shx_f = np.sum(np.conj(model_xf) * self.h_f, axis=2)

            tmp0 = (1 / (T * (self.gamma + mu)) * (self.yf[:, :, None] * model_xf)) - ((1 / (self.gamma + mu)) * self.l_f) + (
                    self.gamma / (self.gamma + mu)) * self.h_f + \
                   (mu / (self.gamma + mu)) * self.g_pre
            tmp1 = 1 / (T *
                        (self.gamma + mu)) * (model_xf *
                                              ((S_xx * self.yf)[:, :, None]))
            tmp2 = mu / (self.gamma + mu) * Sgx_pre_f
            tmp3 = 1 / (self.gamma + mu) * (model_xf * (Slx_f[:, :, None]))
            tmp4 = self.gamma / (self.gamma + mu) * (model_xf *
                                                     Shx_f[:, :, None])
            self.g_f = (tmp0 -
                        (tmp1 + tmp2 - tmp3 + tmp4) / B[:, :, None]).astype(
                            np.complex64)
            #subproblem h
            self.h_f = fft2(
                self.argmin_g(self.reg_window, self.gamma, T,
                              ifft2(self.gamma * (self.g_f + self.l_f))))
            #subproblem mu
            if self.frame_id > 2 and iter < self.admm_max_iterations:
                for i in range(self.g_f.shape[2]):
                    z = np.power(
                        np.linalg.norm(
                            (self.g_f[:, :, i] - self.g_pre[:, :, i]), 2),
                        2) / (2 * self.epsilon)
                    mu = self.ref_mu - z
            # update l
            self.l_f = self.l_f + (self.gamma * (self.g_f - self.h_f))
            # update gama
            self.gamma = min(self.gamma_scale_step * self.gamma,
                             self.gamma_max)
            iter += 1
        self.g_pre = self.g_f
Exemplo n.º 2
0
 def ADMM(self, xf):
     g_f = np.zeros_like(xf)
     h_f = np.zeros_like(g_f)
     l_f = np.zeros_like(g_f)
     mu = 1
     beta = 10
     mumax = 10000
     i = 1
     T = self.feature_map_sz[0] * self.feature_map_sz[1]
     S_xx = np.sum(np.conj(xf) * xf, 2)
     while i <= self.admm_iterations:
         A = mu / (self.admm_gamma + 1)
         B = S_xx + T * A
         S_lx = np.sum(np.conj(xf) * l_f, axis=2)
         S_hx = np.sum(np.conj(xf) * h_f, axis=2)
         # tmp0 = (1 / (T * A) * (self.yf[:, :, None] * xf)) + (self.admm_gamma / A) * (self.M_prev[:, :, None] * xf) - ((1 / A) * l_f) + (mu/A)*h_f
         # tmp1 = 1 / (T * A) * (xf * ((S_xx * self.yf)[:, :, None])) + (self.admm_gamma/A) * (xf * ((S_xx[:,:,np.newaxis] * self.M_prev[:, :, None])))
         tmp0 = ((1 / (T * A)) * (self.yf[:, :, None] * xf)) + (
             self.admm_gamma / A) * (self.M_prev[:, :, None] * xf) - (
                 (1 / A) * l_f) + (mu / A) * h_f
         tmp1 = 1 / (T * A) * (xf * ((S_xx * self.yf)[:, :, None])) + (
             self.admm_gamma / A) * (xf *
                                     ((S_xx * self.M_prev)[:, :, None]))
         tmp2 = (1 / A) * (xf * (S_lx[:, :, None]))
         tmp3 = (mu / A) * xf * S_hx[:, :, None]
         # solve for g
         g_f = (1 / (1 + self.admm_gamma)) * (
             tmp0 - (tmp1 - tmp2 + tmp3) / B[:, :, None])
         # solve for h
         h = (T / ((mu * T) + self.admm_lambda)) * ifft2(mu * g_f + l_f)
         xs, ys, h = self.get_subwindow_no_window(
             h, (np.floor(self.feature_map_sz[0] / 2),
                 np.floor(self.feature_map_sz[1] / 2)),
             self.small_filter_sz)
         t = np.zeros(
             (self.feature_map_sz[1], self.feature_map_sz[0], h.shape[2]),
             dtype=np.complex64)
         # if len(np.shape(h)) == 4:
         #     h = np.sum(h,axis=3)
         t[ys, xs, :] = h
         # t[ys,:][:,xs] = h
         h_f = fft2(t)
         l_f = l_f + (mu * (g_f - h_f))
         mu = min(beta * mu, mumax)
         i += 1
     return g_f
    def update(self, current_frame, frame_id, vis=False):
        self.frame_id = frame_id
        assert len(current_frame.shape) == 3 and current_frame.shape[2] == 3
        #        old_pos=(np.inf,np.inf)
        #        iter=1
        #        while iter<=self.refinement_iterations and (np.abs(old_pos[0]-self._center[0])>1e-2 or
        #                                                    np.abs(old_pos[1]-self._center[1])>1e-2):
        #
        #            sample_scales=self.sc*self.scale_factors

        #position search
        sample_pos = (int(np.round(self._center[0])),
                      int(np.round(self._center[1])))
        sub_window = self.get_sub_window(
            current_frame,
            sample_pos,
            model_sz=self.crop_size,
            scaled_sz=(int(round(self.crop_size[0] * self.sc)),
                       int(round(self.crop_size[1] * self.sc))))
        sub_window = sub_window[:, :, [2, 1, 0]]
        #        xt_hc =self.extrac_feature_test(sub_window, (50,50))
        xt_hc = self.extrac_hc_feature(sub_window, self.cell_size)
        #            for scale in sample_scales:
        #                sub_window = self.get_sub_window(current_frame, sample_pos, model_sz=self.crop_size,
        #                                                 scaled_sz=(int(round(self.crop_size[0] * scale)),
        #                                                            int(round(self.crop_size[1] * scale))))
        #                hc_features=self.extrac_hc_feature(sub_window, self.cell_size)[:,:,:,np.newaxis]
        #                if xt_hc is None:
        #                    xt_hc = hc_features
        #                else:
        #                    xt_hc = np.concatenate((xt_hc, hc_features), axis=3)
        xtw_hc = xt_hc * self.cosine_window[:, :, None]
        xtf_hc = fft2(xtw_hc)
        responsef_hc = np.sum(np.conj(self.g_f)[:, :, :] * xtf_hc, axis=2)
        responsef_padded = resize_dft2(responsef_hc, self.interp_sz)
        response = np.real(ifft2(responsef_padded))
        disp_row, disp_col, sind = resp_newton(response, responsef_padded,
                                               self.newton_iterations, self.ky,
                                               self.kx, self.feature_map_sz)
        if frame_id > 2:
            response_shift = circShift(
                response, [-int(np.floor(disp_row)), -int(np.floor(disp_col))])
            response_pre_shift = circShift(self.response_pre, [
                -int(np.floor(self.disp_row_pre)),
                -int(np.floor(self.disp_col_pre))
            ])
            response_diff = np.abs(
                np.abs(response_shift - response_pre_shift) /
                response_pre_shift)
            self.ref_mu, self.occ = self.updateRefmu(response_diff, self.zeta,
                                                     self.nu, frame_id)
            response_diff = circShift(response_diff, [
                int(np.floor(response_diff.shape[0] / 2)),
                int(np.floor(response_diff.shape[1] / 2))
            ])
            varience = self.delta * np.log(
                response_diff[self.range_h[0]:(self.range_h[-1] + 1),
                              self.range_w[0]:(self.range_w[-1] + 1)] + 1)
            self.reg_window[self.range_h[0]:(self.range_h[-1] + 1),
                            self.range_w[0]:(self.range_w[-1] + 1)] = varience
        self.response_pre = response
        self.disp_row_pre = disp_row
        self.disp_col_pre = disp_col
        #row, col, sind = np.unravel_index(np.argmax(response, axis=None), response.shape)

        #disp_row = (row+ int(np.floor(self.feature_map_sz[1] - 1) / 2)) % self.feature_map_sz[1] - int(
        #    np.floor((self.feature_map_sz[1] - 1) / 2))
        #disp_col = (col + int(np.floor(self.feature_map_sz[0] - 1) / 2)) % self.feature_map_sz[0] - int(
        #    np.floor((self.feature_map_sz[0] - 1) / 2))

        #            if vis is True:
        #                self.score = response[:, :, sind].astype(np.float32)
        #                self.score = np.roll(self.score, int(np.floor(self.score.shape[0] / 2)), axis=0)
        #                self.score = np.roll(self.score, int(np.floor(self.score.shape[1] / 2)), axis=1)
        dx, dy = (disp_col * self.cell_size *
                  self.sc), (disp_row * self.cell_size * self.sc)
        trans_vec = np.array(np.round([dx, dy]))
        #        old_pos = self._center
        self._center = (np.round(sample_pos[0] + dx),
                        np.round(sample_pos[1] + dy))
        #        print(self._center)
        #        self.sc=self.sc*scale_change_factor
        #        self.sc = np.clip(self.sc, self._min_scale_factor, self._max_scale_factor)

        #scale search and update
        self.sc = self.scale_estimator.update(current_frame, self._center,
                                              self.base_target_sz, self.sc)
        #        print(self.sc)
        if self.scale_type == 'normal':
            self.sc = np.clip(self.sc,
                              a_min=self._min_scale_factor,
                              a_max=self._max_scale_factor)

        #training
        shift_sample_pos = np.array(
            2.0 * np.pi * trans_vec /
            (self.sc * np.array(self.crop_size, dtype=float)))
        #        shift_sample() hard to ensure objectiveness....
        #        consider using original ways-----LBW
        # xlf_hc=self.shift_sample(xtf_hc, shift_sample_pos, self.kx, self.ky.T)
        patch = self.get_sub_window(
            current_frame,
            self._center,
            model_sz=self.crop_size,
            scaled_sz=(int(np.round(self.crop_size[0] * self.sc)),
                       int(np.round(self.crop_size[1] * self.sc))))
        patch = patch[:, :, [2, 1, 0]]
        xl_hc = self.extrac_feature_test(patch, (50, 50))
        xl_hc = self.extrac_hc_feature(patch, self.cell_size)
        xlw_hc = xl_hc * self.cosine_window[:, :, None]
        xlf_hc = fft2(xlw_hc)
        mu = self.zeta
        if self.occ == False:
            self.ADMM(xlf_hc, mu)
        target_sz = (self.base_target_sz[0] * self.sc,
                     self.base_target_sz[1] * self.sc)
        return [(self._center[0] - (target_sz[0]) / 2),
                (self._center[1] - (target_sz[1]) / 2), target_sz[0],
                target_sz[1]]
Exemplo n.º 4
0
    def init(self, image, box):

        self.cell_size = self.params.cell_size
        self.cell_selection_thresh = self.params.cell_selection_thresh
        self.search_area_shape = self.params.search_area_shape
        self.search_area_scale = self.params.search_area_scale
        self.filter_max_area = self.params.filter_max_area
        self.interp_factor = self.params.interp_factor
        self.output_sigma_factor = self.params.output_sigma_factor
        self.interpolate_response = self.params.interpolate_response
        self.newton_iterations = self.params.newton_iterations
        self.number_of_scales = self.params.number_of_scales
        self.scale_step = self.params.scale_step
        self.admm_iterations = self.params.admm_iterations
        self.admm_lambda = self.params.admm_lambda
        self.admm_gamma = self.params.admm_gamma
        self.frame = 1
        # self.learning_rate_scale = self.params.learning_rate_scale
        # self.scale_sz_window = self.params.scale_sz_window

        # class ScaleConfig:
        # learning_rate_scale = self.learning_rate_scale
        # scale_sz_window = self.scale_sz_window

        # self.scale_config = ScaleConfig()

        # Get target position and size
        # state = info['init_bbox']
        state = box
        bbox = np.array(state).astype(np.int64)
        x, y, w, h = tuple(bbox)
        self._center = (x + w / 2, y + h / 2)
        # self._center = tuple(np.floor(self._center))
        self.w, self.h = w, h
        self.feature_ratio = self.cell_size
        self.search_area=(self.w/self.feature_ratio*self.search_area_scale)*\
                         (self.h/self.feature_ratio*self.search_area_scale)
        if self.search_area < self.cell_selection_thresh * self.filter_max_area:
            self.cell_size=int(min(self.feature_ratio,max(1,int(np.ceil(np.sqrt(
                self.w*self.search_area_scale/(self.cell_selection_thresh*self.filter_max_area)*\
                self.h*self.search_area_scale))))))
            self.feature_ratio = self.cell_size
            self.search_area = (self.w / self.feature_ratio * self.search_area_scale) * \
                               (self.h / self.feature_ratio * self.search_area_scale)

        if self.search_area > self.filter_max_area:
            self.current_scale_factor = np.sqrt(self.search_area /
                                                self.filter_max_area)
        else:
            self.current_scale_factor = 1.

        self.base_target_sz = (self.w / self.current_scale_factor,
                               self.h / self.current_scale_factor)
        # self.target_sz=self.base_target_sz
        if self.search_area_shape == 'proportional':
            self.crop_size = (int(self.base_target_sz[0] *
                                  self.search_area_scale),
                              int(self.base_target_sz[1] *
                                  self.search_area_scale))
        elif self.search_area_shape == 'square':
            w = np.sqrt(self.base_target_sz[0] *
                        self.base_target_sz[1]) * self.search_area_scale
            self.crop_size = (w, w)
        elif self.search_area_shape == 'fix_padding':
            tmp=int(np.sqrt(self.base_target_sz[0]*self.search_area_scale+(self.base_target_sz[1]-self.base_target_sz[0])/4))+\
                (self.base_target_sz[0]+self.base_target_sz[1])/2
            self.crop_size = (self.base_target_sz[0] + tmp,
                              self.base_target_sz[1] + tmp)
        else:
            raise ValueError
        self.crop_size = (int(
            round(self.crop_size[0] / self.feature_ratio) *
            self.feature_ratio),
                          int(
                              round(self.crop_size[1] / self.feature_ratio) *
                              self.feature_ratio))
        self.feature_map_sz = (self.crop_size[0] // self.feature_ratio,
                               self.crop_size[1] // self.feature_ratio)
        output_sigma = np.sqrt(
            np.floor(self.base_target_sz[0] / self.feature_ratio) *
            np.floor(self.base_target_sz[1] /
                     self.feature_ratio)) * self.output_sigma_factor
        y = gaussian2d_rolled_labels(self.feature_map_sz, output_sigma)
        self.yf = fft2(y)
        if self.interpolate_response == 1:
            self.interp_sz = (self.feature_map_sz[0] * self.feature_ratio,
                              self.feature_map_sz[1] * self.feature_ratio)
        else:
            self.interp_sz = (self.feature_map_sz[0], self.feature_map_sz[1])
        self._window = cos_window(self.feature_map_sz)
        if self.number_of_scales > 0:
            scale_exp = np.arange(
                -int(np.floor((self.number_of_scales - 1) / 2)),
                int(np.ceil((self.number_of_scales - 1) / 2)) + 1)
            self.scale_factors = self.scale_step**scale_exp
            self.min_scale_factor = self.scale_step**(np.ceil(
                np.log(max(5 / self.crop_size[0], 5 / self.crop_size[1])) /
                np.log(self.scale_step)))
            self.max_scale_factor = self.scale_step**(np.floor(
                np.log(
                    min(image.shape[0] / self.base_target_sz[1],
                        image.shape[1] / self.base_target_sz[0])) /
                np.log(self.scale_step)))
        if self.interpolate_response >= 3:
            self.ky = np.roll(
                np.arange(-int(np.floor((self.feature_map_sz[1] - 1) / 2)),
                          int(np.ceil((self.feature_map_sz[1] - 1) / 2 + 1))),
                -int(np.floor((self.feature_map_sz[1] - 1) / 2)))
            self.kx = np.roll(
                np.arange(-int(np.floor((self.feature_map_sz[0] - 1) / 2)),
                          int(np.ceil((self.feature_map_sz[0] - 1) / 2 + 1))),
                -int(np.floor((self.feature_map_sz[0] - 1) / 2))).T
            # self.kx = self.kx.reshape(self.kx.shape[0],1)
            # self.kx=np.roll(np.arange(-int(np.floor((self.feature_map_sz[1]-1)/2)),int(np.ceil((self.feature_map_sz[1]-1)/2+1))),
            #                 -int(np.floor((self.feature_map_sz[1]-1)/2)),axis=0)
            # self.ky=np.roll(np.arange(-int(np.floor((self.feature_map_sz[0]-1)/2)),int(np.ceil((self.feature_map_sz[0]-1)/2+1))),
            #                 -int(np.floor((self.feature_map_sz[0]-1)/2)),axis=0)
        self.M_prev = np.zeros(self.feature_map_sz)

        self.small_filter_sz = (int(
            np.floor(self.base_target_sz[0] / self.feature_ratio)),
                                int(
                                    np.floor(self.base_target_sz[1] /
                                             self.feature_ratio)))

        # self.scale_estimator = LPScaleEstimator(self.target_sz, config=self.scale_config)
        # self.scale_estimator.init(image, self._center, self.base_target_sz, self.current_scale_factor)

        pixels = self.get_sub_window(
            image,
            self._center,
            model_sz=self.crop_size,
            scaled_sz=(int(
                np.round(self.crop_size[0] * self.current_scale_factor)),
                       int(
                           np.round(self.crop_size[1] *
                                    self.current_scale_factor))))
        feature = self.extract_hc_feture(pixels, cell_size=self.feature_ratio)
        self.model_xf = fft2(self._window[:, :, None] * feature)

        self.g_f = self.ADMM(self.model_xf)
        #self.model_xf = self.model_xf[:,:,:,np.newaxis]
        responsef = np.sum(np.conj(self.g_f) * self.model_xf, axis=2)
        if len(np.shape(responsef)) == 4:
            responsef = np.transpose(responsef, (0, 1, 3, 2))
        # responsef = responsef[:,:,np.newaxis] 在函数中加维度
        responsef_padded = resize_dft2(responsef, self.interp_sz)
        # response in the spatial domain
        # MATLAB: response = ifft2(responsef_padded, 'symmetric')
        response = np.real(ifft2(responsef_padded))
        # response = np.real(cifft2(responsef_padded))
        # response = ifft2_sym(responsef_padded)
        self.M_prev = np.squeeze(np.fft.fftshift(response))
        # self.M_prev = np.fft.fftshift(response)
        # self.M_prev = np.sum(self.M_prev,axis=2)
        self.max_M_prev = self.M_prev.max()
        id_max_prev = np.argwhere(self.M_prev == self.max_M_prev)
        self.id_ymax_prev = id_max_prev[:, 0]
        self.id_xmax_prev = id_max_prev[:, 1]
Exemplo n.º 5
0
    def update(self, image, frame_id):
        self.frame = frame_id
        x = None
        for scale_ind in range(self.number_of_scales):
            current_scale = self.current_scale_factor * self.scale_factors[
                scale_ind]
            sub_window = self.get_sub_window(
                image,
                self._center,
                model_sz=self.crop_size,
                scaled_sz=(int(round(self.crop_size[0] * current_scale)),
                           int(round(self.crop_size[1] * current_scale))))
            feature = self.extract_hc_feture(sub_window,
                                             self.cell_size)[:, :, :,
                                                             np.newaxis]
            if x is None:
                x = feature
            else:
                x = np.concatenate((x, feature), axis=3)
        xtf = fft2(x * self._window[:, :, None, None])
        responsef = np.sum(np.conj(self.g_f)[:, :, :, None] * xtf, axis=2)
        if len(np.shape(responsef)) == 4:
            responsef = np.transpose(responsef, (0, 1, 3, 2))

        if self.interpolate_response == 2:
            self.interp_sz = (int(
                np.floor(self.yf.shape[1] * self.feature_ratio *
                         self.current_scale_factor)),
                              int(
                                  np.floor(self.yf.shape[0] *
                                           self.feature_ratio *
                                           self.current_scale_factor)))
        responsef_padded = resize_dft2(responsef, self.interp_sz)
        response = np.real(ifft2(responsef_padded))
        # response = np.real(cifft2(responsef_padded))
        # response = ifft2_sym(responsef_padded)
        if self.interpolate_response == 3:
            raise ValueError

        elif self.interpolate_response == 4:
            disp_row, disp_col, sind = resp_newton(response, responsef_padded,
                                                   self.newton_iterations,
                                                   self.ky, self.kx,
                                                   self.feature_map_sz)
        else:
            row, col, sind = np.unravel_index(np.argmax(response, axis=None),
                                              response.shape)
            disp_row = (row - 1 + int(np.floor(self.interp_sz[1] - 1) /
                                      2)) % self.interp_sz[1] - int(
                                          np.floor(
                                              (self.interp_sz[1] - 1) / 2))
            disp_col = (col - 1 + int(np.floor(self.interp_sz[0] - 1) /
                                      2)) % self.interp_sz[0] - int(
                                          np.floor(
                                              (self.interp_sz[0] - 1) / 2))

        if self.interpolate_response == 0 or self.interpolate_response == 3 or self.interpolate_response == 4:
            factor = self.feature_ratio * self.current_scale_factor * self.scale_factors[
                sind]
        elif self.interpolate_response == 1:
            factor = self.current_scale_factor * self.scale_factors[sind]
        elif self.interpolate_response == 2:
            factor = self.scale_factors[sind]
        else:
            raise ValueError
        dx, dy = int(np.round(disp_col * factor)), int(
            np.round(disp_row * factor))
        self.current_scale_factor = self.current_scale_factor * self.scale_factors[
            sind]
        self.current_scale_factor = max(self.current_scale_factor,
                                        self.min_scale_factor)
        self.current_scale_factor = min(self.current_scale_factor,
                                        self.max_scale_factor)

        # self.current_scale_factor = self.scale_estimator.update(image, self._center, self.base_target_sz,
        #                                       self.current_scale_factor)

        self._center = (self._center[0] + dx, self._center[1] + dy)
        # find peak in the map
        self.M_curr = np.fft.fftshift(response[:, :, sind])
        self.max_M_curr = self.M_curr.max()
        id_max_curr = np.argwhere(self.M_curr == self.max_M_curr)
        # self.id_xmax_curr = id_max_curr[0][0]
        # self.id_ymax_curr = id_max_curr[0][1]
        self.id_ymax_curr = id_max_curr[:, 0]
        self.id_xmax_curr = id_max_curr[:, 1]

        # do shifting of previous response map
        shift_x = self.id_xmax_curr - self.id_xmax_prev
        shift_y = self.id_ymax_curr - self.id_ymax_prev
        sz_shift_y = len(shift_y)
        sz_shift_x = len(shift_x)
        if sz_shift_y > 1:
            shift_y = shift_y[0]
        if sz_shift_x > 1:
            shift_x = shift_x[0]

        self.M_prev = np.roll(self.M_prev, shift_y, 0)
        self.M_prev = np.roll(self.M_prev, shift_x, 1)

        # self.M_prev = np.roll(self.M_prev,shift_x,0)
        # self.M_prev = np.roll(self.M_prev,shift_y,1)

        # map difference
        # map_diff(frame) = norm(abs(M_prev - M_curr))

        pixels = self.get_sub_window(
            image,
            self._center,
            model_sz=self.crop_size,
            scaled_sz=(int(round(self.crop_size[0] *
                                 self.current_scale_factor)),
                       int(round(self.crop_size[1] *
                                 self.current_scale_factor))))
        feature = self.extract_hc_feture(pixels, cell_size=self.cell_size)

        #feature=cv2.resize(pixels,self.feature_map_sz)/255-0.5
        xf = fft2(feature * self._window[:, :, None])

        ##
        self.model_xf = (
            1 - self.interp_factor) * self.model_xf + self.interp_factor * xf
        self.g_f = self.ADMM(self.model_xf)

        self.M_prev = self.M_curr
        self.max_M_prev = self.max_M_curr
        self.id_xmax_prev = self.id_xmax_curr
        self.id_ymax_prev = self.id_ymax_curr

        # target_sz= tuple(np.floor((self.base_target_sz[0]*self.current_scale_factor,self.base_target_sz[1]*self.current_scale_factor)))
        target_sz = (self.base_target_sz[0] * self.current_scale_factor,
                     self.base_target_sz[1] * self.current_scale_factor)
        # new_state = [self._center[0]-np.floor(target_sz[0]/2),self._center[1]-np.floor(target_sz[1]/2),target_sz[0],target_sz[1]]
        new_state = [
            self._center[0] - target_sz[0] / 2,
            self._center[1] - target_sz[1] / 2, target_sz[0], target_sz[1]
        ]
        # out = {'target_bbox': new_state}
        box = np.array(new_state)
        return box