Exemplo n.º 1
0
    def __init__(self):
        self.max_itr = 250

        self.tolerance = 1e-3
        self.freq = 5
        self.epsilon = 0.7
        self.omega = 0.3
        self.beta = 2.0
        self.lambd = 0.5

        self.SIFT = cv2.xfeatures2d.SIFT_create()
        self.SC = ShapeContext()
Exemplo n.º 2
0
    def __init__(self):
        self.height = 224
        self.width = 224
        self.shape = np.array([224.0, 224.0])

        self.sift_weight = 2.0
        self.cnn_weight = 1.0

        self.max_itr = 200

        self.tolerance = 1e-2
        self.freq = 5  # k in the paper
        self.epsilon = 0.5
        self.omega = 0.5
        self.beta = 2.0
        self.lambd = 0.5

        self.cnnph = tf.placeholder("float", [2, 224, 224, 3])
        self.vgg = VGG16mo()
        self.vgg.build(self.cnnph)
        self.SC = ShapeContext()
Exemplo n.º 3
0
class CNN(object):
    def __init__(self):
        self.height = 224
        self.width = 224
        self.shape = np.array([224.0, 224.0])

        self.sift_weight = 2.0
        self.cnn_weight = 1.0

        self.max_itr = 200

        self.tolerance = 1e-2
        self.freq = 5  # k in the paper
        self.epsilon = 0.5
        self.omega = 0.5
        self.beta = 2.0
        self.lambd = 0.5

        self.cnnph = tf.placeholder("float", [2, 224, 224, 3])
        self.vgg = VGG16mo()
        self.vgg.build(self.cnnph)
        self.SC = ShapeContext()

    def register(self, IX, IY):

        # set parameters
        tolerance = self.tolerance
        freq = self.freq
        epsilon = self.epsilon
        omega = self.omega
        beta = self.beta
        lambd = self.lambd

        # resize image
        Xscale = 1.0 * np.array(IX.shape[:2]) / self.shape
        Yscale = 1.0 * np.array(IY.shape[:2]) / self.shape
        IX = cv2.resize(IX, (self.height, self.width))
        IY = cv2.resize(IY, (self.height, self.width))

        # CNN feature
        # propagate the images through VGG16
        IX = np.expand_dims(IX, axis=0)
        IY = np.expand_dims(IY, axis=0)
        cnn_input = np.concatenate((IX, IY), axis=0)
        with tf.Session() as sess:
            feed_dict = {self.cnnph: cnn_input}
            D1, D2, D3 = sess.run(
                [self.vgg.pool3, self.vgg.pool4, self.vgg.pool5_1],
                feed_dict=feed_dict)

        # flatten
        DX1, DY1 = np.reshape(D1[0], [-1, 256]), np.reshape(D1[1], [-1, 256])
        DX2, DY2 = np.reshape(D2[0], [-1, 512]), np.reshape(D2[1], [-1, 512])
        DX3, DY3 = np.reshape(D3[0], [-1, 512]), np.reshape(D3[1], [-1, 512])

        # normalization
        DX1, DY1 = DX1 / np.std(DX1), DY1 / np.std(DY1)
        DX2, DY2 = DX2 / np.std(DX2), DY2 / np.std(DY2)
        DX3, DY3 = DX3 / np.std(DX3), DY3 / np.std(DY3)

        del D1, D2, D3

        # compute feature space distance
        PD1 = pairwise_distance(DX1, DY1)
        PD2 = pd_expand(pairwise_distance(DX2, DY2), 2)
        PD3 = pd_expand(pairwise_distance(DX3, DY3), 4)
        PD = 1.414 * PD1 + PD2 + PD3

        del DX1, DY1, DX2, DY2, DX3, DY3, PD1, PD2, PD3

        seq = np.array([[i, j] for i in range(28) for j in range(28)],
                       dtype='int32')

        X = np.array(seq, dtype='float32') * 8.0 + 4.0
        Y = np.array(seq, dtype='float32') * 8.0 + 4.0

        # normalize
        X = (X - 112.0) / 224.0
        Y = (Y - 112.0) / 224.0

        # prematch and select points
        C_all, quality = match(PD)
        tau_max = np.max(quality)
        while np.where(quality >= tau_max)[0].shape[0] <= 128:
            tau_max -= 0.01

        C = C_all[np.where(quality >= tau_max)]
        cnt = C.shape[0]

        # select prematched feature points
        X, Y = X[C[:, 1]], Y[C[:, 0]]
        PD = PD[np.repeat(np.reshape(C[:, 1], [cnt, 1]), cnt, axis=1),
                np.repeat(np.reshape(C[:, 0], [1, cnt]), cnt, axis=0)]

        N = X.shape[0]
        M = X.shape[0]
        assert M == N

        # precalculation of feature match
        C_all, quality = match(PD)

        # compute \hat{\theta} and \delta
        tau_min = np.min(quality)
        tau_max = np.max(quality)
        while np.where(quality >= tau_max)[0].shape[0] <= 0.5 * cnt:
            tau_max -= 0.01
        tau = tau_max
        delta = (tau_max - tau_min) / 10.0

        SCX = self.SC.compute(X)

        # initialization
        Z = Y.copy()
        GRB = gaussian_radial_basis(Y, beta)
        A = np.zeros([M, 2])
        sigma2 = init_sigma2(X, Y)

        Pr = None

        Q = 0
        dQ = float('Inf')
        itr = 1

        # registration process
        while itr < self.max_itr and abs(dQ) > tolerance and sigma2 > 1e-4:
            Z_old = Z.copy()
            Q_old = Q

            # for every k iterations
            if (itr - 1) % freq == 0:
                # compute C^{conv}_{\theta}
                C = C_all[np.where(quality >= tau)]
                Lt = PD[C[:, 0], C[:, 1]]
                maxLt = np.max(Lt)
                if maxLt > 0: Lt = Lt / maxLt
                L = np.ones([M, N])
                L[C[:, 0], C[:, 1]] = Lt

                # compute C^{geo}_{\theta}
                SCZ = self.SC.compute(Z)
                SC_cost = self.SC.cost(SCZ, SCX)

                # compute C
                L = L * SC_cost

                # linear assignment
                C = lapjv(L)[1]

                # prior probability matrix
                Pr = np.ones_like(PD) * (1.0 - epsilon) / N
                Pr[np.arange(C.shape[0]), C] = 1.0
                Pr = Pr / np.sum(Pr, axis=0)

                tau = tau - delta
                if tau < tau_min: tau = tau_min

            # compute minimization
            Po, P1, Np, tmp, Q = compute(X, Y, Z_old, Pr, sigma2, omega)
            Q = Q + lambd / 2 * np.trace(np.dot(np.dot(A.transpose(), GRB), A))

            # update variables
            dP = np.diag(P1)
            t1 = np.dot(dP, GRB) + lambd * sigma2 * np.eye(M)
            t2 = np.dot(Po, X) - np.dot(dP, Y)
            A = np.dot(np.linalg.inv(t1), t2)
            sigma2 = tmp / (2.0 * Np)
            omega = 1 - (Np / N)
            if omega > 0.99: omega = 0.99
            if omega < 0.01: omega = 0.01
            Z = Y + np.dot(GRB, A)
            lambd = lambd * 0.95
            if lambd < 0.1: lambd = 0.1

            dQ = Q - Q_old
            itr = itr + 1

        print('finish: itr %d, Q %d, tau %d' % (itr, Q, tau))
        return ((X * 224.0) + 112.0) * Xscale, (
            (Y * 224.0) + 112.0) * Yscale, ((Z * 224.0) + 112.0) * Xscale
Exemplo n.º 4
0
class CNN5(object):
    def __init__(self):
        self.height = 224
        self.width = 224
        self.shape = np.array([224.0, 224.0])

        self.sift_weight = 2.0
        self.cnn_weight = 1.0

        self.max_itr = 200

        self.tolerance = 1e-2
        self.freq = 5
        self.epsilon = 0.5
        self.omega = 0.5
        self.beta = 2.0
        self.lambd = 0.5

        self.cnnph = tf.placeholder("float", [2, 224, 224, 3])
        self.vgg = VGG16mo()
        self.vgg.build(self.cnnph)
        self.SC = ShapeContext()

    def register(self, IX, IY):

        # set parameters
        tolerance = self.tolerance
        freq = self.freq
        epsilon = self.epsilon
        omega = self.omega
        beta = self.beta
        lambd = self.lambd

        # resize image
        Xscale = 1.0 * np.array(IX.shape[:2]) / self.shape
        Yscale = 1.0 * np.array(IY.shape[:2]) / self.shape
        IX = cv2.resize(IX, (self.height, self.width))
        IY = cv2.resize(IY, (self.height, self.width))

        # CNN feature
        start_time = time.time()

        IX = np.expand_dims(IX, axis=0)
        IY = np.expand_dims(IY, axis=0)
        cnn_input = np.concatenate((IX, IY), axis=0)
        with tf.Session() as sess:
            feed_dict = {self.cnnph: cnn_input}
            D = sess.run([self.vgg.pool5], feed_dict=feed_dict)[0]

        # generate combined feature

        seq = np.array([[i, j] for i in range(7) for j in range(7)],
                       dtype='int32')
        DX = D[0, seq[:, 0], seq[:, 1]]
        DY = D[1, seq[:, 0], seq[:, 1]]

        X = np.array(seq, dtype='float32') * 32.0 + 16.0
        Y = np.array(seq, dtype='float32') * 32.0 + 16.0

        # normalize

        DX = DX / np.std(DX)
        DY = DY / np.std(DY)

        X = (X - 112.0) / 224.0
        Y = (Y - 112.0) / 224.0

        # prematch and select points
        PD = pairwise_distance(DX, DY)
        C_all, quality = match(PD)

        tau_max = np.max(quality)
        while np.where(quality >= tau_max)[0].shape[0] <= 20:
            tau_max -= 0.001

        C = C_all[np.where(quality >= tau_max)]
        X, Y = X[C[:, 1]], Y[C[:, 0]]
        DX, DY = DX[C[:, 1]], DY[C[:, 0]]

        SCX = self.SC.compute(X)

        N = X.shape[0]
        M = X.shape[0]
        assert M == N

        # feature match precalc
        PD = pairwise_distance(DX, DY)
        C_all, quality = match(PD)
        tau_min = np.min(quality)
        tau_max = np.max(quality)
        while np.where(quality >= tau_max)[0].shape[0] <= 8:
            tau_max -= 0.001
        tau = tau_max
        delta = (tau_max - tau_min) / 10.0

        T = Y.copy()
        GRB = gaussian_radial_basis(Y, beta)
        A = np.zeros([M, 2])
        sigma2 = init_sigma2(X, Y)

        Pm = None

        Q = 0
        dQ = float('Inf')
        itr = 1

        # registration process
        while itr < self.max_itr and abs(dQ) > tolerance and sigma2 > 1e-4:
            T_old = T.copy()
            Q_old = Q

            # refine
            if (itr - 1) % freq == 0:
                C = C_all[np.where(quality >= tau)]
                Lt = PD[C[:, 0], C[:, 1]]
                maxLt = np.max(Lt)
                if maxLt > 0: Lt = Lt / maxLt
                L = np.ones([M, N])
                L[C[:, 0], C[:, 1]] = Lt

                SCT = self.SC.compute(T)
                SC_cost = self.SC.cost(SCT, SCX)
                L = L * SC_cost

                C = lapjv(L)[1]
                Pm = np.ones_like(PD) * (1.0 - epsilon) / N
                Pm[np.arange(C.shape[0]), C] = 1.0
                Pm = Pm / np.sum(Pm, axis=0)

                tau = tau - delta
                if tau < tau_min: tau = tau_min

                # plt.gca().invert_yaxis()
                # plt.scatter(T[:, 1], T[:, 0])
                # plt.show()

            # compute minimization
            Po, P1, Np, tmp, Q = compute(X, Y, T_old, Pm, sigma2, omega)
            Q = Q + lambd / 2 * np.trace(np.dot(np.dot(A.transpose(), GRB), A))

            # update variables
            dP = np.diag(P1)
            t1 = np.dot(dP, GRB) + lambd * sigma2 * np.eye(M)
            t2 = np.dot(Po, X) - np.dot(dP, Y)
            A = np.dot(np.linalg.inv(t1), t2)
            sigma2 = tmp / (2.0 * Np)
            omega = 1 - (Np / N)
            if omega > 0.99: omega = 0.99
            if omega < 0.01: omega = 0.01
            T = Y + np.dot(GRB, A)
            lambd = lambd * 0.95
            if lambd < 0.1: lambd = 0.1

            dQ = Q - Q_old
            itr = itr + 1

            # print(itr, Q, tau)

        print('finish: itr %d, Q %d, tau %d' % (itr, Q, tau))
        return ((X * 224.0) + 112.0) * Xscale, (
            (Y * 224.0) + 112.0) * Yscale, ((T * 224.0) + 112.0) * Xscale
Exemplo n.º 5
0
class SIFT(object):
    def __init__(self):
        self.max_itr = 250

        self.tolerance = 1e-3
        self.freq = 5
        self.epsilon = 0.7
        self.omega = 0.3
        self.beta = 2.0
        self.lambd = 0.5

        self.SIFT = cv2.xfeatures2d.SIFT_create()
        self.SC = ShapeContext()

    def register(self, IX, IY):

        # set parameters
        tolerance = self.tolerance
        freq = self.freq
        epsilon = self.epsilon
        omega = self.omega
        beta = self.beta
        lambd = self.lambd

        # SIFT
        start_time = time.time()
        IX_gray = cv2.cvtColor(IX, cv2.COLOR_BGR2GRAY)
        IY_gray = cv2.cvtColor(IY, cv2.COLOR_BGR2GRAY)

        XS0, DXS0 = self.SIFT.detectAndCompute(IX_gray, None)
        YS0, DYS0 = self.SIFT.detectAndCompute(IY_gray, None)
        XSPD = [(XS0[i], DXS0[i]) for i in range(len(XS0))
                if XS0[i].response > 0.025]
        YSPD = [(YS0[i], DYS0[i]) for i in range(len(YS0))
                if YS0[i].response > 0.025]
        XSPD.sort(key=lambda p: p[0].response, reverse=True)
        YSPD.sort(key=lambda p: p[0].response, reverse=True)
        XSPD = XSPD[:800]
        YSPD = YSPD[:800]

        XS = [t[0].pt for t in XSPD]
        YS = [t[0].pt for t in YSPD]
        DXS = [t[1] for t in XSPD]
        DYS = [t[1] for t in YSPD]

        XS, YS, DXS, DYS = np.array(XS), np.array(YS), np.array(DXS), np.array(
            DYS)
        XS = np.array([XS[:, 1], XS[:, 0]]).transpose()
        YS = np.array([YS[:, 1], YS[:, 0]]).transpose()

        # select points
        PD = pairwise_distance(DXS, DYS)
        C_all, quality = match(PD)
        C = C_all[np.where(quality >= 1.15)]
        X, Y = XS[C[:, 1], :], YS[C[:, 0], :]
        DX, DY = DXS[C[:, 1], :], DYS[C[:, 0], :]

        #normalize
        X_shape = np.array(IX_gray.shape[:2], dtype='float32')
        Y_shape = np.array(IY_gray.shape[:2], dtype='float32')
        X_center = 0.5 * X_shape
        Y_center = 0.5 * Y_shape
        X = (X - X_center) / X_shape
        Y = (Y - Y_center) / Y_shape

        SCX = self.SC.compute(X)

        N = X.shape[0]
        M = Y.shape[0]
        assert M == N

        # prepare feature
        PD = pairwise_distance(DX, DY)
        C_all, quality = match(PD)
        tau = 2.5
        while tau > 1 and np.where(
                quality >= tau)[0].shape[0] < 0.25 * quality.shape[0]:
            tau -= 0.1
        delta = (tau - 1.0) / 10.0

        # registration process
        T = Y.copy()
        GRB = gaussian_radial_basis(Y, beta)
        A = np.zeros([M, 2])
        sigma2 = init_sigma2(X, Y)

        Pm = None

        Q = 0
        dQ = float('Inf')
        itr = 1

        # registration process
        while itr < self.max_itr and abs(dQ) > tolerance and sigma2 > 1e-4:
            T_old = T.copy()
            Q_old = Q

            # refine
            if (itr - 1) % freq == 0:
                C = C_all[np.where(quality >= tau)]
                Lt = PD[C[:, 0], C[:, 1]]
                maxLt = np.max(Lt)
                if maxLt > 0: Lt = Lt / maxLt
                L = np.ones([M, N])
                L[C[:, 0], C[:, 1]] = Lt

                SCT = self.SC.compute(T)
                SC_cost = self.SC.cost(SCT, SCX)
                L = L * SC_cost

                C = lapjv(L)[1]
                Pm = np.ones_like(PD) * (1.0 - epsilon) / N
                Pm[np.arange(C.shape[0]), C] = 1.0
                Pm = Pm / np.sum(Pm, axis=0)

                tau = tau - delta
                if tau < 1: tau = 1

                # plt.scatter(T[:, 0], T[:, 1])
                # plt.show()

            # compute minimization
            Po, P1, Np, tmp, Q = compute(X, Y, T_old, Pm, sigma2, omega)
            Q = Q + lambd / 2 * np.trace(np.dot(np.dot(A.transpose(), GRB), A))

            # update variables
            dP = np.diag(P1)
            t1 = np.dot(dP, GRB) + lambd * sigma2 * np.eye(M)
            t2 = np.dot(Po, X) - np.dot(dP, Y)
            A = np.dot(np.linalg.inv(t1), t2)
            sigma2 = tmp / (2.0 * Np)
            omega = 1 - (Np / N)
            if omega > 0.99: omega = 0.99
            if omega < 0.01: omega = 0.01
            T = Y + np.dot(GRB, A)
            lambd = lambd * 0.95
            if lambd < 0.1: lambd = 0.1

            dQ = Q - Q_old
            itr = itr + 1

            # print(itr, Q, tau)

        print('finish: itr %d, Q %d, tau %d' % (itr, Q, tau))
        return X * X_shape + X_center, Y * Y_shape + Y_center, T * X_shape + X_center