def test_permutation_one_sided_square_matrices_rows_permuted(n): r"""Test one-sided permutation Procrustes with square matrices and permuted rows.""" array_a = np.random.uniform(-10.0, 10.0, (n, n)) perm = generate_random_permutation_matrix(n) # permuted array_b array_b = np.dot(array_a, perm) # procrustes with no translate and scale res = permutation(array_a, array_b) assert_almost_equal(res.t, perm, decimal=6) assert_almost_equal(res.error, 0.0, decimal=6)
def test_permutation_one_sided_with_translate_scale(m, n): r"""Test permutation one_sided by translated and scaled arrays.""" array_a = np.random.uniform(-10.0, 10.0, (m, n)) # array_b is scaled, translated, and permuted array_a perm = generate_random_permutation_matrix(n) # obtain random translation/shift array and permute the array. shift = np.random.uniform(-10.0, 10.0, (n, )) array_b = 3.78 * array_a + shift array_b = np.dot(array_b, perm) # permutation procrustes res = permutation(array_a, array_b, translate=True, scale=True) assert_almost_equal(res.t, perm, decimal=6) assert_almost_equal(res.error, 0.0, decimal=6)
def test_permutation_translate_scale(): r"""Test permutation by scaled arrays.""" # square array array_a = np.array([[1, 5, 8, 4], [1, 5, 7, 2], [1, 6, 9, 3], [2, 7, 9, 4]]) # array_b is scaled, translated, and permuted array_a perm = np.array([[1, 0, 0, 0], [0, 0, 0, 1], [0, 0, 1, 0], [0, 1, 0, 0]]) array_b = 3.78 * array_a + np.array([6, 1, 5, 3]) array_b = np.dot(array_b, perm) # permutation procrustes res = permutation(array_a, array_b, translate=True, scale=True) assert_almost_equal(res["array_u"], perm, decimal=6) assert_almost_equal(res["error"], 0., decimal=6)
def test_permutation_columns(): r"""Test permutation Procrustes with permuted rows.""" # square array array_a = np.array([[1, 5, 8, 4], [1, 5, 7, 2], [1, 6, 9, 3], [2, 7, 9, 4]]) # permutation perm = np.array([[0, 0, 0, 1], [0, 0, 1, 0], [1, 0, 0, 0], [0, 1, 0, 0]]) # permuted array_b array_b = np.dot(array_a, perm) # procrustes with no translate and scale res = permutation(array_a, array_b) assert_almost_equal(res["array_u"], perm, decimal=6) assert_almost_equal(res["error"], 0., decimal=6)
def test_permutation_columns(): # square array array_a = np.array([[1, 5, 8, 4], [1, 5, 7, 2], [1, 6, 9, 3], [2, 7, 9, 4]]) # permutation perm = np.array([[0, 0, 0, 1], [0, 0, 1, 0], [1, 0, 0, 0], [0, 1, 0, 0]]) # permuted array_b array_b = np.dot(array_a, perm) # procrustes with no translate and scale new_a, new_b, array_p, e_opt = permutation(array_a, array_b) assert_almost_equal(array_a, new_a, decimal=6) assert_almost_equal(array_b, new_b, decimal=6) assert_almost_equal(array_p, perm, decimal=6) assert_almost_equal(e_opt, 0., decimal=6)
def test_permutation_translate_scale_padd(): r"""Test permutation by scaled arrays with zero paddings.""" # rectangular array_a array_a = np.array([[118.51, 515.27, 831.61, 431.62], [161.61, 535.13, 763.16, 261.63], [116.31, 661.34, 961.31, 363.15], [236.16, 751.36, 913.51, 451.22]]) # array_b is scaled, translated, and permuted array_a array_b = 51.63 * array_a + np.array([56.24, 79.32, 26.15, 49.52]) perm = np.array([[0., 0., 0., 1.], [0., 1., 0., 0.], [0., 0., 1., 0.], [1., 0., 0., 0.]]) array_b = np.dot(array_b, perm) # check res = permutation(array_a, array_b, translate=True, scale=True) assert_almost_equal(res["array_u"], perm, decimal=6) assert_almost_equal(res["error"], 0., decimal=6)
def test_permutation_one_sided_columns_pad(m, n, ncols, nrows): r"""Test one-sided permutation by permuted columns along with padded zeros.""" array_a = np.random.uniform(-10.0, 10.0, (m, n)) perm = generate_random_permutation_matrix(n) # permuted array_b array_b = np.dot(array_a, perm) # padded array b with zero row and columns array_b = np.concatenate((array_b, np.zeros((m, ncols))), axis=1) array_b = np.concatenate((array_b, np.zeros((nrows, n + ncols))), axis=0) if m < n: array_a = np.concatenate((array_a, np.zeros((n - m, n))), axis=0) # procrustes with no translate and scale res = permutation(array_a, array_b, unpad_col=True, unpad_row=True) # Test that the unpadded b is the same as the original b. assert_almost_equal(res.new_b, np.dot(array_a, perm), decimal=6) # Test that the permutation and the error are the same/zero. assert_almost_equal(res.t, perm, decimal=6) assert_almost_equal(res.error, 0.0, decimal=6)
def test_permutation_columns_pad(): r"""Test permutation by permuted columns along with padded zeros.""" # square array array_a = np.array([[1, 5, 8, 4], [1, 5, 7, 2], [1, 6, 9, 3], [2, 7, 9, 4]]) # permutation perm = np.array([[0, 0, 0, 1], [0, 0, 1, 0], [1, 0, 0, 0], [0, 1, 0, 0]]) # permuted array_b array_b = np.dot(array_a, perm) # padded arrays with zero row and columns array_a = np.concatenate((array_a, np.array([[0], [0], [0], [0]])), axis=1) array_b = np.concatenate((array_b, np.array([[0, 0, 0, 0]])), axis=0) # procrustes with no translate and scale new_a, new_b, array_p, e_opt = permutation(array_a, array_b, remove_zero_col=True, remove_zero_row=True, translate=False, scale=False) assert_almost_equal(new_a, array_a[:, :-1], decimal=6) assert_almost_equal(new_b, array_b[:-1, :], decimal=6) assert_almost_equal(array_p, perm, decimal=6) assert_almost_equal(e_opt, 0., decimal=6)
def softassign( a: np.ndarray, b: np.ndarray, pad: bool = True, translate: bool = False, scale: bool = False, unpad_col: bool = False, unpad_row: bool = False, check_finite: bool = True, weight: Optional[np.ndarray] = None, iteration_soft: int = 50, iteration_sink: int = 200, beta_r: float = 1.10, beta_f: float = 1.0e5, epsilon: float = 0.05, epsilon_soft: float = 1.0e-3, epsilon_sink: float = 1.0e-3, k: float = 0.15, gamma_scaler: float = 1.01, n_stop: int = 3, adapted: bool = True, beta_0: Optional[float] = None, m_guess: Optional[float] = None, iteration_anneal: Optional[int] = None, kopt: bool = False, kopt_k: int = 3, ) -> ProcrustesResult: r""" Find the transformation matrix for 2-sided permutation Procrustes with softassign algorithm. Parameters ---------- a : ndarray The 2D-array :math:`\mathbf{A}_{m \times n}` which is going to be transformed. b : ndarray The 2D-array :math:`\mathbf{B}_{m \times n}` representing the reference. pad : bool, optional Add zero rows (at the bottom) and/or columns (to the right-hand side) of matrices :math:`\mathbf{A}` and :math:`\mathbf{B}` so that they have the same shape. translate : bool, optional If True, both arrays are centered at origin (columns of the arrays will have mean zero). scale : bool, optional If True, both arrays are normalized with respect to the Frobenius norm, i.e., :math:`\text{Tr}\left[\mathbf{A}^\dagger\mathbf{A}\right] = 1` and :math:`\text{Tr}\left[\mathbf{B}^\dagger\mathbf{B}\right] = 1`. unpad_col : bool, optional If True, zero columns (with values less than 1.0e-8) on the right-hand side of the intial :math:`\mathbf{A}` and :math:`\mathbf{B}` matrices are removed. unpad_row : bool, optional If True, zero rows (with values less than 1.0e-8) at the bottom of the intial :math:`\mathbf{A}` and :math:`\mathbf{B}` matrices are removed. check_finite : bool, optional If true, convert the input to an array, checking for NaNs or Infs. Default=True. weight : ndarray, optional The 1D-array representing the weights of each row of :math:`\mathbf{A}`. This defines the elements of the diagonal matrix :math:`\mathbf{W}` that is multiplied by :math:`\mathbf{A}` matrix, i.e., :math:`\mathbf{A} \rightarrow \mathbf{WA}`. iteration_soft : int, optional Number of iterations for softassign loop. iteration_sink : int, optional Number of iterations for Sinkhorn loop. beta_r : float, optional Annealing rate which should greater than 1. beta_f : float, optional The final inverse temperature. epsilon : float, optional The tolerance value for annealing loop. epsilon_soft : float, optional The tolerance value used for softassign. epsilon_sink : float, optional The tolerance value used for Sinkhorn loop. If adapted version is used, it will use the adapted tolerance value for Sinkhorn instead. k : float, optional This parameter controls how much tighter the coverage threshold for the interior loop should be than the coverage threshold for the loops outside. It has be be within the integral :math:`(0,1)`. gamma_scaler : float, optional This parameter ensures the quadratic cost function including self-amplification positive define. n_stop : int, optional Number of running steps after the calculation converges in the relaxation procedure. adapted : bool, optional If adapted, this function will use the tighter covergence threshold for the interior loops. beta_0 : float, optional Initial inverse temperature. beta_f : float, optional Final inverse temperature. m_guess : ndarray, optional The initial guess of the doubly-stochastic matrix. iteration_anneal : int, optional Number of iterations for annealing loop. kopt : bool, optional If True, the k_opt heuristic search will be performed. kopt_k : int, optional Defines the oder of k-opt heuristic local search. For example, kopt_k=3 leads to a local search of 3 items and kopt_k=2 only searches for two items locally. weight : ndarray, optional The weighting matrix. Returns ------- res : ProcrustesResult The Procrustes result represented as a class:`utils.ProcrustesResult` object. Notes ----- Quadratic assignment problem (QAP) has played a very special but fundamental role in combinatorial optimization problems. The problem can be defined as a optimization problem to minimize the cost to assign a set of facilities to a set of locations. The cost is a function of the flow between the facilities and the geographical distances among various facilities. The objective function (also named loss function in machine learning) is defined as .. math:: E_{qap}(M, \mu, \nu) = - \frac{1}{2}\Sigma_{aibj}C_{ai;bj}M_{ai}M_{bj} + \Sigma_{a}{\mu}_a (\Sigma_i M_{ai} -1) \\ + \Sigma_i {\nu}_i (\Sigma_i M_{ai} -1) - \frac{\gamma}{2}\Sigma_{ai} {M_{ai}}^2 + \frac{1}{\beta} \Sigma_{ai} M_{ai}\log{M_{ai}} where :math:`C_{ai,bj}` is the benefit matrix, :math:`M` is the desired :math:`N \times N` permutation matrix. :math:`E` is the energy function which comes along with a self-amplification term with `\gamma`, two Lagrange parameters :math:`\mu` and :math:`\nu` for constrained optimization and :math:`M_{ai} \log{M_{ai}}` servers as a barrier function which ensures positivity of :math:`M_{ai}`. The inverse temperature :math:`\beta` is a deterministic annealing control parameter. Examples -------- >>> import numpy as np >>> array_a = np.array([[4, 5, 3, 3], [5, 7, 3, 5], ... [3, 3, 2, 2], [3, 5, 2, 5]]) # define a random matrix >>> perm = np.array([[0., 0., 1., 0.], [1., 0., 0., 0.], ... [0., 0., 0., 1.], [0., 1., 0., 0.]]) # define b by permuting array_a >>> b = np.dot(perm.T, np.dot(a, perm)) >>> new_a, new_b, M_ai, error = softassign(a,b,unpad_col=False,unpad_row=False) >>> M_ai # the permutation matrix array([[0., 0., 1., 0.], [1., 0., 0., 0.], [0., 0., 0., 1.], [0., 1., 0., 0.]]) >>> error # the error 0.0 """ # pylint: disable-msg=too-many-arguments # pylint: disable-msg=too-many-branches # todo: add linear_cost_func with default value 0 # Check beta_r if beta_r <= 1: raise ValueError("Argument beta_r cannot be less than 1.") new_a, new_b = setup_input_arrays(a, b, unpad_col, unpad_row, pad, translate, scale, check_finite, weight) # Check that A & B are square and that they match each other. if new_a.shape[0] != new_a.shape[1]: raise ValueError(f"Matrix A should be square but A.shape={new_a.shape}" "Check pad, unpad_col, and unpad_row arguments.") if new_b.shape[0] != new_b.shape[1]: raise ValueError(f"Matrix B should be square but B.shape={new_b.shape}" "Check pad, unpad_col, and unpad_row arguments.") if new_a.shape != new_b.shape: raise ValueError(f"New matrix A {new_a.shape} should match the new" f" matrix B shape {new_b.shape}.") # Initialization # Compute the benefit matrix array_c = np.kron(new_a, new_b) # Get the shape of A (B and the permutation matrix as well) row_num = new_a.shape[0] c_tensor = array_c.reshape(row_num, row_num, row_num, row_num) # Compute the beta_0 gamma = _compute_gamma(array_c, row_num, gamma_scaler) if beta_0 is None: c_gamma = array_c + gamma * (np.identity(row_num * row_num)) eival_gamma = np.amax(np.abs(np.linalg.eigvalsh(c_gamma))) beta_0 = gamma_scaler * max(1.0e-10, eival_gamma / row_num) beta_0 = 1 / beta_0 else: beta_0 *= row_num beta = beta_0 # We will use iteration_anneal if provided even if the final inverse temperature is specified # iteration_anneal is not None, beta_f can be None or not if iteration_anneal is not None: beta_f = beta_0 * np.power(beta_r, iteration_anneal) * row_num # iteration_anneal is None and beta_f is not None elif iteration_anneal is None and beta_f is not None: beta_f *= row_num # Both iteration_anneal and beta_f are None else: raise ValueError( "We must specify at least one of iteration_anneal and beta_f and " "specify only one is recommended.") # Initialization of m_ai # check shape of m_guess if m_guess is not None: if np.any(m_guess < 0): raise ValueError( "The initial guess of permutation matrix cannot contain any negative values." ) if m_guess.shape[0] == row_num and m_guess.shape[1] == row_num: array_m = m_guess else: warnings.warn( f"The shape of m_guess does not match ({row_num}, {row_num})." "Use random initial guess instead.") array_m = np.abs( np.random.normal(loc=1.0, scale=0.1, size=(row_num, row_num))) else: # m_relax_old = 1 / N + np.random.rand(N, N) array_m = np.abs( np.random.normal(loc=1.0, scale=0.1, size=(row_num, row_num))) array_m[array_m < 0] = 0 array_m = array_m / row_num nochange = 0 if adapted: epsilon_sink = epsilon_soft * k while beta < beta_f: # relaxation m_old_beta = deepcopy(array_m) # softassign loop for _ in np.arange(iteration_soft): m_old_soft = deepcopy(array_m) # Compute Z in relaxation step # C_gamma_tensor = C_gamma.reshape(N, N, N, N) # Z = -np.einsum('ijkl,jl->ik', C_gamma_tensor, M) # Z -= linear_cost_func array_z = np.einsum("aibj,bj->ai", c_tensor, array_m) array_z += gamma * array_m # soft_assign array_m = np.exp(beta * array_z) # Sinkhorn loop for _ in np.arange(iteration_sink): # Row normalization array_m = array_m / array_m.sum(axis=1, keepdims=1) # Column normalization array_m = array_m / array_m.sum(axis=0, keepdims=1) # Compute the delata_M_sink if np.amax(np.abs(array_m.sum(axis=1, keepdims=1) - 1)) < epsilon_sink: array_m = array_m / array_m.sum(axis=1, keepdims=1) break change_soft = np.amax(np.abs(array_m - m_old_soft)) # pylint: disable-msg=no-else-break if change_soft < epsilon_soft: break else: if adapted: epsilon_sink = change_soft * k else: continue change_annealing = np.amax(np.abs(array_m - m_old_beta)) if change_annealing < epsilon: nochange += 1 if nochange > n_stop: break else: nochange = 0 beta *= beta_r if adapted: epsilon_soft = change_soft * k epsilon_sink = epsilon_soft * k # Compute the error array_m = permutation(np.eye(array_m.shape[0]), array_m)["t"] # k-opt heuristic if kopt: fun_error = lambda p: compute_error(new_a, new_b, p, p.T) array_m, error = kopt_heuristic_single(fun_error, p0=array_m, k=kopt_k) else: error = compute_error(new_a, new_b, array_m, array_m.T) return ProcrustesResult(error=error, new_a=new_a, new_b=new_b, t=array_m, s=None)
def softassign(array_a, array_b, iteration_soft=50, iteration_sink=200, beta_r=1.10, beta_f=1.e5, epsilon=0.05, epsilon_soft=1.e-3, epsilon_sink=1.e-3, k=0.15, gamma_scaler=1.01, n_stop=3, pad_mode='row-col', remove_zero_col=True, remove_zero_row=True, translate=False, scale=False, check_finite=True, adapted=True, beta_0=None, m_guess=None, iteration_anneal=None): r""" Find the transformation matrix for 2-sided permutation Procrustes with softassign algorithm. Parameters ---------- array_a : numpy.ndarray The 2d-array :math:`\mathbf{A}_{m \times n}` which is going to be transformed. array_b : numpy.ndarray The 2d-array :math:`\mathbf{B}_{m \times n}` representing the reference. iteration_soft : int, optional Number of iterations for softassign loop. Default=50. iteration_sink : int, optional Number of iterations for Sinkhorn loop. Default=50. linear_cost_func : numpy.ndarray Linear cost function. Default=0. beta_r : float, optional Annealing rate which should greater than 1. Default=1.10. beta_f : float, optional The final inverse temperature. Default=1.e5. epsilon : float, optional The tolerance value for annealing loop. Default=0.05. epsilon_soft : float, optional The tolerance value used for softassign. Default=1.e-3. epsilon_sink : float, optional The tolerance value used for Sinkhorn loop. If adapted version is used, it will use the adapted tolerance value for Sinkhorn instead. Default=1.e-3. k : float, optional This parameter controls how much tighter the coverage threshold for the interior loop should be than the coverage threshold for the loops outside. It has be be within the integral :math:`\(0,1\)`. Default=0.15. gamma_scaler : float This parameter ensures the quadratic cost function including self-amplification positive define. Default=1.01. n_stop : int, optional Number of running steps after the calculation converges in the relaxation procedure. Default=10. pad_mode : str, optional Specifying how to pad the arrays, listed below. Default="row-col". - "row" The array with fewer rows is padded with zero rows so that both have the same number of rows. - "col" The array with fewer columns is padded with zero columns so that both have the same number of columns. - "row-col" The array with fewer rows is padded with zero rows, and the array with fewer columns is padded with zero columns, so that both have the same dimensions. This does not necessarily result in square arrays. - "square" The arrays are padded with zero rows and zero columns so that they are both squared arrays. The dimension of square array is specified based on the highest dimension, i.e. :math:`\text{max}(n_a, m_a, n_b, m_b)`. remove_zero_col : bool, optional If True, zero columns (values less than 1e-8) on the right side will be removed. Default=True. remove_zero_row : bool, optional If True, zero rows (values less than 1e-8) on the bottom will be removed. Default=True. translate : bool, optional If True, both arrays are translated to be centered at origin, ie columns of the arrays will have mean zero. Default=False. scale : bool, optional If True, both arrays are normalized to one with respect to the Frobenius norm, ie :math:`Tr(A^T A) = 1`. Default=False. check_finite : bool, optional If true, convert the input to an array, checking for NaNs or Infs. Default=True. adapted : bool, optional If adapted, this function will use the tighter covergence threshold for the interior loops. Default=True. beta_0 : float, optional Initial inverse temperature. Default=None. beta_f : float, optional Final inverse temperature. Default=None. m_guess : numpy.ndarray, optional The initial guess of the doubly-stochastic matrix. Default=None. iteration_anneal : int, optional Number of iterations for annealing loop. Default=None. Returns ------- new_a : numpy.ndarray The transformed numpy.ndarray A. new_b : numpy.ndarray The transformed numpy.ndarray B. m_ai : numpy.ndarray The optimum permutation transformation matrix. e_opt : float Two-sided Procrustes error. Notes ----- Quadratic assignment problem (QAP) has played a very special but fundamental role in combinatorial optimization problems. The problem can be defined as a optimization problem to minimize the cost to assign a set of facilities to a set of locations. The cost is a function of the flow between the facilities and the geographical distances among various facilities. The objective function (also named loss function in machine learning) is defined as [1]_ .. math:: E_{qap}(M, \mu, \nu) = - \frac{1}{2}\Sigma_{aibj}C_{ai;bj}M_{ai}M_{bj} + \Sigma_{a}{\mu}_a (\Sigma_i M_{ai} -1) \\ + \Sigma_i {\nu}_i (\Sigma_i M_{ai} -1) - \frac{\gamma}{2}\Sigma_{ai} {M_{ai}}^2 + \frac{1}{\beta} \Sigma_{ai} M_{ai}\log{M_{ai}} where :math:`C_{ai,bj}` is the benefit matrix, :math:`M` is the desired :math:`N \times N` permutation matrix. :math:`E` is the energy function which comes along with a self-amplification term with `\gamma`, two Lagrange parameters :math:`\mu` and :math:`\nu` for constrained optimization and :math:`M_{ai} \log{M_{ai}}` servers as a barrier function which ensures positivity of :math:`M_{ai}`. The inverse temperature :math:`\beta` is a deterministic annealing control parameter. More detailed information about the algorithm can be referred to Rangarajan's paper and the details about the algorithm can be found at [2]_. References ---------- .. [1] Rangarajan, Anand and Yuille, Alan L and Gold, Steven and Mjolsness, Eric, "A convergence proof for the softassign quadratic assignment algorithm", Advances in Neural Information Processing Systems, page 620-626, 1997. .. [2] Stefan Roth, "Analysis of a Deterministic Annealing Method for Graph Matching and Quadratic Assignment", Ph.D. thesis, University of Mannheim, 2001 Examples -------- >>> import numpy as np >>> array_a = np.array([[4, 5, 3, 3], [5, 7, 3, 5],\ [3, 3, 2, 2], [3, 5, 2, 5]]) # define a random matrix >>> perm = np.array([[0., 0., 1., 0.], [1., 0., 0., 0.],\ [0., 0., 0., 1.], [0., 1., 0., 0.]]) # define array_b by permuting array_a >>> array_b = np.dot(perm.T, np.dot(array_a, perm)) >>> new_a, new_b, M_ai, e_opt = softassign(array_a, array_b,\ remove_zero_col=False,\ remove_zero_row=False) >>> M_ai # the permutation matrix array([[0., 0., 1., 0.], [1., 0., 0., 0.], [0., 0., 0., 1.], [0., 1., 0., 0.]]) >>> e_opt # the error 0.0 """ # pylint: disable-msg=too-many-arguments # pylint: disable-msg=too-many-branches # todo: add linear_cost_func with default value 0 # Check beta_r if beta_r <= 1: raise ValueError("Argument beta_r cannot be less than 1.") array_a, array_b = setup_input_arrays(array_a, array_b, remove_zero_col, remove_zero_row, pad_mode, translate, scale, check_finite) # Initialization # Compute the benefit matrix array_c = np.kron(array_a, array_b) # Get the shape of A (B and the permutation matrix as well) row_num = array_a.shape[0] c_tensor = array_c.reshape(row_num, row_num, row_num, row_num) # Compute the beta_0 gamma = _compute_gamma(array_c, row_num, gamma_scaler) if beta_0 is None: c_gamma = array_c + gamma * (np.identity(row_num * row_num)) eival_gamma = np.amax(np.abs(np.linalg.eigvalsh(c_gamma))) beta_0 = gamma_scaler * max(1.e-10, eival_gamma / row_num) beta_0 = 1 / beta_0 else: beta_0 *= row_num beta = beta_0 # We will use iteration_anneal if provided even if the final inverse temperature is specified # iteration_anneal is not None, beta_f can be None or not if iteration_anneal is not None: beta_f = beta_0 * np.power(beta_r, iteration_anneal) * row_num # iteration_anneal is None and beta_f is not None elif iteration_anneal is None and beta_f is not None: beta_f *= row_num # Both iteration_anneal and beta_f are None else: raise ValueError("We must specify at least one of iteration_anneal and beta_f and " "specify only one is recommended.") # Initialization of m_ai # check shape of m_guess if m_guess is not None: if np.any(m_guess < 0): raise ValueError( "The initial guess of permutation matrix cannot contain any negative values.") if m_guess.shape[0] == row_num and m_guess.shape[1] == row_num: array_m = m_guess else: warnings.warn("The shape of m_guess does not match ({0}, {0})." "Use random initial guess instead.".format(row_num)) array_m = np.abs(np.random.normal(loc=1.0, scale=0.1, size=(row_num, row_num))) else: # m_relax_old = 1 / N + np.random.rand(N, N) array_m = np.abs(np.random.normal(loc=1.0, scale=0.1, size=(row_num, row_num))) array_m[array_m < 0] = 0 array_m = array_m / row_num nochange = 0 if adapted: epsilon_sink = epsilon_soft * k while beta < beta_f: # relaxation m_old_beta = deepcopy(array_m) # softassign loop for _ in np.arange(iteration_soft): m_old_soft = deepcopy(array_m) # Compute Z in relaxation step # C_gamma_tensor = C_gamma.reshape(N, N, N, N) # Z = -np.einsum('ijkl,jl->ik', C_gamma_tensor, M) # Z -= linear_cost_func array_z = np.einsum('aibj,bj->ai', c_tensor, array_m) array_z += gamma * array_m # soft_assign array_m = np.exp(beta * array_z) # Sinkhorn loop for _ in np.arange(iteration_sink): # Row normalization array_m = array_m / array_m.sum(axis=1, keepdims=1) # Column normalization array_m = array_m / array_m.sum(axis=0, keepdims=1) # Compute the delata_M_sink if np.amax(np.abs(array_m.sum(axis=1, keepdims=1) - 1)) < epsilon_sink: array_m = array_m / array_m.sum(axis=1, keepdims=1) break change_soft = np.amax(np.abs(array_m - m_old_soft)) # pylint: disable-msg=no-else-break if change_soft < epsilon_soft: break else: if adapted: epsilon_sink = change_soft * k else: continue change_annealing = np.amax(np.abs(array_m - m_old_beta)) if change_annealing < epsilon: nochange += 1 if nochange > n_stop: break else: nochange = 0 beta *= beta_r if adapted: epsilon_soft = change_soft * k epsilon_sink = epsilon_soft * k # Compute the error _, _, array_m, _ = permutation(np.eye(array_m.shape[0]), array_m) e_opt = error(array_a, array_b, array_m, array_m) return array_a, array_b, array_m, e_opt