Asymptotic Log-Det Sum-of-Ranks Minimization via Tensor (Alternating) Iteratively Reweighted Least Squares

06/29/2021
βˆ™
by   Sebastian KrΓ€mer, et al.
βˆ™
0
βˆ™

Affine sum-of-ranks minimization (ASRM) generalizes the affine rank minimization (ARM) problem from matrices to tensors. Here, the interest lies in the ranks of a family 𝒦 of different matricizations. Transferring our priorly discussed results on asymptotic log-det rank minimization, we show that iteratively reweighted least squares with weight strength p = 0 remains a, theoretically and practically, particularly viable method denoted as IRLS-0𝒦. As in the matrix case, we prove global convergence of asymptotic minimizers of the log-det sum-of-ranks function to desired solutions. Further, we show local convergence of IRLS-0𝒦 in dependence of the rate of decline of the therein appearing regularization parameter Ξ³β†˜ 0. For hierarchical families 𝒦, we show how an alternating version (AIRLS-0𝒦, related to prior work under the name SALSA) can be evaluated solely through tensor tree network based operations. The method can thereby be applied to high dimensions through the avoidance of exponential computational complexity. Further, the otherwise crucial rank adaption process becomes essentially superfluous even for completion problems. In numerical experiments, we show that the therefor required subspace restrictions and relaxation of the affine constraint cause only a marginal loss of approximation quality. On the other hand, we demonstrate that IRLS-0𝒦 allows to observe the theoretical phase transition also for generic tensor recoverability in practice. Concludingly, we apply AIRLS-0𝒦 to larger scale problems.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset