Sparse Estimation of High-Dimensional Inverse Covariance Matrices with Explicit Eigenvalue Constraints

2021 
Firstly, this paper proposes a generalized log-determinant optimization model with the purpose of estimating the high-dimensional sparse inverse covariance matrices. Under the normality assumption, the zero components in the inverse covariance matrices represent the conditional independence between pairs of variables given all the other variables. The generalized model considered in this study, because of the setting of the eigenvalue bounded constraints, covers a large number of existing estimators as special cases. Secondly, rather than directly tracking the challenging optimization problem, this paper uses a couple of alternating direction methods of multipliers (ADMM) to solve its dual model where 5 separable structures are contained. The first implemented algorithm is based on a single Gauss–Seidel iteration, but it does not necessarily converge theoretically. In contrast, the second algorithm employs the symmetric Gauss–Seidel (sGS) based ADMM which is equivalent to the 2-block iterative scheme from the latest sGS decomposition theorem. Finally, we do numerical simulations using the synthetic data and the real data set which show that both algorithms are very effective in estimating high-dimensional sparse inverse covariance matrix.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    40
    References
    0
    Citations
    NaN
    KQI
    []