论文标题
$ \ ell_1 $ -norm是否在拉普拉斯(Laplacian)约束图形模型下学习稀疏图?
Does the $\ell_1$-norm Learn a Sparse Graph under Laplacian Constrained Graphical Models?
论文作者
论文摘要
我们考虑在拉普拉斯(Laplacian)约束的高斯图形模型下学习稀疏图的问题。该问题可以作为对拉普拉斯约束精度矩阵的受惩罚最大似然估计。就像在经典的图形套索问题中一样,最近的作品利用了$ \ ell_1 $ -norm正则化,目的是在拉普拉斯(Laplacian)约束的精度矩阵估计中促进稀疏性。但是,我们发现广泛使用的$ \ ell_1 $ -norm在此问题中施加稀疏解决方案无效。通过经验证据,我们观察到随着正则化参数的增加,非零图的数量会增长。从理论的角度来看,我们证明了一个大的正则化参数将出奇地导致完整的图形,即每对顶点都通过边缘连接。为了解决这个问题,我们介绍了非凸稀疏惩罚,并通过求解一系列加权$ \ ell_1 $ norm惩罚的子问题来提出一个新的估计器。我们在优化误差和统计误差上建立了非反应优化性能保证,并证明所提出的估计器可以以高概率正确恢复边缘。为了解决每个子问题,我们开发了一种预测的梯度下降算法,该算法具有线性收敛速率。最后,通过施加其他等级约束,提出了学习断开图的扩展。我们根据乘数的交替方向方法提出了一种数值算法,并建立了其理论序列收敛。涉及合成和现实世界数据集的数值实验证明了该方法的有效性。
We consider the problem of learning a sparse graph under the Laplacian constrained Gaussian graphical models. This problem can be formulated as a penalized maximum likelihood estimation of the Laplacian constrained precision matrix. Like in the classical graphical lasso problem, recent works made use of the $\ell_1$-norm regularization with the goal of promoting sparsity in Laplacian constrained precision matrix estimation. However, we find that the widely used $\ell_1$-norm is not effective in imposing a sparse solution in this problem. Through empirical evidence, we observe that the number of nonzero graph weights grows with the increase of the regularization parameter. From a theoretical perspective, we prove that a large regularization parameter will surprisingly lead to a complete graph, i.e., every pair of vertices is connected by an edge. To address this issue, we introduce the nonconvex sparsity penalty, and propose a new estimator by solving a sequence of weighted $\ell_1$-norm penalized sub-problems. We establish the non-asymptotic optimization performance guarantees on both optimization error and statistical error, and prove that the proposed estimator can recover the edges correctly with a high probability. To solve each sub-problem, we develop a projected gradient descent algorithm which enjoys a linear convergence rate. Finally, an extension to learn disconnected graphs is proposed by imposing additional rank constraint. We propose a numerical algorithm based on based on the alternating direction method of multipliers, and establish its theoretical sequence convergence. Numerical experiments involving synthetic and real-world data sets demonstrate the effectiveness of the proposed method.