全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...
-  2017 

基于正则化秩k矩阵逼近的稀疏主成分分析
Sparse principal component analysis via regularized rank-k matrix approximation

DOI: 10.13700/j.bh.1001-5965.2016.0462

Keywords: 降维,稀疏主成分,正则化,块坐标下降法,奇异值分解,阈值
dimension reduction
,sparse principal component,regularization,block coordinate descent method,singular value decomposition,threshold

Full-Text   Cite this paper   Add to My Lib

Abstract:

摘要 在计算稀疏主成分(PCs)时,由于同时求k个主成分的做法可以减少计算所产生的累积误差,因此提出了基于正则化秩k矩阵逼近的稀疏主成分模型,并设计了求解该模型的块坐标下降法(BCD-sPCA-rSVD)。该算法的主要思想是先把变量按坐标分成2k个块,当固定其他2k-1个坐标块的变量时,求解关于单个坐标块的子问题并给出子问题的显式解,循环地求解这些子问题直至满足终止条件。该算法每次迭代的计算复杂度关于样本个数与变量维数都是线性的,并且证明了它是收敛的。该算法不仅易于实现,数值仿真结果表明,该算法应用到真实数据与合成数据上都是可行且有效的。它不仅使累积误差降低,而且具有较低的计算复杂度,因而可以有效地求解大规模稀疏主成分分析问题。
Abstract:In calculating the sparse principal components (PCs), attaining k PCs simultaneously can reduce the accumulated error arising from the calculation process. We proposed the sparse principal component model via regularized rank-k matrix approximation and designed a block coordinate descent method (BCD-sPCA-rSVD) to solve this problem. Its main idea is to first divide variables into 2k blocks by coordinates, and then solve sub-problem with respect to each single coordinate block when keeping other 2k-1 variables fixed. By solving these sub-problems with explicit solutions recursively until the stopping criterion is satisfied, the BCD-sPCA-rSVD algorithm can be easily constructed. Its per-iteration complexity is linear in both sample size and variable dimensionality. The algorithm is convergent and easy to implement. Numerical simulation results show that the algorithm is feasible and effective when applied to real and synthetic data sets. The proposed method reduces the accumulated error and has lower computational complexity, which makes it well suited to handling large-scale problems of sparse principal component analysis.

References

[1]  SIGG C,BUHMANN J.Expectation-maximization for sparse and nonnegative PCA[C]//Proceedings of the 25th International Conference on Machine Learning.NewYork:ACM,2008:960-967.
[2]  JOURNEE M,NESTEROV Y,RICHTARIK P,et al.Generalized power method for sparse principal component analysis[J].Journal of Machine Learning Research,2010,11(2):517-553.
[3]  LU Z S,ZHANG Y.An augmented Lagrangian approach for sparse principal component analysis[J].Math Program Series A,2012,135(1-2):149-193.
[4]  ZHAO Q,MENG D Y,XU Z B,et al.A block coordinates descent approach for sparse principal component analysis[J].Neurocomputing,2015,153(4):180-190.
[5]  TIBSHIRANI R. Regression shrinkage and selection via the LASSO[J].Journal of the Royal Statistical Society Series B,1996,58(3):267-268.
[6]  TSENG P.Convergence of a block coordinates descent method for nondifferentiable minimization[J].Journal of Optimization Theory Apply,2001,109(3):475-494.
[7]  JEFFERS J N R.Two case studies in the application of principal component analysis[J].Applied Statistics,1967,16(3):225-236.
[8]  ALON U,BARKAI N,NOTTERMAN D A,et al.Broad patterns of gene expression revealed by clustering of tumor and normal colon tissues probed by oligonucleotide arrays[J].Proceedings of the National Academy of Sciences of the United States of America,1999,96(12):6745-6750.
[9]  TRENDAFILOV N T,JOLLIFFE I T.Projected gradient approach to the numerical solution of the SCoTLASS[J].Computational Statistics and Data Analysis,2006,50(1):242-253.
[10]  MOGHADDAM B,WEISS Y,AVIDAN S.Spectral bounds for sparse PCA:Exact and greedy algorithms[C]//Advances in Neural Information Processing Systems.Montreal:Neural Information Processing System Foundation,2006:915-922.
[11]  D'ASPREMONT A,BACH F R,GHAOUI L E.Optimal solutions for sparse principal component analysis[J].Machine Learning,2008,9(7):1269-1294.
[12]  LUSS R, TEBOULLE M.Conditional gradient algorithms for rank-one matrix approximations with a sparsity constraint[J].SIAM Review,2013,55(1):65-98.
[13]  LUSS R,TEBOULLE M.Convex approximations to sparse PCA via Lagrangian duality[J].Operations Research Letters,2011,39(1):57-61.
[14]  ZOU H,HASTIE T.Sparse principal component analysis[J].Journal of Computational and Graphical Statistics,2006,15(2):265-286.
[15]  D'ASPREMONT A,GHAOUI L E,JORDAN M I,et al.A direct formulation for sparse PCA using semidefinite programming[J].SIAM Review,2007,48(3):434-448.
[16]  SHEN H,HUANG J Z.Sparse principal component analysis via regularized low rank matrix approximation[J].Journal of Multivariate Analysis,2008,99(6):1015-1034.

Full-Text

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133