南京大学学报(自然科学版) ›› 2018, Vol. 54 ›› Issue (1): 31–.

• • 上一篇    下一篇

 基于L1-范数的鲁棒稀疏的张量PCA人脸图像分析

 唐肝翌*,卢桂馥   

  • 出版日期:2018-01-31 发布日期:2018-01-31
  • 作者简介: 安徽工程大学计算机与信息学院,芜湖,241000
  • 基金资助:
    基金项目:国家自然科学基金(61572033),安徽省高校自然科学研究重大项目(KJ2015ZD08),安徽省高等教育提升计划(TSKJ2015B14)
    收稿日期:2017-12-12
    *通讯联系人,E-mail:tangganhyi.tony@qq.com

 Robust and sparse tensor PCA with L1-norm for faces analysis

 Tang Ganyi*,Lu Guifu   

  • Online:2018-01-31 Published:2018-01-31
  • About author: School of Computer and Information,Anhui Polytechnic University,Wuhu,241000,China

摘要:  张量主成分分析(Tensor Principal Component Analysis,TPCA)是主成分分析(Principal Component Analysis,PCA)在多维空间上的推广,能充分利用图像/视频的空间关联,在图像分析和视频处理中扮演了重要的角色.传统的张量PCA方法提取的特征向量是非稀疏的,这使得其很难进行解释.近年来出现了众多稀疏PCA方法,能提取只包含少量非零元的特征.把稀疏特征提取引入到张量分析,提出一种鲁棒稀疏的张量PCA方法(TPCA-L1S).首先,设计了能实现稀疏特征提取的目标函数.一方面,用L1范数代替Frobenius-范数,使得算法对异常数据更加鲁棒;另一方面,在目标函数中引入弹性网,联合使用Lasso与Ridge惩罚因子来实现稀疏化,增强了算法的语义解释性.然后,设计了一种基于二阶张量的投影矩阵交替求解算法,二阶张量便于数学描述,也易于推广到更高阶张量.此求解算法分为两个步骤(V,U分别表示左投影矩阵和右投影矩阵),先固定U优化V,再固定V的值优化U,两个步骤反复交替执行,直到收敛.每个步骤都采用贪心算法以迭代的方式逐个特征提取以求得U或V.最后,对迭代过程的单调性做了理论证明.基于ORL,Yale和Feret库,将TPCA-L1S应用于人脸图像分析并与其他常见方法作比较,实验结果验证了该模型的有效性.

Abstract:  Tensor Principal Component Analysis(TPCA),which is a generalization of classical PCA(Principal Component Analysis)in multiple dimension space and can make full use of the spatial relationship of images/videos,plays an important role in image analysis and vision processing.The basis vectors of tensor PCA,however,are still dense,which makes it difficult to explain.Various sparse PCA methods,which extract principal components of the given data with sparse non-zero loadings,have been proposed in recent years.In this paper,we extend sparse feature extraction of PCA to tensor analysis and propose a robust and sparse tensor PCA with L1-norm(TPCA-L1S).Firstly,an objective function for TPCA-L1S is designed.On the one hand,we replace Frobenius-norm with L1-norm in the objective function,which makes it more robust to outlier.On the other hand,the elastic net,which generalizes the sparsity-inducing lasso penalty by combining the ridge penalty,is integrated into the objective function to enhance the interpretability of the tensor PCA.Then,an alternative tensor projection optimization procedure is developed in the second-order tensor PCA which is easy to be formulated and can be extended to much higher order.It is convenient to use U and V to denote the projection matrices and refer V and U as to the left and right projection matrices respectively.The procedure is divided into two steps.At the first step, V is computed while U is fixed,and then at the second step, U is computed while V is fixed.The two steps repeat until convergence.A greedy procedure,which is for computing projection matrices U and V,is designed to extract basic features one by one.Finally,the monotonicity of the iterative greedy procedure is theoretically guaranteed.The proposed TPCA-L1S is applied to several face image analysis problems upon ORL,Yale and Feret databases and compared with some other conventional methods.Experimental results confirm its effectiveness.

 [1] Zhao W,Chellappa R,Phillips P J,et al.Face recognition:A literature survey.ACM Computing Surveys(CSUR),2003,35(4):399-458.
[2] Yu Q,Wang R,Yang X J,et al.Diagonal principal component analysis with non-greedy l1-norm maximization for face recognition.Neurocomputing,2016,171:57-62.
[3] Wang H X.Block principal component analysis with L1-norm for image analysis.Pattern Recognition Letters,2012,33(5):537-542.
[4] Jolliffe I T.Principal component analysis.New York:Springer,https://doi.org/10.1007/b98835,2002. 
[5] Yang J,Zhang D,Frangi A F,et al.Two-dimensional PCA:A new approach to appearance-based face representation and recognition.IEEE Transactions on Pattern Analysis and Machine Intelligence,2004,26(1):131-137.
[6] Kwak N.Principal component analysis based on L1-norm maximization.IEEE Transactions on Pattern Analysis and Machine Intelligence,2008,30(9):1672-1680.
[7] Li X L,Pang Y W,Yuan Y.L1-norm-based 2DPCA.IEEE Transactions on Systems,Man,and Cybernetics,Part B(Cybernetics),2010,40(4):1170-1175.
[8] Wang H X,Wang J.2DPCA with L1-norm for simultaneously robust and sparse modelling.Neural Networks,2013,46:190-198.
[9] Kolda T G,Bader B W.Tensor decompositions and applications.SIAM Review,2009,51(3):455-500.
[10] Ye J P.Generalized low rank approximations of matrices.Machine Learning,2005,61(1-3):167-191.
[11] Sun J M,Tao D C,Papadimitriou S,et al.Incremental tensor analysis:Theory and applications.ACM Transactions on Knowledge Discovery From Data(TKDD),2008,2(3):Article No.11.
[12] Pang Y W,Li X L,Yuan Y.Robust tensor analysis with L1-norm.IEEE Transactions on Circuits and Systems for video technology,2010,20(2):172-178.
[13] Zhao L M,Jia W M,Wang R,et al.Robust Tensor analysis with non-greedy l1-norm maximization.Radioengineering,2016,25(1):200-207.
[14] Zou H,Hastie T,Tibshirani R.Sparse principal component analysis.Journal of Computational and Graphical Statistics,2006,15(2):265-286.
[15] Jenatton R,Obozinski G,Bach F.Structured sparse principal component analysis.Journal of Machine Learning Research,2009,9(2):131-160.
No related articles found!
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!