南京大学学报(自然科学版) ›› 2023, Vol. 59 ›› Issue (2): 263272.doi: 10.13232/j.cnki.jnju.2023.02.009
Yu Song, Yuzhu Xiao(), Xueli Song()
摘要:
无监督特征选择是无标签高维数据预处理过程中一种有效的数据降维技术,然而大多数无监督特征选择算法忽略了数据样本本身的类簇结构特性,选择具有低判别性信息的特征.基于此,提出一种基于伪标签回归和流形正则化的无监督特征选择算法.具体地,联合伪标签回归和最大化类间散度来保证算法在迭代过程中学习伪标签,同时,自适应学习数据样本之间的局部几何结构,获得更加精准的标签信息和结构信息,进而选择具有高判别性且能保持数据流形结构的特征.在四个公开数据集上的对比实验表明,提出算法的特征选择结果优于现有的一些无监督特征选择算法.
中图分类号:
1 | 蒋胜利. 高维数据的特征选择与特征提取研究. 博士学位论文. 西安:西安电子科技大学,2011. |
Jiang S L. Research on feature selection and feature extraction of high?dimensional data. Ph.D. Dissertation. Xi'an:Xidian University,2011. | |
2 | 林书亮. 联合L2,1范数正则约束的特征选择方法. 科技与企业,2013(24):383-384. |
3 | Liang S Q, Xu Q, Zhu P F,et al. Unsupervised feature selection by manifold regularized self?represen?tation∥Proceedings of 2017 IEEE International Conference on Image Processing. Beijing,China:IEEE,2017:2398-2402. |
4 | 方威. 自适应图正则非负矩阵分解聚类算法的研究. 硕士学位论文. 扬州:扬州大学,2021. |
Fang W. Research on clustering algorithm of adaptive graph regularized non?negative matrix factorization. Master Dissertation. Yangzhou:Yangzhou University,2021. | |
5 | 章永来,周耀鉴. 聚类算法综述. 计算机应用,2019,39(7):1869-1882. |
Zhang Y L, Zhou Y J. Review of clustering algorithms. Journal of Computer Applications,2019,39(7):1869-1882. | |
6 | 杜世强. 基于维数约简的无监督聚类算法研究. 博士学位论文. 兰州:兰州大学,2017. |
Du S Q. Unsupervised clustering algorithm based on dimension reduction. Ph.D. Dissertation. Lanzhou:Lanzhou University,2017. | |
7 | 汪志远. 无监督特征选择方法研究. 硕士学位论文. 太原:太原理工大学,2020. |
Wang Z Y. Research on unsupervised feature selection. Master Dissertation. Taiyuan:Taiyuan University of Technology,2020. | |
8 | Li Z C, Yang Y, Liu J,et al. Unsupervised feature selection using nonnegative spectral analysis∥Proceedings of the 26th AAAI Conference on Artificial Intelligence. Toronto,Canada:AAAI,2012,26(1):1026-1032. |
9 | Yang Y, Shen H T, Ma Z G,et al . L 2,1?norm regu?larized discriminative feature selection for unsupervised learning∥Proceedings of the 22nd International Joint Conference on Artificial Intelligence. Barcelona,Spain:AAAI,2011:1589-1594. |
10 | Cai D, Zhang C Y, He X F. Unsupervised feature selection for multi?cluster data∥Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Washington,DC,USA:ACM,2010:333-342. |
11 | Hou C P, Nie F P, Yi D Y,et al. Feature selection via joint embedding learning and sparse regression∥Proceedings of the 22nd International Joint Conference on Artificial Intelligence. Barcelona,Spain:AAAI,2011:1324-1329. |
12 | Tang C, Liu X W, Li M M,et al. Robust unsupervised feature selection via dual self?representation and manifold regularization. Knowledge?Based Systems,2018(145):109-120. |
13 | Du L, Shen Y D. Unsupervised feature selection with adaptive structure learning∥Proceedings of the 21th ACM SIGKDD International Conference on Know?ledge Discovery and Data Mining. Sydney,Australia:ACM,2015:209-218. |
14 | Zhang R, Zhang Y X, Li X L. Unsupervised feature selection via adaptive graph learning and constraint. IEEE Transactions on Neural Networks and Learning Systems,2022,33(3):1355-1362. |
15 | Zhu P F, Zhu W C, Hu Q H,et al. Subspace clus?tering guided unsupervised feature selection. Pattern Recognition,2017(66):364-374. |
16 | 盛超,宋鹏,郑文明,等. 基于子空间学习和伪标签回归的无监督特征选择. 信号处理,2021,37(9):1701-1708. |
Sheng C, Song P, Zheng W M,et al. Subspace learning and virtual label regression based unsupervised feature selection. Journal of Signal Processing,2021,37(9):1701-1708. | |
17 | 周志华. 机器学习. 北京:清华大学出版社,2016:1-425. |
18 | Liu X W, Wang L, Zhang J,et al. Global and local structure preservation for feature selection. IEEE Transactions on Neural Networks and Learning Systems,2014,25(6):1083-1095. |
19 | Nie F P, Huang H, Cai X,et al. Efficient and robust feature selection via joint L2,1?norms minimization∥Proceedings of the 23rd International Conference on Neural Information Processing Systems. Vancouver,Canada:Curran Associates Inc.,2010:1813-1821. |
20 | 丛思安,王星星. K?means算法研究综述. 电子技术与软件工程,2018(17):155-156. |
[1] | 时照群, 刘兆伟, 刘惊雷. 基于相关熵和流形正则化的图像聚类[J]. 南京大学学报(自然科学版), 2022, 58(3): 469-482. |
|