南京大学学报(自然科学版) ›› 2015, Vol. 51 ›› Issue (1): 139–147.

• • 上一篇    下一篇

一种基于正则化最小二乘的多标记分类算法

吕静,何志芬   

  • 出版日期:2015-01-04 发布日期:2015-01-04
  • 作者简介:(1. 南京师范大学计算机科学与技术学院南京210023; 2. 南京,南京,2100
  • 基金资助:
     江苏省自然科学基金重点重大专项(BK2011005)

A multi-label classification algorithm based on Regularized Least Squares

Lv Jing1*, He Zhifen1,2   

  • Online:2015-01-04 Published:2015-01-04
  • About author:(1.School of Computer Science and Technology, Nanjing Normal University, Nanjing, 210023, China; 2. School?of?Mathematical?Sciences, Nanjing Normal University, Nanjing, 210023, China)

摘要: 在传统的监督学习中,每个对象由单个实例表示且只属于一个类别标记。然而,在多标记学习中,每个对象由一个实例表示但可能同时属于多个类别标记,其任务是预测未知样本的类别标记集合。本文提出了基于正则化最小二乘的多标记分类算法,即将传统的正则化最小二乘分类推广到多标记学习中。首先,将多标记学习问题转化为多个独立的二分类问题(每个对应一个类别标记);其次,为了充分利用类别标记之间的相关信息,构建了基于类别标记的邻接图, 其中每个节点代表一个类别标记,每条边的权重反映了相应类别标记对之间的相似性。最后,构建了建立在核函数基础上的多标记正则化最小二乘模型,并可以转化为求解一个Sylvester方程。在8个基准数据集上用5种不同的评价准则进行度量的实验结果表明了本文算法优于其他6种最新的多标记分类算法。

Abstract: In traditional supervised learning, each object is represented by a single instance and associated with only one class label. Recently, multi-label learning attracts much attention from researchers of machine learning, pattern recognition etc., due to its ability in obtain more information on predicting multiple class labels. In multi-label learning, each object is represented by an instance, while it may be assigned to multiple class labels simultaneously. Hence, its task is to predict a class label set for the unknown instance. The resulting model is usually an ill-conditioned quadratic program, which requires regularization. In this paper, a multi-label classification algorithm based on regularized least squares classification is proposed, which is derived from the traditional regularized least squares classification. To establish our model, we first transform the multi-label learning problem into a multiple independent binary classification problem (each for one class label). Then, in order to fully exploit the class label correlations information, an adjacent graph based on all the class labels is built, where each node stands for one class label and the weight of each edge reflects the similarity between corresponding pairwise-label. Finally, a multi-label regularized least squares model is constructed on the basis of kernel function, whose first order optimality condition is a Sylvester equation, which can be solved efficiently by using numerical linear algebra technique. We perform experiment on eight benchmark data sets in terms of five different evaluation criteria and compare it with the other six state-of-the-art multi-label learning algorithms. The results show that our algorithm is competitive.

[1] Schapire R E, Singer Y. Boostexter: A boosting-based system for text categorization. Machine Learning, 2000, 39(2):135~168.
[2] Zhang M L, Zhou Z H. A review on multi-label learning algorithms. IEEE Transactions on Knowledge and Data Engineering, 2014, 26(8):1819~1837.
[3] Boutell M R, Luo J, Shen X, et al. Learning multi-label scene classification. Pattern recognition, 2004, 37(9):1757~1771.
[4] Zhang M L, Zhou Z H. Multilabel neural networks with applications to functional genomics and text categorization. IEEE Transactions on Knowledge and Data Engineering, 2006, 18(10):1338~1351.
[5] Elisseeff A, Weston J. A kernel method for multi-labelled classification. In: Advances in Neural Information Processing Systems. Cambridge, MA: MIT Press, 2001: 681~687.
[6] Lo H Y, Wang J C, Wang H M, et al. Cost-sensitive multi-label learning for audio tag annotation and retrieval. IEEE Transactions on Multimedia, 2011, 13(3):518~529.
[7] Huang S J, Zhou Z H. Multi-label learning by exploiting label correlations locally. In: Proceedings of the 26th AAAI Conference on Artificial Intelligence. AAAI Press, 2012: 949~955.
[8] Huang S J, Yu Y, Zhou Z H. Multi-label hypothesis reuse. In: Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. NY: ACM, 2012: 525~533.
[9] Tsoumakas G, Katakis I, Vlahavas I. Random k-labelsets for multilabel classification. IEEE Transactions on Knowledge and Data Engineering, 2011, 23(7): 1079~1089.
[10] Hüllermeier E, Fürnkranz J, Cheng W, et al. Label ranking by learning pairwise preferences. Artificial Intelligence, 2008, 172(16): 1897~1916.
[11] Fürnkranz J, Hüllermeier E, Mencía E L, et al. Multilabel classification via calibrated label ranking. Machine Learning, 2008, 73(2): 133~153.
[12] 郑 伟, 王朝坤, 刘 璋等. 一种基于随机游走模型的多标签分类算法. 计算机学报, 2010, 33(8): 1418~1426.
[13] Zhang M L, Zhou Z H. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognition, 2007, 40(7): 2038~2048.
[14] 张敏灵. 一种新型多标记懒惰学习算法. 计算机研究与发展, 2012, 49(11): 2271~2282.
[15] Rifkin R, Yeo G, Poggio T. Regularized least-squares classification. Nato Science Series Sub Series III Computer and Systems Sciences, 2003, 190: 131~154.
[16] Schölkopf B, Herbrich R, Smola A J. A generalized representer theorem. Computational Learning Theory. Springer Berlin Heidelberg, 2001: 416~426.
[17] Belkin M, Niyogi P, Sindhwani V. Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. The Journal of Machine Learning Research, 2006, 7: 2399~2434.
[18] Chen G, Song Y. Semi-supervised multi-label learning by solving a Sylvester equation. In: Jonker W, Petkovi? M. SDM 2008. LNCS, Springer, Heidelberg, 2008: 5159.
[19] Hu D Y, Reichel L. Krylov-subspace methods for the Sylvester equation. Linear Algebra and Its Applications, 1992, 172: 283~313.
[20] Read J, Pfahringer B, Holmes G, et al. Classifier chains for multi-label classification. Machine Learning, 2011, 85(3): 333~359.
[21] Tsoumakas G, Xioufis E S, Vilcek J, et al. MULAN: A java library for multi-label learning. Journal of Machine Learning Research, 2011, 12(7): 2411~2414.
No related articles found!
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!