×

Image representation using Laplacian regularized nonnegative tensor factorization. (English) Zbl 1218.68134

Summary: Tensor provides a better representation for image space by avoiding information loss in vectorization. Nonnegative tensor factorization (NTF), whose objective is to express an \(n\)-way tensor as a sum of \(k\) rank-1 tensors under nonnegative constraints, has recently attracted a lot of attentions for its efficient and meaningful representation. However, NTF only sees Euclidean structures in data space and is not optimized for image representation as image space is believed to be a sub-manifold embedded in high-dimensional ambient space. To avoid the limitation of NTF, we propose a novel Laplacian regularized nonnegative tensor factorization (LRNTF) method for image representation and clustering in this paper. In LRNTF, the image space is represented as a 3-way tensor and we explicitly consider the manifold structure of the image space in factorization. That is, two data points that are close to each other in the intrinsic geometry of image space shall also be close to each other under the factorized basis. To evaluate the performance of LRNTF in image representation and clustering, we compare our algorithm with NMF, NTF, NCut and GNMF methods on three standard image databases. Experimental results demonstrate that LRNTF achieves better image clustering performance, while being more insensitive to noise.

MSC:

68T05 Learning and adaptive systems in artificial intelligence
68U10 Computing methodologies for image processing

Software:

BCLS
Full Text: DOI

References:

[1] Shashua, A., On photometric issues in 3D visual recognition from a single 2D image, International Journal of Computer Vision, 21, 1, 99-122 (1997)
[2] Lee, D.; Seung, H., Learning the parts of objects by non-negative matrix factorization, Nature, 401, 6755, 788-791 (1999) · Zbl 1369.68285
[3] Lee, D.; Seung, H., Algorithms for non-negative matrix factorization, Advances in Neural Information Processing Systems, 13, 556-562 (2001)
[4] Li, X.; Pang, Y.; Yuan, Y., L1-norm-based 2DPCA, IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, 40, 4, 1170-1175 (2010)
[5] Tao, D.; Li, X.; Wu, X.; Hu, W.; Maybank, S., Supervised tensor learning, Knowledge and Information Systems, 13, 1, 1-42 (2007)
[6] Welling, M.; Weber, M., Positive tensor factorization, Pattern Recognition Letters, 22, 12, 1255-1261 (2001) · Zbl 0990.68123
[7] Shashua, A.; Hazan, T., Non-negative tensor factorization with applications to statistics and computer vision, (Proceedings of the 22nd International Conference on Machine Learning (2005), ACM), 799
[8] A. Cichocki, R. Zdunek, S. Choi, R. Plemmons, S. Amari, Non-negative tensor factorization using alpha and beta divergences, in: International Conference on Acoustics, Speech and Signal Processing (ICASSP 07), Honolulu, Hawaii, USA, 2007.; A. Cichocki, R. Zdunek, S. Choi, R. Plemmons, S. Amari, Non-negative tensor factorization using alpha and beta divergences, in: International Conference on Acoustics, Speech and Signal Processing (ICASSP 07), Honolulu, Hawaii, USA, 2007.
[9] M. Mørup, L. Hansen, J. Parnas, S. Arnfred, Decomposing the time-frequency representation of EEG using nonnegative matrix and multi-way factorization, Technical University of Denmark, Informatics and Mathematical Modelling, Copenhagen, Denmark.; M. Mørup, L. Hansen, J. Parnas, S. Arnfred, Decomposing the time-frequency representation of EEG using nonnegative matrix and multi-way factorization, Technical University of Denmark, Informatics and Mathematical Modelling, Copenhagen, Denmark.
[10] B. Bader, M. Berry, M. Browne, Discussion tracking in Enron email using PARAFAC, Survey of Text Mining II: Clustering, Classification, and Retrieval, second ed., 2007, pp. 147-163.; B. Bader, M. Berry, M. Browne, Discussion tracking in Enron email using PARAFAC, Survey of Text Mining II: Clustering, Classification, and Retrieval, second ed., 2007, pp. 147-163.
[11] T. Hazan, S. Polak, A. Shashua, Sparse image coding using a 3D non-negative tensor factorization, in: International Conference of Computer Vision (ICCV), Citeseer, 2005, pp. 50-57.; T. Hazan, S. Polak, A. Shashua, Sparse image coding using a 3D non-negative tensor factorization, in: International Conference of Computer Vision (ICCV), Citeseer, 2005, pp. 50-57.
[12] Friedlander, M.; Hatz, K., Computing non-negative tensor factorizations, Optimization Methods & Software, 23, 4, 631-647 (2008) · Zbl 1158.65321
[13] Kolda, T.; Bader, B., Tensor decompositions and applications, SIAM Review, 51, 3, 455-500 (2009) · Zbl 1173.65029
[14] Cichocki, A.; Zdunek, R.; Amari, S., Nonnegative matrix and tensor factorization, IEEE Signal Processing Magazine, 25, 1, 142 (2008)
[15] He, X., Incremental semi-supervised subspace learning for image retrieval, (Proceedings of the 12th Annual ACM International Conference on Multimedia (2004), ACM: ACM New York, NY, USA), 2-8
[16] He, X.; Cai, D.; Han, J., Learning a maximum margin subspace for image retrieval, IEEE Transactions on Knowledge and Data Engineering, 20, 2, 189-201 (2008)
[17] He, X.; Ma, W.; Zhang, H., Learning an image manifold for retrieval, (Proceedings of the 12th Annual ACM International Conference on Multimedia (2004), ACM: ACM New York, NY, USA), 17-23
[18] Lin, Y.; Liu, T.; Chen, H., Semantic manifold learning for image retrieval, (Proceedings of the 13th Annual ACM International Conference on Multimedia (2005), ACM: ACM New York, NY, USA), 258
[19] Wang, C.; Zhao, J.; He, X.; Chen, C.; Bu, J., Image retrieval using nonlinear manifold embedding, Neurocomputing, 72, 16-18, 3922-3929 (2009)
[20] Ding, C.; Li, T., Adaptive dimension reduction using discriminant analysis and k-means clustering, (Proceedings of the 24th International Conference on Machine Learning (2007), ACM), 521-528
[21] Cai, D.; He, X.; Wu, X.; Han, J., Non-negative matrix factorization on manifold, (Proceedings of the International Conference on Data Mining (ICDM2008) (2008))
[22] Cai, D.; He, X.; Wang, X.; Bao, H.; Han, J., Locality preserving nonnegative matrix factorization, (Proceedings of the 21st International Joint Conference on Artificial Intelligence (2009), Morgan Kaufmann Publishers Inc.), 1010-1015
[23] Gu, Q.; Zhou, J., Local learning regularized nonnegative matrix factorization, (Proceedings of the 21st International Joint Conference on Artificial Intelligence (2009), Morgan Kaufmann Publishers Inc.), 1046-1051
[24] Wachsmuth, E.; Oram, M.; Perrett, D., Recognition of objects and their component parts: responses of single units in the temporal cortex of the macaque, Cerebral Cortex, 4, 5, 509 (1994)
[25] Logothetis, N.; Sheinberg, D., Visual object recognition, Annual Review of Neuroscience, 19, 577-621 (1996)
[26] S. Li, X. Hou, H. Zhang, Q. Cheng, Learning spatially localized, parts-based representation, in: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1, IEEE Computer Society, 1999, 2001.; S. Li, X. Hou, H. Zhang, Q. Cheng, Learning spatially localized, parts-based representation, in: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1, IEEE Computer Society, 1999, 2001.
[27] Xu, W.; Liu, X.; Gong, Y., Document clustering based on non-negative matrix factorization, (Proceedings of the 26th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (2003), ACM: ACM New York, NY, USA), 267-273
[28] Brunet, J.; Tamayo, P.; Golub, T.; Mesirov, J., Metagenes and molecular pattern discovery using matrix factorization, Proceedings of the National Academy of Sciences of the United States of America, 101, 12, 4164 (2004)
[29] Shashua, A.; Levin, A., Linear image coding for regression and classification using the tensor-rank principle, (Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, vol. 1 (2001)) · Zbl 1039.68670
[30] Tucker, L., The extension of factor analysis to three-dimensional matrices, Contributions to Mathematical Psychology, 109-127 (1964)
[31] Carroll, J.; Chang, J., Analysis of individual differences in multidimensional scaling via an N-way generalization of Eckart-Young decomposition, Psychometrika, 35, 3, 283-319 (1970) · Zbl 0202.19101
[32] Harshman, R., Foundations of the PARAFAC procedure: models and conditions for an “explanatory” multi-modal factor analysis, UCLA Working Papers in Phonetics, 16, 1, 84 (1970)
[33] Tenenbaum, J.; Silva, V.; Langford, J., A global geometric framework for nonlinear dimensionality reduction, Science, 290, 5500, 2319 (2000)
[34] Roweis, S.; Saul, L., Nonlinear dimensionality reduction by locally linear embedding, Science, 290, 5500, 2323 (2000)
[35] Belkin, M.; Niyogi, P., Laplacian eigenmaps and spectral techniques for embedding and clustering, Advances in Neural Information Processing Systems, 14 1, 585-592 (2002)
[36] Martínez, A.; Kak, A., Pca versus lda, IEEE Transactions on Pattern Analysis and Machine Intelligence, 23, 2, 228-233 (2002)
[37] Lovász, L.; Plummer, M., Matching Theory (2009), Chelsea Pub Co. · Zbl 1175.05002
[38] Shi, J.; Malik, J., Normalized cuts and image segmentation, IEEE Transactions on Pattern Analysis and Machine Intelligence, 22, 8, 888-905 (2000)
[39] He, X.; Cai, D.; Niyogi, P., Tensor subspace analysis, Advances in Neural Information Processing Systems, 18, 499 (2006)
[40] Wu, M.; Schölkopf, B., A local learning approach for clustering, Advances Neural Information Processing Systems, 18, 1529-1536 (2006)
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. In some cases that data have been complemented/enhanced by data from zbMATH Open. This attempts to reflect the references listed in the original paper as accurately as possible without claiming completeness or a perfect matching.