×

Self-organizing maps on “what-where” codes towards fully unsupervised classification. (English) Zbl 1521.92014

Summary: Interest in unsupervised learning architectures has been rising. Besides being biologically unnatural, it is costly to depend on large labeled data sets to get a well-performing classification system. Therefore, both the deep learning community and the more biologically-inspired models community have focused on proposing unsupervised techniques that can produce adequate hidden representations which can then be fed to a simpler supervised classifier. Despite great success with this approach, an ultimate dependence on a supervised model remains, which forces the number of classes to be known beforehand, and makes the system depend on labels to extract concepts. To overcome this limitation, recent work has been proposed that shows how a self-organizing map (SOM) can be used as a completely unsupervised classifier. However, to achieve success it required deep learning techniques to generate high quality embeddings. The purpose of this work is to show that we can use our previously proposed what-where encoder in tandem with the SOM to get an end-to-end unsupervised system that is Hebbian. Such system, requires no labels to train nor does it require knowledge of which classes exist beforehand. It can be trained online and adapt to new classes that may emerge. As in the original work, we use the MNIST data set to run an experimental analysis and verify that the system achieves similar accuracies to the best ones reported thus far. Furthermore, we extend the analysis to the more difficult fashion-MNIST problem and conclude that the system still performs.

MSC:

92B20 Neural networks for/in biological studies, artificial life and related topics

References:

[1] Bishop CM (2006) Pattern recognition and machine learning. Springer, Berlin. http://www.library.wisc.edu/selectedtocs/bg0137.pdf · Zbl 1107.68072
[2] Cardoso, Â.; Wichert, A., Neocognitron and the map transformation cascade, Neural Netw, 23, 74-88 (2010) · doi:10.1016/j.neunet.2009.09.004
[3] Goodfellow, I.; Bengio, Y.; Courville, A., Deep learning (2016), Cambridge: MIT press, Cambridge · Zbl 1373.68009
[4] Harkness, L.; Bennet-Clark, H., The deep fovea as a focus indicator, Nature, 272, 5656, 814-816 (1978) · doi:10.1038/272814a0
[5] Haykin, S., Neural networks and learning machines (2008), London: Pearson, London
[6] Hertz, J.; Krogh, A.; Palmer, RG, Introduction to the theory of neural computation (1991), Boca Raton: CRC Press, Boca Raton
[7] Illing, B.; Gerstner, W.; Brea, J., Biologically plausible deep learning-but how far can we go with shallow networks?, Neural Netw, 118, 90-101 (2019) · doi:10.1016/j.neunet.2019.06.001
[8] Khacef L, Miramond B, Barrientos D, Upegui A (2019) Self-organizing neurons: toward brain-inspired unsupervised learning. In: International Joint Conference on Neural Networks (IJCNN), pp 1-9. doi:10.1109/IJCNN.2019.8852098. IEEE
[9] Khacef L, Rodriguez L, Miramond B (2020) Improving self-organizing maps with unsupervised feature extraction. In: International Conference on Neural Information Processing (ICONIP), pp 474-486. doi:10.1007/978-3-030-63833-7_40
[10] Kohonen, T., Self-organization and associative memory (1984), Berlin: Springer, Berlin · Zbl 0528.68062
[11] Kohonen, T., The self-organizing map, Proc IEEE, 78, 9, 1464-1480 (1990) · doi:10.1109/5.58325
[12] Krotov, D.; Hopfield, JJ, Unsupervised learning by competing hidden units, Proc Natl Acad Sci, 116, 16, 7723-7731 (2019) · Zbl 1433.68359 · doi:10.1073/pnas.1820458116
[13] LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. In: Proceedings of the IEEE, vol 86, pp 2278-2324. doi:10.1109/5.726791
[14] LeCun Y, Cortes C, Burges C. MNIST handwritten digit database. http://yann.lecun.com/exdb/mnist/ Accessed 12 Apr 2020
[15] Le-Khac PH, Healy G, Smeaton AF (2020) Contrastive representation learning: a framework and review. IEEE Access 8:193907-193934. doi:10.1109/ACCESS.2020.3031549
[16] Liversedge, SP; Findlay, JM, Saccadic eye movements and cognition, Trends Cogn Sci, 4, 1, 6-14 (2000) · doi:10.1016/S1364-6613(99)01418-7
[17] Lloyd S (1982) Least squares quantization in pcm. IEEE Trans Inf Theory 28(2):129-137. doi:10.1109/TIT.1982.1056489 · Zbl 0504.94015
[18] Marr, D., Vision: a computational investigation into the human representation and processing of visual information (1982), Cambridge: MIT press, Cambridge
[19] Melnykov, I.; Melnykov, V., On k-means algorithm with the use of mahalanobis distances, Stat Probab Lett, 84, 88-95 (2014) · Zbl 1284.62383 · doi:10.1016/j.spl.2013.09.026
[20] Murphy, KP, Machine learning: a probabilistic perspective (2012), Cambridge: MIT press, Cambridge · Zbl 1295.68003
[21] Ravichandran NB, Lansner A, Herman P (2020) Learning representations in bayesian confidence propagation neural networks. In: International Joint Conference on Neural Networks (IJCNN), pp 1-7. doi:10.1109/IJCNN48605.2020.9207061
[22] Rumelhart, DE; Zipser, D., Feature discovery by competitive learning, Cogn Sci, 9, 75-112 (1985) · doi:10.1016/S0364-0213(85)80010-0
[23] Sa-Couto, L.; Wichert, A., Attention inspired network: steep learning curve in an invariant pattern recognition model, Neural Netw, 114, 38-46 (2019) · doi:10.1016/j.neunet.2019.01.018
[24] Sa-Couto, L.; Wichert, A., Storing object-dependent sparse codes in a willshaw associative network, Neural Comput, 32, 136-152 (2020) · Zbl 1473.68163 · doi:10.1162/neco_a_01243
[25] Sa-Couto, L.; Wichert, A., Simple convolutional-based models: Are they learning the task or the data?, Neural Comput, 33, 12, 3334-3350 (2021) · Zbl 1482.92024 · doi:10.1162/neco_a_01446
[26] Sa-Couto, L.; Wichert, A., “what-where” sparse distributed invariant representations of visual patterns, Neural Comput Appl, 34, 8, 6207-6214 (2022) · doi:10.1007/s00521-021-06759-0
[27] Sa-Couto, L.; Wichert, A., Using brain inspired principles to unsupervisedly learn good representations for visual pattern recognition, Neurocomputing, 495, 97-104 (2022) · doi:10.1016/j.neucom.2022.04.130
[28] Sculley D (2010) Web-scale k-means clustering. In: 19th International Conference on World Wide Web, pp. 1177-1178. doi:10.1145/1772690.1772862
[29] Trappenberg T (2009) Fundamentals of computational neuroscience. OUP, Oxford · Zbl 1179.92012
[30] Xiao H, Rasul K, Vollgraf R (2017) Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. In some cases that data have been complemented/enhanced by data from zbMATH Open. This attempts to reflect the references listed in the original paper as accurately as possible without claiming completeness or a perfect matching.