key: cord-1016227-xmmyqbgc authors: Vellido, Alfredo; Angulo, Cecilio; Gibert, Karina title: A self-organizing world: special issue of the 13th edition of the workshop on self-organizing maps and learning vector quantization, clustering and data visualization, WSOM + 2019 date: 2021-07-19 journal: Neural Comput Appl DOI: 10.1007/s00521-021-06307-w sha: 5718ca2a38f7e1e7954ec71236e3d363820db489 doc_id: 1016227 cord_uid: xmmyqbgc nan Padova, Italy (Self-organizing Maps for Structured Data: from Unsupervised to Deep Supervised Models). The variety of these talks reflected the own variety of the advances presented by the papers accepted in the conference. The conference was structured in nine sessions on seven different topics: SOM theory, methods and practical applications covered three sessions; correspondingly, LVQ theory, methods and practical applications covered two sessions; then, two sessions were devoted to general theory and methods and general practical applications and, finally, two sessions to specific applications to the life sciences. The latter bear witness to the rapidly increasing interest in machine learning in general and unsupervised machine learning in particular shown by biology and medicine, mostly driven by developments in the omics sciences and the area of systems biology. A total of thirty-three papers were accepted for oral presentation. Out of these, ten were selected for extension and these are the ones that we now proudly present to the Neural Computing & Applications readership in this special issue. The four first papers of this issue are devoted to SOM models. The first of them [1] stems from one of the invited talks and merges self-organization and deep learning. Pasa, Navarin and Sperduti propose a SOM-based aggregation approach for Graph Convolutional Neural Networks. Deep learning of graphs is a hot topic due to the increasing availability of data that are naturally represented in a graph form (life sciences, social networks, etc.). A main component of a graph neural network is the aggregation operator required to generate a graph-level representation from a set of node-level embeddings. The proposed SOM-based DeepSets aggregation operator allows to compute node representations that embed information about their mutual similarity. The second SOM-related paper [2] also addresses graphs, but from a very different standpoint. Taylor and Merenyi present DM-Prune: not a clustering method as such, but as a sparsification step for cumulative adjacency (CADJ) graphs as inputs to graph-based clustering (GBC) methods for partitioning the learned prototypes of a vector quantizer. Even if CADJ graphs may originate from the prototypes of any vector quantizer, authors focus on prototypes obtained using SOM. The method is illustrated with GBC of images. The third paper [3] by Ma, Kirby and Peterson takes a more theoretical road and describes the mathematical framework for implementing SOM mappings on flag manifolds. A flag is described as a nested sequence of vector spaces and a flag manifold as a manifold whose points parameterize all flags of a fixed type in a fixed vector space. Flags are argued to arise in data analysis contexts such as wavelet, Fourier, and singular value decompositions. Finally, in [4] , Bacciu, Bertoncini and Morelli address an eminently practical problem: quality inspection and intelligent filtering of smart-bracelet data in the context of preventive analysis of heart activity sensors, plagued by noise issues introduced by motion artifacts. SOM models are shown to be effective for filtering sections of heart data collected from unreliable devices, such as wrist-worn devices. The next four papers of the WSOM ? 2019 special issue deal with LVQ models, again from different perspectives. The first [5] draws us towards a current -at the time of publication-affair: the COVID-19 pandemic. Kaden and co-workers investigate SARS-CoV-2 virus sequences based on alignment-free dissimilarity measures for RNA sequence comparison leading to virus type classification using an LVQ variant. The second LVQ-related paper [6] proposes a quantum-inspired LVQ for prototypebased classification. In their paper, Villmann and coworkers are interested in interpretable variants such as Generalized LVQ (GLVQ) and propose a quantum computing-inspired realization of this model. Kernelized GLVQ methods are described and linked to quantum space ideas using quantum bit vector space in the quantum state space. The third paper [7] takes the reader to another timely topic: concept drift handling, with a focus on streaming data. Heusinger, Raab and Schleif modify Soft LVQ and GLVQ to handle streaming data. Here, momentum-based stochastic gradient descent techniques are applied to tackle concept drift passively. The fourth and last paper of this batch [8] also addresses the problem of concept drift. Straat and co-workers apply methods from statistical physics to develop a modeling framework for the mathematical analysis of the training dynamics in non-stationary environments, showing that standard LVQ algorithms are, to some extent, suitable for the training in non-stationary environments without modification. Authors go beyond LVQ to investigate gradient-based training of layered neural networks with sigmoidal activation functions, comparing them with the use of rectified linear units (ReLU) , showing that the sensitivity to concept drift differs significantly between these types of activation function. The remaining two papers in the issue tackle quite different problems. The work by Karna and Gibert [9] provides solutions for the classical problem of determining the most adequate number of clusters in hierarchical clustering settings. Usually, this determination from the data partition dendrogram requires human expertise as input, especially in real application settings like the one which is the focus of the paper: decision support in an industry 4.0 environment, namely 3D printers based on sensor data in real-time. The final paper [10] takes us back to the problem of concept drift. Here, Szadkowski addresses, in a combined manner, two challenges for life-long classifiers: concept drift and catastrophic forgetting. For that, the process of continual learning is defined as a wrap modification, where a wrap is a manifold that can be trained to cover or uncover a given set of samples and is used as a theoretical building block for a general life-long learning scheme implemented as an ensemble of variational autoencoders. The chairpersons of WSOM ? 2019, as editors and curators of this special issue, hope that you will enjoy reading the collected contributions and find them useful. We encourage you to submit your work to, and follow the future editions of, our WSOM ? conference. We express our gratitude to the editor-in-chief for his invitation to produce this issue of the journal. We also thank the editorial staff for their kind help on its elaboration. SOM-based aggregation for graph convolutional neural networks DM-pruning CADJ graphs for SOM clustering Self-organizing mappings on the flag manifold with applications to hyper-spectral image data analysis Topographic mapping for quality inspection and intelligent filtering of smart-bracelet data Learning vector quantization as an interpretable classifier for the detection of SARS-CoV-2 types based on their RNA sequences Quantum-inspired learning vector quantizers for prototype-based classification Passive concept drift handling via variations of learning vector quantization Supervised learning in the presence of concept drift: a modelling framework Automatic identification of the number of clusters in hierarchical clustering Continually trained lifelong classification