@inproceedings{oai:kyutech.repo.nii.ac.jp:00001536, author = {Furukawa, Tetsuo and 古川, 徹生 and Tokunaga, Kazuhiro and Morishita, Kenji and Yasui, Syozo}, book = {Proceedings of International Joint Conference on Neural Networks}, month = {Aug}, note = {Kohonen's Self-Organizing Map (SOM), which performs topology-preserving transformation from a highdimensional data vector space to a low-dimensional map space, provides a powerful tool for data analysis, classification and visualization in many application fields. Despite its powerfulness, SOM can only deal with vectorized data, although many expansions have been proposed for various data-type cases. This study aims to develop a novel generalization of SOM called modular Iletwork SOM (mIlSOM), which enables users to deal with general data classes in a consistent manner. mnSOM has an array structure consisting of function modules that are trainable neural networks, e.g. multi-layer perceptrons (MLPs), instead of the vector units of the conventional SOM family. In the case of MLPmodules, mnSOM learns a group of systems or functions in terms of the input-output relationships, and at the same time mnSOM generates a feature map that shows distances between the learned systems. Thus, mnSOM with MLP modules is an SOM in function space rather than in vector space. From this point of view, the conventional SOM of Kohonen's can be regarded as a special case of mnSOM, the modules consisting of fixed-value bias units. In this paper, mnSOM with MLP modules is described along with some application examples., 2005 IEEE International Joint Conference on Neural Networks, 31 July - 4 August, 2005, Montreal, Québec, Canada}, pages = {1581--1586}, publisher = {IEEE}, title = {Modular Network SOM (mnSOM): From Vector Space to Function Space}, volume = {3}, year = {2005}, yomi = {フルカワ, テツオ} }