1 / 103

The Self-Organizing Map and Applications

The Self-Organizing Map and Applications. Jennie Si, Ph.D. Professor Dept. of Electrical Engineering Center for Systems Science and Engineering Research Arizona State University (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu (email). Structure of the presentation Background

harlant
Télécharger la présentation

The Self-Organizing Map and Applications

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. The Self-Organizing Map and Applications Jennie Si, Ph.D. Professor Dept. of Electrical Engineering Center for Systems Science and Engineering Research Arizona State University (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu (email) (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  2. Structure of the presentation • Background • The algorithm • Analysis • Case 1 – motor data analysis • Case 2 – auditory data analysis • Case 3 – supplier change control • A qualitative comparison • Advanced issue – topology preserving • Conclusions (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  3. References J. Si, S. Lin, and M. A. Vuong, “Dynamic topology representing network”, Neural Networks, the Official Journal of International Neural Network Society. 13 (6): 617-627, 2000. S. Lin, and J. Si, “Weight convergence and weight density distribution of the SOFM network with discrete input”, Neural Computation 10 (4): 807-814, 1998. S. Lin, and J. Si, and A. B. Schwartz, "Self-Organization of Firing Activities in Monkey's Motor Cortex: Trajectory Computation from Spike Signals.” Neural Computation, the MIT Press, March 1997, pp. 607-621. R. Davis, Industry data mining using the self-organizing map. M.S. thesis. Arizona State University, May 2001. (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  4. Evolution of the SOM algorithm • Von der Masburg, 1970’s, the self-organization of orientation sensitive nerve cells in the striate cortex • Willshaw and von der Masburg, 1976, the first paper on the formation of self-organizing maps on biological grounds to explain retinotopic mapping from the retina to the visual cortex (in higher vertebrates). • Kohonen, 1982, the paper on the self-organizing map, “Self-organized formation of topologically correct featured maps”, in Biological cybernetics (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  5. SOM - a computational shortcut • To mimic basic functions similar to biological neural networks • Implementation details of biological systems “ignored” • To create an “Ordered” map of input signals • Internal structure of the input signals themselves • Coordination of the unit activities through the lateral connections between the units • A statistical data modeling tool (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  6. Two distinct properties of SOM • Clustering of multidimensional input data • Spatially ordering the output map so that similar input patterns tend to produce a response in units that are close to each other in the output map. • Topology preserving • nodes in the output layer represent clustering information from the input data (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  7. Applications of SOM • Speech processing • Vector quantization • Image coding • Biological signal analysis • Visualize results from multi-dimensional data analysis. • And many many more… (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  8. Structure of the presentation • Background • The algorithm • Analysis • Case 1 – motor data analysis • Case 2 – auditory data analysis • Case 3 – supplier change control • A qualitative comparison • Advanced issue – topology preserving • Conclusions (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  9. A finite set of nodes Undirected edges adjacent to , if The output map - undirected graph (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  10. i c The self-organizing map Output graph G Weight vector W Input vector (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  11. Learning the topological mapping from input • The graph (output map) is usually pre-specified as a one-dimensional chain or two dimensional lattice. • SOM intends to learn the topological mapping by means of self-organization driven by samples X • X is assumed to be connected in parallel to every node in the output map. • A node in G, associated with a weight vector W, can be represented by its index i or its position (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  12. SOM building block • Find the winner and the neighborhood of the winner • comparing the inner products WiTX, for i = 1,2…L and selecting the node with the largest inner product. • If the weight vectors Wiare normalized, the inner product criterion is equivalent to the minimum Euclidean distance measure. c(X) = arg min ||X-Wi||, i = 1,2…L • With c(X) to indicate the output node of which the weight vector “matches” the input vector X the best (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  13. SOM building block (continued) • Adaptive process (by Oja 1982): • The weight vectors inside the neighborhood of the winner are usually updated by Hebbian type learning law. • the negative component is a nonlinear forgetting term. (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  14. Discrete-time update format for the adaptive process • Simplifying the equation, • yi(t) = 1 if node i inside the neighborhood of the winner c • yi(t) = 0 otherwise • Obtained discrete-time format (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  15. SOM Algorithm implementation 1. Select a winner ‘c’ in the map by : 2. Update the weights in the neighborhood of ‘c’ by Where ‘c’ is the neighborhood function, defined as: (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  16. Neighborhood Function • bell-shaped neighborhood Neighborhood is large at first, shrinks over time (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  17. square neighborhood (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  18. Learning Rate α(t) • Essential for convergence • Large enough for the network to adapt quickly for the new training patterns • Small enough for stability - the network would not forget the experience from the past training patterns • A decreasing function of time. (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  19. SOM Software Implementation • Matlab Neural Networks Tool box • SOM toolbox created by a group at the Helsinki Institute of Technology • SOMToolbox downloaded from website at http://www.cis.hut.fi/projects/somtoolbox/about.html (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  20. Structure of the presentation • Background • The algorithm • Analysis • Case 1 – motor data analysis • Case 2 – auditory data analysis • Case 3 – supplier change control • A qualitative comparison • Advanced issue – topology preserving • Conclusions (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  21. Weight convergence (Assumptions) • the input has discrete probability density • the learning rate α(k) satisfies conditions (Robbins and Monro, 1951): (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  22. Weight convergence (Results) • SOM algorithm (locally or globally) minimizes the objective function • Weights converge almost truly to a stationary solution if the stationary solution exists Lin, Si, “Weight Value Convergence of the SOM Algorithm for Discrete Input” (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  23. Voronoi polyhedra Voronoi polyhedra on Rn: Masked Voronoi polyhedra on: (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  24. Some extreme cases … • Assume neighborhood function is constant Nc(k)=Nc in the final learning phase, then where (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  25. Extreme case 1 Neighborhood covers the entire output map, i.e. Each weight vector converges to the same stationary state which is the mass center of the training data set. To eliminate the effect of initial conditions, we should use a neighborhood function covering a large range of the output map. (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  26. Extreme case 2 Neighborhood equals 0, i.e. where Wi* become the centroids of the cells of Voronoi partition of the inputs and the final iterations of SOM becomes a sequential updating process of vector quantization. SOM could be used for vector quantization by shrinking the range of the neighborhood function to zero during the learning process. (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  27. Observations • Robbins-Monro algorithm ensures weight convergence to the root dJ/dWi=0 almost truly if the root exists. • In practice the weights would only converge to local minima. • It has been observed that SOM is capable to some extent of escaping from local minima when it is used for vector quantization (Mcauliffe 1990). • Topology ordering of weights is not explicitly proved but it remains as a well observed practice in many applications. (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  28. Structure of the presentation • Background • The algorithm • Analysis • Case 1 – motor data analysis • Case 2 – auditory data analysis • Case 3 – supplier change control • A qualitative comparison • Advanced issue – topology preserving • Conclusions (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  29. Monkey’s motor cortical data analysis to interpret its movement intention Collaborators Andy Schwartz, Siming Lin (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  30. Motor Experiment – Overview (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  31. I3 I2 I1 time Firing rate calculation dt: bin size, Ti / Tj: start / end time of the i-th bin dij: the firing rate of the i-th bin [Ti , Tj], Tj = Ti+dt. Calculation of dij: the number of spike intervals overlapping with the i-th bin is first determined to be 3. As shown (counting from left to right), 30% of the first interval, 100% of the second interval, and 50% of the third interval are located in the I-th bin. Thus the equation above. (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  32. Self-Organizing Application – Motor Cortical Information processing • Spike signal and feature extraction • Computation models using SOM • Visualization of firing patterns of motor cortex • Neural trajectory computation • Weights are adaptively updated by the average discharge rates Input Average discharge rates of 81 cells Output Two-dimensional grid, each node codes the movement directions from the average discharge rates (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  33. The Self-Organized Map of Discharge Rates from 81 Neurons in the Center - Out Task (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  34. 1 8 1 1 7 7 8 7 1 7 7 1 1 8 6 6 2 2 2 6 8 2 6 8 6 2 3 5 5 3 3 4 3 3 4 4 4 5 5 5 The Self-Organized Map of Discharge Rates from 81 Neurons in the Center - Out Task (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  35. 10 10 10 8 10 7 9 8 7 9 9 11 7 8 11 9 8 8 10 8 7 10 8 7 12 11 12 6 8 7 9 6 6 12 11 13 11 6 6 9 10 12 10 10 5 5 11 13 11 5 5 13 5 4 12 13 12 4 2 13 13 3 14 13 3 4 14 14 2 3 13 14 15 14 2 3 14 15 16 15 3 1 15 15 16 2 1 15 16 1 1 1 15 16 16 15 1 2 1 2 The Self-Organized Map of Discharge Rates – from 81 Neurons in the Spiral Task (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  36. Neural Directions: Four Trials for Training, One for Testing in Spiral Task (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  37. Neural Trajectory: Four Trials for Training, one for testing Left: monkey finger trajectory; Right: SOM predicted trajectory (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  38. Neural Trajectory: Data from Spiral Tasks and Center-out Tasks for Training Left: monkey finger trajectory; Right: SOM predicted trajectory (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  39. Neural Trajectory: Average Testing Result from Five Trials Using Leave-K-Out Left: monkey finger trajectory; Right: SOM predicted trajectory (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  40. Trajectory Computation Error in 100 Bins (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  41. Structure of the presentation • Background • The algorithm • Analysis • Case 1 – motor data analysis • Case 2 – auditory data analysis • Case 3 – supplier change control • A qualitative comparison • Advanced issue – topology preserving • Conclusions (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  42. Guinea pig auditory cortical data analysis to interpret its perception of sound Collaborators Russ Witte , Jing Hu, Daryl Kipke (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  43. 1 mm Surgical implant, neural recordings Sample Signal from a single electrode Microvolts Time (sec) Each of 60 frequencies spanning 6 octaves were repeated 10 times.Each stimulus interval lasted 700ms, including 200ms of tone on time and 500ms of off time(interstimulus interval). (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  44. Stream of Neural Activity g180926a Time (sec) Raster Plot- 6s snapshot (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  45. Averaged Spike count of 22 chan 1701Hz 1473Hz (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  46. Spikerate of Channel1, Stimulus 1-60 (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  47. Data Processing • Channel selection - 30 channels were selected • Bin width - Basic unit to hold spike count from experiment data, 5ms, 10ms, 20ms, or higher. 70ms-bin size used • Noise filtering - Apply Gaussian filter to binned data. • Frequency grouping – 12 out of 60 stimuli are selected. It is therefore approximately one frequency per half Octave. • Trial loop selection/leave-1-out - Among the 10 loops of experimental data, take 9 for training, leave one for testing (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  48. SOM Training and Testing • Input vector: Bins from all channels are combined. Training/testing patterns are from all loops. • Output: 2-Dimensional grid. The position of each node in the map corresponds to certain spike pattern. • Training parameters: map size (eg 10 10), learning rate (0.02/0.0001), neighborhood function (Gaussian) and radius (eg 10/0), training/fine-tuning epochs, reducing schedules.etc. • Calibration after training: label the output using the labels (frequencies) of the training data (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  49. Tonotopic maps: natural vs. SOM 1473 1028 717 Neuronal activities of one channel that leads to the predicted stimulus map - Preserved topology Auditory cortex map Channel 12 has a narrow tuning curve. It is mostly tuned into 700-1500Hz auditory tones. (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

  50. Results from 10 sessions using leave_k_out (480) 965-6133 (voice) (480) 965-0461 (fax) si@asu.edu

More Related