1 / 19

Enhancing SOM Efficiency through Data Order Adjustment

This study delves into optimizing Self-Organizing Maps (SOM) by adjusting data order to reduce competition load and enhance learning speed. The process involves modifying the input data order based on inter-class distances, culminating in a converged learning point. Experiment results show a significant improvement in efficiency with small distance data streams. However, data with larger distances exhibit similar results to conventional SOM. The study suggests that terminal conditions play a crucial role in the efficiency of SOM.

Download Presentation

Enhancing SOM Efficiency through Data Order Adjustment

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Efficient SOM Learning by Data Order Adjustment Authors: Miyoshi et al. Advisor: Dr. Hsu Graduate :Yu-Wei Su

  2. Outline • Motivation • Objective • SOM • Data order and learning convergence • Data order • Ending point of convergence • Experiments • Conclusion • opinion

  3. Motivation • In SOM, there are many factors to aggravate computational load and competition

  4. Objective • Reducing the competition load and increasing the speed of SOM

  5. SOM • SOM is a algorithm that map dimension from high to low, always two dimension • Step 1: finding BMU of each datum • Step 2: modifying the value of BMU and neighborhood nodes

  6. Data order and learning convergence • The SOM spends lots of time to learn because of large map size, large quantity of input data and many dimensions in data et al. • In the beginning stage of learning process, SOM map is dynamically and widely and that is depended on the distance of each input data

  7. Data order and learning convergence( cont.) • Adjusting data order based on the distance between data classes to reduce the competition load

  8. Data order • To change order of input data, using class distance that is calculated by class center • First select typical data as class center in each class • And calculate Euclidian distance between all class centers as class distance

  9. Data order( cont.) • Order 1: random order • Order 2: the largest distance order based on previous data class scli: selected data class i ucli: still unselected data class i cd(A,B) :class distance between A and B

  10. Data order( cont.) • Order 3: the smallest distance order based on previous data class • Order 4: the smallest distance order based on all classes scli: selected data class i ucli: still unselected data class i cd(A,B) :class distance between A and B

  11. Data order( cont.) • Order 5: the largest distance order based on all classes • Order 6: average distance order based on all classes scli: selected data class i ucli: still unselected data class i cd(A,B) :class distance between A and B

  12. Ending point of convergence • Definition of converging point of learning • Keep maximum Euclidian distance for all nodes in each step of learning • Test the difference between (|xdn-xdn-1|) whether (|xdn-xdn-1|) is smaller than Th1 • If it is, test how long the distance are continued smaller • If it continues long enough than Th2 it is determined as the ending point of learning

  13. Ending point of convergence (cont.) Th1 : threshold of difference Th2 : threshold of period xdn : n-th max distance through all input data and output nodes ed(A,B) : Euclidian distance between A and B dti : input data I onj: output node j dmax : total of input data nmax : total of output nodes

  14. Experiments • Experiment data • Synthetic data, 5 dim, 7 classes each 49 data • Parameters • Size of map 8x8, initial neighborhood from 3 to 5, initial learning rate from 0.2 to 0.8, 300 maps that initialized at random

  15. Experiments ( cont.) • Learning rate function • Neighborhood function

  16. Experiments ( cont.)

  17. Experiments ( cont.)

  18. Conclusion • The data stream of small distance makes maximum 9% improvement • The data stream of large distance still similar with conventional SOM • All order make no remarkable difference in result map

  19. Opinion • No experiments of comparison with others • The terminal condition is a good idea

More Related