Computational Maps in the Visual Cortex
     Figure 15.4
MiikkulainenBednarChoeSirosh
Home    
About the Authors
Back Cover    
Table of Contents 
Sample Chapter 
Figures    
References    
Errata    
Demos     
Talks/Courses 
Software    
Credits    
Purchase online at:

springeronline.com
amazon.com

Click on the image to see a PDF version (for zooming in)

Fig. 15.4. Training time and memory usage in LISSOM vs. GLISSOM. Data are shown for a LISSOM network of 144 × 144 units and a GLISSOM network grown from 36 × 36 to 144 × 144 units as described in Section 15.4.2. (a) Each line shows a 20-point running average of the time spent in training for one iteration, with a data point measured every 10 iterations. Only training time is shown; times for initialization, plotting images, pruning, and scaling networks are not included. Computational requirements of LISSOM peak at the early iterations, falling as the excitatory radius (and thus the number of neurons activated by a given pattern) shrinks and as the neurons become more selective. In contrast, GLISSOM requires little computation time until the final iterations. Because the total training time is determined by the area under each curve, GLISSOM is much more efficient to train overall. (b) Each line shows the number of connections simulated at a given iteration. LISSOM's memory usage peaks at early iterations, decreasing at first in a series of small drops as the lateral excitatory radius shrinks, and then later in a few large drops as long-range inhibitory weights are pruned at iterations 6500, 12,000, and 16,000. Similar shrinking and pruning takes place in GLISSOM, while the network size is scaled up at iterations 4000, 6500, 12,000, and 16,000. Because the GLISSOM map starts out small, memory usage peaks much later, and remains bounded because connections are pruned as the network is grown. As a result, the peak number of connections (which determines the memory usage) in GLISSOM is as low as the smallest number of connections in LISSOM.