next up previous
Next: Experiments Up: A Neural Network Model Previous: Introduction

Architecture


 
Figure 2: Architecture of the RF-LISSOM network.
  A tiny RF-LISSOM network and retina are shown, along with connections to a single neuron (shown as a large circle). The input is an oriented Gaussian activity pattern on the retinal ganglion cells. The afferent connections form a local anatomical receptive field on the simulated retina. Neighboring neurons have different but highly overlapping RFs. Each neuron computes an initial response as a dot product of its receptive field and its afferent weight vector. The responses then repeatedly propagate within the cortex through the lateral connections and evolve into an activity ``bubble''. After the activity stabilizes, weights of the active neurons are adapted.
\begin{figure}
\centering
 \begin{minipage}
{3in}
 \vspace{0.1in} \centering
 \i...
 .../rf-lissom-architecture-bw.ps}}
\fi
 \end{minipage} \vspace{-0.1in} \end{figure}

The cortical architecture for the model has been simplified and reduced to the minimum necessary configuration to account for the observed phenomena. Because the focus is on the two-dimensional organization of the cortex, each ``neuron'' in the model cortex corresponds to a vertical column of cells through the six layers of the human cortex. The cortical network is modeled with a sheet of interconnected neurons and the retina with a sheet of retinal ganglion cells (figure 2). Neurons receive afferent connections from broad overlapping patches on the retina. The N × N network is projected on to the retina of R × R ganglion cells, and each neuron is connected to ganglion cells in a circular area of radius r around the projections. Thus, neurons at a particular cortical location receive afferents from the corresponding location on the retina. Since the LGN accurately reproduces the receptive fields of the retina, it has been bypassed for simplicity.

Each neuron also has reciprocal excitatory and inhibitory lateral connections with itself and other neurons. Lateral excitatory connections are short-range, connecting each neuron with itself and its close neighbors. Lateral inhibitory connections run for comparatively long distances, but also include connections to the neuron itself and to its neighbors.

The input to the model consists of 2-D ellipsoidal Gaussian patterns representing retinal ganglion cell activations. For training, the orientations of the Gaussians are chosen randomly from the uniform distribution in the range $[0,\pi)$. The elongated spots approximate natural visual stimuli after the edge detection and enhancement mechanisms in the retina. They can also be seen as a model of the intrinsic retinal activity waves that occur in late pre-natal development in mammals (Meister, Wong, Baylor, and Shatz, 1991). The RF-LISSOM network models the self-organization of the visual cortex based on these natural sources of elongated features.

The afferent weights are initially set to random values, and the lateral weights are preset to a smooth Gaussian profile. The connections are organized through an unsupervised learning process. At each training step, neurons start out with zero activity. The initial response $\eta_{ij}$ of neuron (i,j) is calculated as a weighted sum of the retinal activations:

 \begin{displaymath}
\eta_{ij} = \sigma \left( \sum_{a,b} \xi_{ab} \mu_{ij,ab} \right),\end{displaymath}

where $\xi_{ab}$ is the activation of retinal ganglion (a,b) within the anatomical RF of the neuron, $\mu_{ij,ab}$ is the corresponding afferent weight, and $\sigma$ is a piecewise linear approximation of the sigmoid activation function. The response evolves over a very short time scale through lateral interaction. At each time step, the neuron combines the above afferent activation $\sum \xi \mu$ with lateral excitation and inhibition:  \begin{displaymath}
\eta_{ij}(t) = \sigma \left( \sum \xi \mu +
 \gamma_e \sum_{...
 ...-1) -
 \gamma_i \sum_{k,l} I_{ij,kl} \eta_{kl}(t-1) \right) ,
 \end{displaymath}
where Eij,kl is the excitatory lateral connection weight on the connection from neuron (k,l) to neuron (i,j) , Iij,kl is the inhibitory connection weight, and $\eta_{kl}(t-1)$ is the activity of neuron (k,l) during the previous time step. The scaling factors $\gamma_e$ and $\gamma_i$ determine the relative strengths of excitatory and inhibitory lateral interactions.

While the cortical response is settling, the retinal activity remains constant. The activity pattern starts out diffuse and spread over a substantial part of the map, but within a few iterations of equation 2, converges into a small number of stable focused patches of activity, or activity bubbles. After the activity has settled, the connection weights of each neuron are modified. Both afferent and lateral weights adapt according to the same mechanism: the Hebb rule, normalized so that the sum of the weights is constant:

 \begin{displaymath}
w_{ij,mn}(t+\delta t)=\frac{ w_{ij,mn}(t) + \alpha \eta_{ij}...
 ...um_{mn} \left[ w_{ij,mn}(t) + \alpha \eta_{ij} X_{mn} \right]},\end{displaymath}

where $\eta_{ij}$ stands for the activity of neuron (i,j) in the final activity bubble, wij,mn is the afferent or lateral connection weight ($\mu$, E or I ), $\alpha$ is the learning rate for each type of connection ($\alpha_A$ for afferent weights, $\alpha_E$ for excitatory, and $\alpha_I$ for inhibitory) and Xmn is the presynaptic activity ($\xi$ for afferent, $\eta$ for lateral). The larger the product of the pre- and post-synaptic activity $\eta_{ij} X_{mn}$, the larger the weight change. Therefore, when the pre- and post-synaptic neurons fire together frequently, the connection becomes stronger. Both excitatory and inhibitory connections strengthen by correlated activity; normalization then redistributes the changes so that the sum of each weight type for each neuron remains constant.

At long distances, very few neurons have correlated activity and therefore most long-range connections eventually become weak. The weak connections can be eliminated periodically, resulting in patchy lateral connectivity similar to that observed in the visual cortex. The radius of the lateral excitatory interactions starts out large, but as self-organization progresses, it is decreased until it covers only the nearest neighbors. Such a decrease is necessary for global topographic order to develop and for the receptive fields to become well-tuned at the same time.


next up previous
Next: Experiments Up: A Neural Network Model Previous: Introduction
James A. Bednar
9/15/1997