Volume 10 Supplement 1

Eighteenth Annual Computational Neuroscience Meeting: CNS*2009

Open Access

Sharpening projections

BMC Neuroscience200910(Suppl 1):P214

DOI: 10.1186/1471-2202-10-S1-P214

Published: 13 July 2009

It is known that neuronal arbors can project roughly topographically from their source to their target area [1]. However, developmental rules in the cortex often provide only roughly reciprocal connections through chemical cues governing the physical location of distant axonal arbors, leaving open the question of whether more precise reciprocal connectivity is possible in these cases [2, 3]. We address this question of whether it is possible for a biologically plausible learning rule to adjust the synaptic weights so that the projection effectively becomes more precise than the anatomy alone provides. We have discovered a biologically plausible set of learning rules that can adjust the synaptic weights so that precisely reciprocal ones are strengthened while others are weakened, thus effectively increasing the specificity of the projections.

The question introduced above can be generalized to any number of areas connected in a feed-forward cycle with feedback only coming as a result of information traveling around the entire cycle. We examined the cases of two or three areas connected in a directed cycle. Each area was represented by a pool of linear threshold units having a sigmoidal threshold function, with inter-area weights initialized to random weight matrices, corresponding to the worst case of a neuronal arbor completely covering the target area. For the dynamics of the unit activities and the weights, we combined three techniques that plausibly have biological counterparts: winner-take-all circuitry [4, 5], activity regulation [6], and Hebbian learning [7, 8]. On the shortest time scale, winner-take-all circuitry within each pool ensures that the current configuration of activity consists of a clearly localized region of activation. On an intermediate time scale, activity regulation within each unit dampens the likelihood for units to win the winner-take-all competition if they have been highly active but increases their chance to win if they have been inactive, thereby enforcing fair use of the units. On the slowest time scale, Hebbian learning increases the weights along active cycles, making them more likely to recur. Over time, this encourages specific cycles to be strengthened, while the activity regulation ensures that these specific cycles cover all of the units in each pool.

Depending on the combinations of parameters used, it was possible to generate two kinds of reciprocal connectivity. In the first kind, strong precise reciprocal connections developed, with weaker strengths for connections close to being reciprocal, and much weaker connections elsewhere. In other words, the product of the weight matrices was a blurred identity matrix. In the second kind, each area partitioned itself into distinct subgroups, and then the subgroups formed fully connected cycles. In this case, the product of the matrices was a crisp block identity matrix. Both forms of connectivity can be useful within larger artificial architectures and we claim they could easily occur in brains.

Declarations

Acknowledgements

The authors, who are listed alphabetically, would like to thank ETH Research Grant ETH-23 08-1 and EU Project Grant FET-IP-216593.

Authors’ Affiliations

(1)
Institute of Neuroinformatics, University of Zurich and ETH Zurich
(2)
Institute of Theoretical Computer Science, ETH Zurich

References

  1. Essen D, Zeki S: The topographic organization of rhesus monkey prestriate cortex. J Physiol. 1978, 277: 193-226.PubMed CentralPubMedView ArticleGoogle Scholar
  2. Felleman D, Essen D: Distributed hierarchical processing in the primate cerebral cortex. Cereb Cortex. 1991, 1: 1-47. 10.1093/cercor/1.1.1-a.PubMedView ArticleGoogle Scholar
  3. Price D, Kennedy H, Dehay C, Zhou L, Mercier M: The development of cortical connections. Eur J Neurosci. 2006, 23: 910-920. 10.1111/j.1460-9568.2006.04620.x.PubMedView ArticleGoogle Scholar
  4. Dayan P, Abbott L: Theoretical neuroscience: computational and mathematical modeling of neural systems. 2001, London: MIT PressGoogle Scholar
  5. Hertz J, Krogh A, Palmer R: Introduction to the theory of neural computation. 1991, Boulder: Westview PressGoogle Scholar
  6. Turrigiano G, Nelson S: Homeostatic plasticity in the developing nervous system. Nat Rev Neurosci. 2004, 5: 97-107. 10.1038/nrn1327.PubMedView ArticleGoogle Scholar
  7. Buonomano D, Merzenich M: Cortical plasticity: From synapses to maps. Annu Rev Neurosci. 1998, 21: 149-186. 10.1146/annurev.neuro.21.1.149.PubMedView ArticleGoogle Scholar
  8. Hebb D: The organization of behavior: A neuropsychological theory. 1949, New York: WileyGoogle Scholar

Copyright

© Cook et al; licensee BioMed Central Ltd. 2009

This article is published under license to BioMed Central Ltd.

Advertisement