Welcome to The Neuromorphic Engineer
Learning to correct orientation estimates using spatial memory

PDF version | Permalink

Tarek Massoud and Timothy Horiuchi

6 September 2011

A simple learning rule that implements a spatial map can be used for online correction of position estimates in a neuromorphic head direction cell system.

The survival of many animals is dependent on knowing where they are in relation to sources of food and shelter. While spatial navigation clearly requires memory (i.e., storage of and access to knowledge) and an estimate of one's position within this spatial knowledge, the algorithms and supporting neural mechanisms are poorly understood. One way to estimate one's position is to maintain a quantitative estimate based on the integration of velocities (odometry, or dead reckoning) from a reference point or it can be inferred approximately through the sensory recognition of position-linked sensory memories. In mammals, neurons in the hippocampal formation (head-direction,1 grid,2 and place cells3) have been discovered that appear to support both of these functions. In the absence of external sensory cues, animals can navigate successfully but errors accumulate over time. When sensory cues are present, drifts in navigation are not observed.4 This suggests that animals use noisy internal estimates of their position to navigate and sensory cues are used to correct for drifts caused by the noise.

To begin exploring how the brain could perform such online corrections, we have used a rotational odometry (‘head-direction’) system that is being developed as part of a larger bat echolocation modeling project to link sensed objects to different directions in memory.5, 6 Our mixed hardware/software system offers a biologically-plausible model of how the brain could use this memory to keep its noisy estimate of the orientation aligned with the environment.

In the system, a sonar head is mounted on a rotating platform from which the rotation velocity can be measured (see Figure 1 for a schematic of the environment). The head is equipped with a simple sonar system that can detect objects and report their range. A neuromorphic VLSI-based head-direction (HD) system integrates angular velocity to maintain an estimate of the head's orientation (see Figure 2 for a detailed block diagram). The system suffers from errors in the integration, which accumulate over time, causing the estimate to drift away from the actual position in the space. To mitigate this problem, a group of conjunctive cells use Hebbian learning to combine incoming sonar data and the location of activity in the HD system: the neurons are used to correct for drifts in the position estimates when data is available. Expectation cells are also important. Like the conjunctive cells, these learn the association between orientation and sensory input, but they reflect the activity of the object cells (i.e., sensory input). By observing both the object and expectation cells, the system can determine whether the HD estimate is aligned with the environment or a correction is required.


A sonar transducer is mounted on a rotating platform from which the rotation velocity can be measured. The grey cone represents the effective field of view of the sonar. The individual targets are distinguished by their radial distance from the head.6


Block diagram of the system. Blocks enclosed in the dashed-line box are implemented in software and the other blocks are in hardware. The black arrows indicate hardwired synaptic connections, the arrows in grey show the plastic synaptic connections, and the white arrows indicate hardwired synaptic connections that provide teacher signals to guide the learning process.6

We conducted an experiment with two objects at 45° and 100° (see Figure 3). In the two cases shown, the head orientation and the HD estimate are initially aligned. As the head begins to rotate, the noisy integration accumulates errors and begins drifting away from the true orientation. In the case without learning, the errors accumulate and the orientation error grows over time (left panel). In contrast, the learning example (right panel) shows that once the objects are associated with specific orientations, the next encounter realigns the HD system to the stored orientation.


Results from an experiment with two targets present (at 45°and 100°). With no correction; the HDS estimate accumulates error with time (left). With correction, as the HDS estimate drifts, spatial memories of the targets are used to reset it to the accurate position (right).6

In summary, we have demonstrated how sensory cues in the environment can be dynamically associated with internal states to compensate for (and potentially calibrate against) drifts in integration. We plan to extend this system as we expand to two-dimensional odometry (i.e., place cells) and more sophisticated object recognition.




Authors

Tarek Massoud
University of Maryland

Tarek Massoud joined the University of Maryland in 2005 where he is currently working toward a PhD. His general research interests are analog very-large-scale integration (VLSI) and computational neuroscience.

Timothy Horiuchi
University of Maryland

Timothy Horiuchi received both his BS in Electrical Engineering (1989) and PhD in Computation and Neural Systems (1997) from the California Institute of Technology. He worked as a postdoctoral scholar at the Johns Hopkins University until moving to the University of Maryland in 1999. He is currently an associate professor, mainly interested in the design and fabrication of neuromorphic VLSI circuits and the implementation of neural computation in silicon. His primary focus has been on the modeling of spike-based neural processing models of the auditory system of the echolocating bat.


References
  1. J. S. Taube, R. U. Muller and J. B., Jr. Ranck, Head-direction cells recorded from the postsubiculum in freely moving rats. I. Description and quantitative analysis, J. Neurosci. 10, pp. 420-435, 1990.

  2. T. Hafting, M. Fyhn, S. Molden, M. B. Moser and E. I. Moser, Microstructure of a spatial map in the entorhinal cortex, Nature 436, pp. 801-806, 2005.

  3. J. O'Keefe and J. Dostrovsky, The hippocampus as a spatial map. Preliminary evidence from unit activity in the freely-moving rat, Brain Res. 34, pp. 171-175, 1971.

  4. J. J. Knierim, H. S. Kudrimoti and B. L. McNaughton, Place cells, head direction cells, and the learning of landmark stability, J. Neurosci. 15, pp. 1648-1659, 1995.

  5. T. M. Massoud and T. K. Horiuchi, A neuromorphic VLSI head direction cell system, IEEE Trans. Circuits Syst. I: Regular papers 58, pp. 150-163, 2011.

  6. T. M. Massoud and T. K. Horiuchi, Online correction of orientation estimates using spatial memory in a neuromorphic head direction system, Int'l Symp. Circuits Syst., pp. 2429-2432, 2011.


 
DOI:  10.2417/1201108.003828

@NeuroEng



Tell us what to cover!

If you'd like to write an article or know of someone else who is doing relevant and interesting stuff, let us know. E-mail the editor and suggest the subject for the article and, if you're suggesting someone else's work, tell us their name, affiliation, and e-mail.