News & Events


Neural mechanisms of motion and form perception to control visual behavior

Prof. Heiko Neumann, Universität Ulm


Observers extract relevant information from a continuous stream of visual information that impinges the retina. Task-relevant features need to be extracted and represented enabling an agent to interact with the environment. A biologically inspired network architecture is presented that builds upon generic principles of cortical organization and canonical computation, such as topographically organized sheets of interconnected columnar representations, compartmentalization into layers, and converging feedforward and feedback signal flow.


For several key visual tasks it is demonstrated how they can be accomplished using such generic processing principles. Examples are the segmentation of a scene into surfaces by segregating them from the background, calculating optical flow for self-motion estimation and depth encoding, and predicting the behavior of observed articulated body motion. As a brief outlook, recent developments in the modeling of visual processing using event-based sensors are presented.

Date: 09.05.2014

Time: 15:30 h

Location: Cartesium, Bremen




Heiko_Neumann_01.pdf117 K