Abstract
AbstractAttention leads the gaze of the observer towards interesting items, allowing a detailed analysis only for selected regions of a scene. A robot can take advantage of the perceptual organisation of the features in the scene to guide its attention to better understand its environment. Current bottom–up attention models work with standard RGB cameras requiring a significant amount of time to detect the most salient item in a frame-based fashion. Event-driven cameras are an innovative technology to asynchronously detect contrast changes in the scene with a high temporal resolution and low latency. We propose a new neuromorphic pipeline exploiting the asynchronous output of the event-driven cameras to generate saliency maps of the scene. In an attempt to further decrease the latency, the neuromorphic attention model is implemented in a spiking neural network on SpiNNaker, a dedicated neuromorphic platform. The proposed implementation has been compared with its bio-inspired GPU counterpart, and it has been benchmarked against ground truth fixational maps. The system successfully detects items in the scene, producing saliency maps comparable with the GPU implementation. The asynchronous pipeline achieves an average of 16 ms latency to produce a usable saliency map.
Reference52 articles.
1. Bio-inspired visual attention in agile sensing for target detection;Liu;Int. J. Sensor Netw.,2009
2. Computational models of visual attention;Tsotsos;Scholarpedia,2011
3. The icub humanoid robot: an open platform for research in embodied cognition;Metta,2008
4. Recurrent models of visual attention;Mnih,2014
5. Fixation prediction through multimodal analysis;Min;ACM Trans. Multimed. Comput. Commun. Appl.,2016
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献