Author:
Ghosh Suman,D’Angelo Giulia,Glover Arren,Iacono Massimiliano,Niebur Ernst,Bartolozzi Chiara
Abstract
AbstractTo interact with its environment, a robot working in 3D space needs to organise its visual input in terms of objects or their perceptual precursors, proto-objects. Among other visual cues, depth is a submodality used to direct attention to visual features and objects. Current depth-based proto-object attention models have been implemented for standard RGB-D cameras that produce synchronous frames. In contrast, event cameras are neuromorphic sensors that loosely mimic the function of the human retina by asynchronously encoding per-pixel brightness changes at very high temporal resolution, thereby providing advantages like high dynamic range, efficiency (thanks to their high degree of signal compression), and low latency. We propose a bio-inspired bottom-up attention model that exploits event-driven sensing to generate depth-based saliency maps that allow a robot to interact with complex visual input. We use event-cameras mounted in the eyes of the iCub humanoid robot to directly extract edge, disparity and motion information. Real-world experiments demonstrate that our system robustly selects salient objects near the robot in the presence of clutter and dynamic scene changes, for the benefit of downstream applications like object segmentation, tracking and robot interaction with external objects.
Publisher
Springer Science and Business Media LLC
Reference66 articles.
1. Tsotsos, J. K. Analyzing vision at the complexity level. Behav. Brain Sci. 13, 423–445 (1990).
2. Rea, F., Metta, G. & Bartolozzi, C. Event-driven visual attention for the humanoid robot icub. Front. Neurosci. 7, 234. https://doi.org/10.3389/fnins.2013.00234 (2013).
3. Clark, J. J. & Ferrier, N. J. Modal control of an attentive vision system. In ICCV, 514–523 (1988).
4. Pahlavan, K., Uhlin, T. & Eklundh, J.-O. Integrating primary ocular processes. In European Conference on Computer Vision, 526–541 (Springer, 1992).
5. Bruce, N. D. & Tsotsos, J. K. An attentional framework for stereo vision. In The 2nd Canadian Conference on Computer and Robot Vision (CRV’05), 88–95 (IEEE, 2005).
Cited by
8 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献