Author:
Wang Tao,Cai Yuanzheng,Liang Lingyu,Ye Dongyi
Abstract
We address the problem of localizing waste objects from a color image and an optional depth image, which is a key perception component for robotic interaction with such objects. Specifically, our method integrates the intensity and depth information at multiple levels of spatial granularity. Firstly, a scene-level deep network produces an initial coarse segmentation, based on which we select a few potential object regions to zoom in and perform fine segmentation. The results of the above steps are further integrated into a densely connected conditional random field that learns to respect the appearance, depth, and spatial affinities with pixel-level accuracy. In addition, we create a new RGBD waste object segmentation dataset, MJU-Waste, that is made public to facilitate future research in this area. The efficacy of our method is validated on both MJU-Waste and the Trash Annotation in Context (TACO) dataset.
Funder
National Natural Science Foundation of China
Natural Science Foundation of Fujian Province
Department of Education, Fujian Province
Minjiang University
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference86 articles.
1. Visual attention within and around the field of focal attention: A zoom lens model
2. Spatial Frequency and Selective Attention to Local and Global Information
3. The Psychology of Attention;Pashler,1999
4. Vision Science: Photons to Phenomenology;Palmer,1999
5. TACO: Trash Annotations in Context for Litter Detection;Proença;arXiv,2020
Cited by
47 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献