Transport Object Detection in Street View Imagery Using Decomposed Convolutional Neural Networks
-
Published:2023-09-07
Issue:18
Volume:11
Page:3839
-
ISSN:2227-7390
-
Container-title:Mathematics
-
language:en
-
Short-container-title:Mathematics
Author:
Bai Yunpeng1ORCID, Shang Changjing1, Li Ying2, Shen Liang3ORCID, Jin Shangzhu4, Shen Qiang2
Affiliation:
1. Department of Computer Science, Aberystwyth University, Aberystwyth SY23 3DB, UK 2. School of Computer Science, Northwestern Polytechnical University, Xi’an 710129, China 3. School of Information Engineering, Fujian Business University, Fuzhou 350506, China 4. Information Office, Chongqing University of Science and Technology, Chongqing 401331, China
Abstract
Deep learning has achieved great successes in performing many visual recognition tasks, including object detection. Nevertheless, existing deep networks are computationally expensive and memory intensive, hindering their deployment in resource-constrained environments, such as mobile or embedded devices that are widely used by city travellers. Recently, estimating city-level travel patterns using street imagery has been shown to be a potentially valid way according to a case study with Google Street View (GSV), addressing a critical challenge in transport object detection. This paper presents a compressed deep network using tensor decomposition to detect transport objects in GSV images, which is sustainable and eco-friendly. In particular, a new dataset named Transport Mode Share-Tokyo (TMS-Tokyo) is created to serve the public for transport object detection. This is based on the selection and filtering of 32,555 acquired images that involve 50,827 visible transport objects (including cars, pedestrians, buses, trucks, motors, vans, cyclists and parked bicycles) from the GSV imagery of Tokyo. Then a compressed convolutional neural network (termed SVDet) is proposed for street view object detection via tensor train decomposition on a given baseline detector. The method proposed herein yields a mean average precision (mAP) of 77.6% on the newly introduced dataset, TMS-Tokyo, necessitating just 17.29 M parameters and a computational capacity of 16.52 G FLOPs. As such, it markedly surpasses the performance of existing state-of-the-art methods documented in the literature.
Subject
General Mathematics,Engineering (miscellaneous),Computer Science (miscellaneous)
Reference67 articles.
1. Object detection using YOLO: Challenges, architectural successors, datasets and applications;Diwan;Multimed. Tools Appl.,2023 2. Tools, techniques, datasets and application areas for object detection in an image: A review;Kaur;Multimed. Tools Appl.,2022 3. Bai, Z., Wu, G., Qi, X., Liu, Y., Oguchi, K., and Barth, M.J. (2022, January 4–9). Infrastructure-based object detection and tracking for cooperative driving automation: A survey. Proceedings of the 2022 IEEE Intelligent Vehicles Symposium (IV), Aachen, Germany. 4. Edge YOLO: Real-time intelligent object detection system based on edge-cloud cooperation in autonomous vehicles;Liang;IEEE Trans. Intell. Transp. Syst.,2022 5. Huang, Y., Chen, J., and Huang, D. (March, January 22). UFPMP-Det: Toward accurate and efficient object detection on drone imagery. Proceedings of the AAAI Conference on Artificial Intelligence, Online.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|