A Survey on Visual Mamba
-
Published:2024-06-28
Issue:13
Volume:14
Page:5683
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Zhang Hanwei123, Zhu Ying4, Wang Dan4, Zhang Lijun1, Chen Tianxiang5, Wang Ziyang6ORCID, Ye Zi2ORCID
Affiliation:
1. Automotive Software Innovation Center, Chongqing 401331, China 2. Institute of Intelligent Software, Guangzhou 511458, China 3. Department of Computer Science, Saarland University, 66424 Homburg, Germany 4. Hangzhou Institute for Advanced Study, University of Chinese Academy of Sciences, Hangzhou 310024, China 5. School of Cyber Space and Technology, University of Science and Technology of China, Hefei 230026, China 6. Department of Computer Science, University of Oxford, Oxford OX3 7LD, UK
Abstract
State space models (SSM) with selection mechanisms and hardware-aware architectures, namely Mamba, have recently shown significant potential in long-sequence modeling. Since the complexity of transformers’ self-attention mechanism is quadratic with image size, as well as increasing computational demands, researchers are currently exploring how to adapt Mamba for computer vision tasks. This paper is the first comprehensive survey that aims to provide an in-depth analysis of Mamba models within the domain of computer vision. It begins by exploring the foundational concepts contributing to Mamba’s success, including the SSM framework, selection mechanisms, and hardware-aware design. Then, we review these vision Mamba models by categorizing them into foundational models and those enhanced with techniques including convolution, recurrence, and attention to improve their sophistication. Furthermore, we investigate the widespread applications of Mamba in vision tasks, which include their use as a backbone in various levels of vision processing. This encompasses general visual tasks, medical visual tasks (e.g., 2D/3D segmentation, classification, image registration, etc.), and remote sensing visual tasks. In particular, we introduce general visual tasks from two levels: high/mid-level vision (e.g., object detection, segmentation, video classification, etc.) and low-level vision (e.g., image super-resolution, image restoration, visual generation, etc.). We hope this endeavor will spark additional interest within the community to address current challenges and further apply Mamba models in computer vision.
Reference118 articles.
1. Rosenblatt, F. (1957). The Perceptron, a Perceiving and Recognizing Automaton Project Para, Cornell Aeronautical Laboratory. 2. Rosenblatt, F., Jones, B., Smith, T., Brown, C., Green, M., Wilson, A., Taylor, J., White, P., King, R., and Johnson, L. (1962). Principles of Neurodynamics: Perceptrons and the Theory of Brain Mechanisms, Spartan Books. 3. Gradient-based learning applied to document recognition;LeCun;Proc. IEEE,1998 4. Imagenet classification with deep convolutional neural networks;Krizhevsky;Adv. Neural Inf. Process. Syst.,2012 5. Long short-term memory;Hochreiter;Neural Comput.,1997
|
|