M2AST:MLP-Mixer-based Adaptive Spatial-Temporal Graph Learning for Human Motion Prediction

Author:

Tang Junyi1,Liu Yuanwei1,Su Yong1,An Simin1

Affiliation:

1. Tianjin Normal University

Abstract

Abstract Human motion prediction is a challenging task in human-centric computer vision that involves forecasting future poses based on historical sequences. Despite recent progress in modeling spatial-temporal relationships of motion sequences using complex structured graphs, few approaches have been able to provide an adaptive and compact representation for varying graph structures of human motion. Inspired by the advantages of MLP-Mixer, a lightweight architecture developed for learning complex interactions in multi-dimensional data, we explore its potential as a backbone for motion prediction. Human motion prediction is a challenging task in human-centric computer vision, involving forecasting future poses based on historical sequences. Despite recent progress in modeling spatial-temporal relationships of motion sequences using complex structured graphs, few approaches have provided an adaptive and lightweight representation for varying graph structures of human motion. Taking inspiration from the advantages of MLP-Mixer, a lightweight architecture designed for learning complex interactions in multi-dimensional data, we explore its potential as a backbone for motion prediction. To this end, we propose a novel MLP-Mixer-based adaptive spatial-temporal pattern learning framework (M\(^2\)AST). Our framework includes an adaptive spatial mixer to model the spatial relationships between joints, an adaptive temporal mixer to learn temporal smoothness, and a local dynamic mixer to capture fine-grained cross-dependencies between joints of adjacent poses. The final method achieves a compact representation of human motion dynamics by adaptively considering spatial-temporal dependencies from coarse to fine. Unlike the trivial spatial-temporal MLP-Mixer, our proposed approach can more effectively capture both local and global spatial-temporal relationships simultaneously. We extensively evaluated our proposed framework on three commonly used benchmarks (Human3.6M, AMASS, 3DPW MoCap), demonstrating comparable or better performance than existing state-of-the-art methods in both short and long-term predictions, despite having significantly fewer parameters. Overall, our proposed framework provides a novel and efficient solution for human motion prediction with adaptive graph learning.

Publisher

Research Square Platform LLC

Reference61 articles.

1. Julieta Martinez and Michael J. Black and Javier Romero (2017) On Human Motion Prediction Using Recurrent Neural Networks. 4674--4683, {IEEE} Conference on Computer Vision and Pattern Recognition, {CVPR}

2. Chen Li and Zhen Zhang and Wee Sun Lee and Gim Hee Lee (2018) Convolutional Sequence to Sequence Model for Human Dynamics. 5226--5234, {IEEE} Conference on Computer Vision and Pattern Recognition, {CVPR}

3. Zhenguang Liu and Shuang Wu and Shuyuan Jin and Qi Liu and Shijian Lu and Roger Zimmermann and Li Cheng (2019) Towards Natural and Accurate Future Motion Prediction of Humans and Animals. 10004--10012, {IEEE} Conference on Computer Vision and Pattern Recognition, {CVPR}

4. Wei Mao and Miaomiao Liu and Mathieu Salzmann and Hongdong Li (2019) Learning Trajectory Dependencies for Human Motion Prediction. 9488--9496, {IEEE} International Conference on Computer Vision, {ICCV}

5. Maosen Li and Siheng Chen and Yangheng Zhao and Ya Zhang and Yanfeng Wang and Qi Tian (2020) Dynamic Multiscale Graph Neural Networks for 3D Skeleton Based Human Motion Prediction. 211--220, {IEEE} Conference on Computer Vision and Pattern Recognition, {CVPR}

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3