Learning Human Motion with Temporally Conditional Mamba

Quang Nguyen1       Tri Le1       Baoru Huang2       Minh Nhat Vu3       Ngan Le4       Thieu Vo5       Anh Nguyen2

1FPT Software AI Center   2University of Liverpool   3TU Wien  
4University of Arkansas   5National University of Singapore  

Abstract

Learning human motion based on a time-dependent input signal presents a challenging yet impactful task with various applications. The goal of this task is to generate or estimate human movement that consistently reflects the temporal patterns of conditioning inputs. Existing methods typically rely on cross-attention mechanisms to fuse the condition with motion. However, this approach primarily captures global interactions and struggles to maintain step-by-step temporal alignment. To address this limitation, we introduce Temporally Conditional Mamba, a new mamba-based model for human motion generation. Our approach integrates conditional information into the recurrent dynamics of the Mamba block, enabling better temporally aligned motion. To validate the effectiveness of our method, we evaluate it on a variety of human motion tasks. Extensive experiments demonstrate that our model significantly improves temporal alignment, motion realism, and condition consistency over state-of-the-art approaches.

BibTeX

Comming soon...
  

Acknowledgements

We borrow the page template from HyperNeRF. Special thanks to them!