Papers
arxiv:2603.09930

Fine-grained Motion Retrieval via Joint-Angle Motion Images and Token-Patch Late Interaction

Published on Mar 10
· Submitted by
Zhuchenyang Liu
on Mar 16
Authors:
,
,
,

Abstract

A novel text-motion retrieval approach uses joint-angle-based motion representation and Vision Transformer-compatible pseudo-images to achieve superior performance with interpretable fine-grained alignments.

AI-generated summary

Text-motion retrieval aims to learn a semantically aligned latent space between natural language descriptions and 3D human motion skeleton sequences, enabling bidirectional search across the two modalities. Most existing methods use a dual-encoder framework that compresses motion and text into global embeddings, discarding fine-grained local correspondences, and thus reducing accuracy. Additionally, these global-embedding methods offer limited interpretability of the retrieval results. To overcome these limitations, we propose an interpretable, joint-angle-based motion representation that maps joint-level local features into a structured pseudo-image, compatible with pre-trained Vision Transformers. For text-to-motion retrieval, we employ MaxSim, a token-wise late interaction mechanism, and enhance it with Masked Language Modeling regularization to foster robust, interpretable text-motion alignment. Extensive experiments on HumanML3D and KIT-ML show that our method outperforms state-of-the-art text-motion retrieval approaches while offering interpretable fine-grained correspondences between text and motion. The code is available in the supplementary material.

Community

Paper author Paper submitter

This paper introduces a novel approach to text-motion retrieval by addressing the limitations of traditional global-embedding paradigms. Most existing methods compress 3D human motion sequences and textual descriptions into single global vectors, which is computationally efficient but inevitably discards fine-grained local correspondences. To overcome this information bottleneck, the authors propose replacing global embeddings with a structurally grounded, fine-grained late interaction mechanism.

The core innovation relies on explicitly decoupling local joint movements from the body's global trajectory using a joint-angle-based motion representation. These joint-level features are mapped into a structured 224x224 pseudo-image, making them compatible with pre-trained Vision Transformers. For the retrieval process, the framework utilizes MaxSim, a token-wise late interaction operator that matches text tokens to motion patches, which is further stabilized by a Masked Language Modeling regularization on the text encoder.

Extensive experiments conducted on the HumanML3D and KIT-ML datasets demonstrate that this approach consistently outperforms state-of-the-art text-motion retrieval baselines. Furthermore, the token-to-patch matching produces highly interpretable correspondence maps that transparently align textual semantics with specific body joints and temporal phases. This level of transparency and granularity provides a robust foundation for downstream applications like language-driven motion generation and localized editing.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2603.09930 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2603.09930 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2603.09930 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.