Encoder-Free Human Motion Understanding
via Structured Motion Descriptions

Yao Zhang1, Zhuchenyang Liu1, Thomas Ploetz2, Yu Xiao1
1Aalto University    2Georgia Institute of Technology
Comparison of encoder-based paradigm vs. our SMD-based paradigm
Figure 1. Previous encoder-based approaches (top) require a learned motion encoder and multi-stage alignment training tied to a specific LLM. Our approach (bottom) converts motion to Structured Motion Descriptions via deterministic rule-based computation, so any LLM can process it directly with LoRA fine-tuning as the only training step.

Abstract

The world knowledge and reasoning capabilities of text-based large language models (LLMs) are advancing rapidly, yet current approaches to human motion understanding — including motion question answering and captioning — have not fully exploited these capabilities. Existing LLM-based methods typically learn motion–language alignment through dedicated encoders that project motion features into the LLM's embedding space, remaining constrained by cross-modal representation and alignment. Inspired by biomechanical analysis, where joint angles and body-part kinematics have long served as a precise descriptive language for human movement, we propose Structured Motion Description (SMD), a rule-based, deterministic approach that converts joint position sequences into structured natural language descriptions of joint angles, body-part movements, and global trajectory. By representing motion as text, SMD enables LLMs to apply their pretrained knowledge of body parts, spatial directions, and movement semantics directly to motion reasoning, without requiring learned encoders or alignment modules.

SMD goes beyond state-of-the-art results on both motion question answering (66.7% on BABEL-QA, 90.1% on HuMMan-QA) and motion captioning (R@1 = 0.584, CIDEr = 53.16 on HumanML3D), surpassing prior methods. The same text input works across different LLMs with only lightweight LoRA adaptation (validated on 8 LLMs from 6 model families), and its human-readable representation enables interpretable attention analysis over motion descriptions.

Method

SMD construction: joint angles + temporal segmentation + global trajectory
Figure 2. SMD construction. For each motion, we compute 26 biomechanically meaningful joint angles from the 22-joint skeleton, segment each angle's time series into peak–valley episodes, and combine them with global body trajectory into a single structured text. Both the angle descriptors and trajectory descriptions are written in terms the LLM already understands (e.g., "Right Shoulder Flexion (raising arm forward): increases -1° → 84° [3.4s–4.5s]").

Example SMD

One real SMD from HumanML3D (Top-3 joint variant, ~1K tokens):

Motion: 7.0s (140 frames at 20 FPS)

Trajectory: displacement 0.01m, height change +0.00m, avg height 0.95m

Global Trajectory:
  Height: static at 0.954m throughout
  Forward Position: static at 0.001m throughout
  Lateral Position: static at 0.002m throughout

Joint Angles:
 [Right Shoulder]
  Right Shoulder Flexion (raising arm forward):
    holds at 3° [0.0s–3.4s] (3.4s)
    increases: -1° → 84° [3.4s–4.5s]
    holds at 82° [4.6s–5.5s] (0.9s)
    ...

More samples in the GitHub examples/ and the full 17 variants on the HuggingFace dataset repo.

Main Results

Motion Question Answering (accuracy %)

MethodInputBackboneBABEL-QA ↑HuMMan-QA ↑
IMoRe [Li et al., ICCV 2025]jointsSpecialized60.175.2
MotionGPT3-Qwen (our controlled baseline)VAE tokensQwen2.5-7B50.122.0
SMD + Qwen2.5-7B (LoRA)textQwen2.5-7B66.790.1

MotionGPT3-Qwen is a controlled baseline we constructed that replaces SMD with VAE-encoded motion tokens on the same Qwen2.5-7B backbone — isolating the contribution of the motion representation.

Motion Captioning on HumanML3D

MethodR@1 ↑R@3 ↑MM-Dist ↓CIDEr ↑BERTScore ↑
MotionGPT3 [Zhu et al., 2025]0.5730.8642.4340.635.2
SMD + Qwen2.5-7B (LoRA)0.5840.8832.3553.1645.58

See the paper for the complete tables, backbone-portability study across 8 LLMs, and ablations on trajectory / joint-count / segmentation variants.

Interpretability

Attention heatmap over SMD tokens
Figure 3. Because SMD is human-readable, we can directly visualize which angle descriptors and trajectory segments the fine-tuned LLM attends to when producing a caption — something encoder-based approaches cannot offer without a separate probing pipeline.

Resources

Citation

@article{zhang2026smd, title = {Encoder-Free Human Motion Understanding via Structured Motion Descriptions}, author = {Zhang, Yao and Liu, Zhuchenyang and Ploetz, Thomas and Xiao, Yu}, journal = {arXiv preprint arXiv:2604.21668}, year = {2026} }