Diffuse-CLoC: Guided Diffusion for Physics-based Character Look-ahead Control

UC Berkeley, Stanford University, RAI Institute
SIGGRAPH 2025

*Indicates Joint First Authors

Diffuse-CLoC enables intuitive steering of physics-based character motion through guided diffusion, combining the flexibility of kinematic motion generation with physical realism.

Zero-Shot Generalization to Downstream Tasks

Dynamic Obstacle Avoidance
Character smoothly navigates around dynamic obstacles.
Static Obstacle Navigation
Character moves while avoiding static obstalces
Task-space Control
Desired right end-effector tracking while maintaining whole-body coordination.
Long-horizon Planning
Complex multi-stage tasks executed with a single pre-trained model.
Motion In-betweening
Generates physically plausible transitions between keyframe poses.
Jumping Between Pillars
Can generalize to jumping over increasing pillar heights despite only seeing jumping on flatground.
Waypoint
Can easily guide the character between different locations.
Interactive Joystick Control
Real-time responsive motion adaptation to user input.

Abstract

We present Diffuse-CLoC, a guided diffusion framework for physics-based look-ahead control that enables intuitive, steerable, and physically realistic motion generation. While existing kinematics motion generation with diffusion models offer intuitive steering capabilities with inference-time conditioning, they often fail to produce physically viable motions. In contrast, recent diffusion-based control policies have shown promise in generating physically realizable motion sequences, but the lack of kinematics prediction limits their steerability. Diffuse-CLoC addresses these challenges through a key insight: modeling the joint distribution of states and actions within a single diffusion model makes action generation steerable by conditioning it on the predicted states. This approach allows us to leverage established conditioning techniques from kinematic motion generation while producing physically realistic motions. As a result, we achieve planning capabilities without the need for a high-level planner. Our method handles a diverse set of unseen long-horizon downstream tasks through a single pre-trained model, including static and dynamic obstacle avoidance, motion in-betweening, and task-space control. Experimental results show that our method significantly outperforms the traditional hierarchical framework of high-level motion diffusion and low-level tracking.

Presentation

Baseline Comparisons

Diffuse-CLoC vs. Kinematic Motion Model with Universal RL motion Tracker

Additional Tasks

BibTeX

@article{huang2025diffusecloc,
  title={Diffuse-CLoC: Guided Diffusion for Physics-based Character Look-ahead Control},
  author={Huang, Xiaoyu and Truong, Takara and Zhang, Yunbo and Yu, Fangzhou and Sleiman, Jean Pierre and Hodgins, Jessica and Sreenath, Koushil and Farshidian, Farbod},
  journal={ACM Transactions on Graphics (TOG)},
  volume={44},
  number={4},
  pages={1--13},
  year={2025},
  publisher={ACM},
  note={SIGGRAPH 2025}
}