Perceiving Systems Conference Paper 2025

InterDyn: Controllable Interactive Dynamics with Video Diffusion Models

project arXiv
Thumb ticker sm rick akkerman
Perceiving Systems
Master's Thesis Student
Thumb ticker sm img 4026
Perceiving Systems
  • Doctoral Researcher
Thumb ticker sm headshot2021
Perceiving Systems
Director
Thumb ticker sm 04 1.58.1   crop3
Perceiving Systems
  • Guest Scientist
Thumb ticker sm prof pic
Perceiving Systems
  • Research Scientist
Thumb xxl interdyn

Predicting the dynamics of interacting objects is essential for both humans and intelligent systems. However, existing approaches are limited to simplified, toy settings and lack generalizability to complex, real-world environments. Recent advances in generative models have enabled the prediction of state transitions based on interventions, but focus on generating a single future state which neglects the continuous dynamics resulting from the interaction. To address this gap, we propose InterDyn, a novel framework that generates videos of interactive dynamics given an initial frame and a control signal encoding the motion of a driving object or actor. Our key insight is that large video generation models can act as both neural renderers and implicit physics ``simulators'', having learned interactive dynamics from large-scale video data. To effectively harness this capability, we introduce an interactive control mechanism that conditions the video generation process on the motion of the driving entity. Qualitative results demonstrate that InterDyn generates plausible, temporally consistent videos of complex object interactions while generalizing to unseen objects. Quantitative evaluations show that InterDyn outperforms baselines that focus on static state transitions. This work highlights the potential of leveraging video generative models as implicit physics engines

Author(s): Rick Akkerman and Haiwen Feng and Michael J. Black and Dimitrios Tzionas and Victoria Fernández Abrevaya
Links:
Book Title: IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR)
Year: 2025
Month: June
Day: 13
Bibtex Type: Conference Paper (inproceedings)
Event Place: Nashville, TN
State: Published

BibTex

@inproceedings{InterDyn:2025,
  title = {{InterDyn}: Controllable Interactive Dynamics with Video Diffusion Models},
  booktitle = {IEEE/CVF Conf.~on Computer Vision and Pattern Recognition (CVPR)},
  abstract = {Predicting the dynamics of interacting objects is essential for both humans and intelligent systems. However, existing approaches are limited to simplified, toy settings and lack generalizability to complex, real-world environments. Recent advances in generative models have enabled the prediction of state transitions based on interventions, but focus on generating a single future state which neglects the continuous dynamics resulting from the interaction. To address this gap, we propose InterDyn, a novel framework that generates videos of interactive dynamics given an initial frame and a control signal encoding the motion of a driving object or actor. Our key insight is that large video generation models can act as both neural renderers and implicit physics ``simulators'', having learned interactive dynamics from large-scale video data. To effectively harness this capability, we introduce an interactive control mechanism that conditions the video generation process on the motion of the driving entity. Qualitative results demonstrate that InterDyn generates plausible, temporally consistent videos of complex object interactions while generalizing to unseen objects. Quantitative evaluations show that InterDyn outperforms baselines that focus on static state transitions. This work highlights the potential of leveraging video generative models as implicit physics engines},
  month = jun,
  year = {2025},
  slug = {interdyn-2025},
  author = {Akkerman, Rick and Feng, Haiwen and Black, Michael J. and Tzionas, Dimitrios and Abrevaya, Victoria Fernández},
  month_numeric = {6}
}