πŸ“ ECCV 2026 Workshop

Human Motion-Informed
World Models

Bridging human motion modeling, visual world models, and embodied AI for socially intelligent perception and action.

πŸƒ Motion Modeling 🌍 World Models πŸ€– Embodied AI πŸš— Autonomous Systems
About the Workshop

Human motion and activity provide a critical signal for understanding, predicting, and interacting with dynamic environments. In recent years, computer vision has made significant progress in human motion perception, activity understanding, and motion generation, providing a strong foundation for modeling human behavior and dynamics. Integrating these advances into visual world models that support prediction, planning, and decision-making is a key next step toward enabling embodied intelligent systems to reason and act effectively in human-populated scenes.

This workshop focuses on the challenge of integrating rich models of human motion and behavior into world models, an increasingly important yet still under-explored direction in visual world modeling. Topics include: 1. modeling human motion and activity in complex, interactive scenes; 2. learning dynamic world models that incorporate human behavior to represent scenes, objects, and affordances; 3. enabling efficient and robust real-world deployment, including integrated perception-planning, safe navigation and autonomous driving, and improved generalization under noise, occlusions, and distribution shifts.

This topic is closely aligned with recent progress in visual world modeling and generative simulation which aim to capture vision-based representations of the scene structure, object relations, and human dynamics. The workshop will bring together research on human motion and activity modeling, dynamic scene understanding, and world models that explicitly account for human behavior as a central component of the environment. By uniting perspectives from computer vision, embodied AI, robotics, and graphics, the workshop provides a forum to explore human-centered world models that enable socially intelligent perception and action in applications such as dynamic scene understanding, predictive navigation, autonomous driving, and human-robot interaction.

Topics of Interest
πŸƒβ€β™‚οΈ

Human Motion & Activity Modeling

  • Trajectory, pose, mesh, and flow representations of human motion
  • Vision-based motion perception, tracking, and forecasting
  • Human motion generation and digital human modeling
🌐

Dynamic World Models

  • Generative and predictive world models
  • Visual world models incorporating human motion and behavior
  • Scene, object, and affordance modeling conditioned on human motion
  • Scene representations for dynamic environments
πŸš€

Efficient and Robust Real-World Deployment

  • Integrated perception and planning in human motion-informed models
  • Design choices under computational and real-time constraints
  • Robustness to noise, occlusions, and distribution shifts
  • Generalization across environments, agents, and human behaviors
Keynote Speakers
πŸ‘€

Gerard Pons-Moll

Professor, University of TΓΌbingen
Human motion, 3D tracking, virtual humans
πŸ‘€

Kai O. Arras

Professor, University of Stuttgart
Embodied intelligence, social robotics
πŸ‘€

Matthieu Cord

Professor, Sorbonne University / Director, valeo.ai
World models, vision-language models
πŸ‘€

Lingni Ma

Research Scientist, Meta Reality Labs
3D scene modeling, digital humans
πŸ‘€

Yuji Yasui

Executive Chief Engineer, Honda R&D
Autonomous driving, human-machine collaboration
Workshop Schedule

Half-day program

09:00

Welcome & Opening Remarks

09:10

Invited Keynote Talk 1

09:40

Invited Keynote Talk 2

10:10

Invited Keynote Talk 3

10:40

Coffee Break & Poster Session

11:10

Invited Keynote Talk 4

11:40

Workshop Best Paper Presentation & Award

Selected paper oral presentation

11:55

Invited Keynote Talk 5

12:25

Panel Discussion

Open Q&A with speakers

12:55

Closing Remarks

Organizers
PhD Student Organizers
πŸ‘©β€πŸ”¬

Yasamin Borhani

EPFL, Switzerland

3D localization & trajectory forecasting

πŸ‘©β€πŸ’»

Mariam Hassan

EPFL, Switzerland

World models for autonomous driving

πŸ§‘β€πŸ”¬

Yang Gao

EPFL, Switzerland

Motion prediction & 3D perception

πŸ‘©β€πŸŽ“

Yi Yang

KTH / Scania, Sweden

Prediction & planning in traffic

πŸ‘©β€πŸ”¬

Yufei Zhu

Γ–rebro University, Sweden

Probabilistic human motion patterns

Senior Organizers
πŸ§‘β€πŸ”¬

Andrey Rudenko

TU Munich, Germany

Motion prediction & HRI

πŸ‘©β€πŸ”¬

Wanyu Ma

The Chinese University of Hong Kong, China

HRI & robotic manipulation

πŸ‘¨β€πŸ«

Achim J. Lilienthal

TU Munich, Germany

Multi-modal perception & navigation

πŸ‘¨β€πŸ«

Martin Magnusson

Γ–rebro University, Sweden

3D mapping & autonomous systems

πŸ§‘β€πŸ’»

Chuan Guo

Meta Reality Labs, USA

Generative AI for digital humans

πŸ§‘β€πŸ”¬

NarΕ«nas VaΕ‘kevicius

Bosch Center for Artificial Intelligence, Germany

Dynamic perception & 3D scene graphs

πŸ§‘β€πŸ’Ό

Luigi Palmieri

Bosch Center for Artificial Intelligence, Germany

Predictive navigation & RL

Important Dates

Exact dates will be announced soon

Paper Submission
TBA
Notification
TBA
Camera-Ready
TBA
Workshop Day
ECCV 2026