GEM A GENeralist Model for Human MOtion
Kimodo Scaling Controllable Human Motion Generation
Data-Driven AI for Robotics DAIR @ NVIDIA Research

Welcome to the homepage of NVIDIA’s Data-Driven AI for Robotics (DAIR) group, led by Umar Iqbal. We are part of the Learning and Perception Research (LPR) organization within NVIDIA Research.

Our group investigates how robots can learn directly from human data, such as videos, motion capture, and large-scale demonstrations, to acquire skills that generalize across tasks, embodiments, and environments. We work at the intersection of computer vision, machine learning, and robotics, developing models that understand, reconstruct, and imitate human behaviors.

Our research contributes to NVIDIA’s broader vision of foundation models for robotics, combining advances in human motion modeling, human–object and human–scene interaction modeling, physics-based simulation, and embodied intelligence to enable scalable robot learning. Ultimately, we aim to bridge the gap between human understanding and robotic intelligence, advancing the goal of robots that learn by watching humans.

News

March 2026
We released a whole new ecosystem for Human(oid) Motion including SOMA, Kimodo, GEM, SOMA-Retargeter and, BONES-SEED dataset.
December 2025
We released SONIC, a state-of-the-art generalist humanoid controller.
July 2025
Four papers accepted to ICCV 2025 including GENMO, GeoMan, AdaHuman, and HumanOLAT.
Feb 2025
SimAvatar accepted to CVPR 2025!

Members

Avatar

Umar Iqbal

Team Leader

Avatar

Jiefeng Li

Computer Vision, Machine Learning, Generative AI

Avatar

Ye Yuan

3D Vision, Embodied AI, Reinforcement Learning

Avatar

Xueting Li

Computer Vision, 3D Vision

Avatar

Yufei Ye

3D Vision, Robotics, Human Motion

Avatar

Jinhyung (David) Park

Emboddied Intelligence, 3D Scene Understanding, Human Modeling

Avatar

Tianyi Xie

Computer Graphics, Generative AI, 3D Vision

Publications

GEM: A GENeralist Model for Human MOtion