Data-Driven AI for Robotics (DAIR)

Welcome to the homepage of NVIDIA’s Data-Driven AI for Robotics (DAIR) group, led by Umar Iqbal. We are part of the Learning and Perception Research (LPR) organization within NVIDIA Research.

Our group investigates how robots can learn directly from human data, such as videos, motion capture, and large-scale demonstrations, to acquire skills that generalize across tasks, embodiments, and environments. We work at the intersection of computer vision, machine learning, and robotics, developing models that understand, reconstruct, and imitate human behaviors.

Our research contributes to NVIDIA’s broader vision of foundation models for robotics, combining advances in human motion modeling, human–object and human–scene interaction modeling, physics-based simulation, and embodied intelligence to enable scalable robot learning. Ultimately, we aim to bridge the gap between human understanding and robotic intelligence, advancing the goal of robots that learn by watching humans.

News

October 2025 - DAIR is now Data-Driven AI for Robotics.

July 2025 - Four papers accepted to ICCV 2025 including GENMO, GeoMan, AdaHuman, and HumanOLAT.

Feb 2025 - SimAvatar accepted to CVPR 2025!

Feb 2025 - The DAIR website goes live!!

Members

Avatar

Jiefeng Li

Computer Vision, Machine Learning, Generative AI

Avatar

Umar Iqbal

Team Leader

Avatar

Xueting Li

Computer Vision, 3D Vision

Avatar

Ye Yuan

3D Vision, Embodied AI, Reinforcement Learning

Current and Past Interns

.js-id-Intern

Publications

GENMO: A GENeralist Model for Human MOtion