Esports and expertise: what competitive gaming can teach us about mastery

Historically, much research and development in human computer interaction has focused on atomic and generalizable tasks, where task completion time indicates productivity. However, the emergence of competitive games and esports reminds us of an alternative perspective on human performance in HCI: mastery of higher-level, holistic practices. Just as a world-renowned artist is rarely evaluated for their individual brush strokes, so skilled competitive gamers rarely succeed solely by completing individual mouse movements or keystrokes as quickly as possible.

Szu-Wei Fu

Szu-Wei Fu joined NVIDIA Research in November 2022. His current interests include ML-based audio-visual processing, speech processing/enhancement, and quality estimation. Before Joined NVIDIA, he was an applied scientist in Microsoft.

Mingjie Liu

Mingjie Liu is currently a Research Scientist at NVIDIA, where he actively conduct research on Electronic Design Automation. He received his PhD degree in electrical and computer engineering from the The University of Texas at Austin in 2022. His research interest include applied machine learning for design automation and design automation for analog and mixed-signal integrated circuits.

Min-Hung Chen

Min-Hung (Steve) Chen is a Senior Research Scientist at NVIDIA Research Taiwan, working on Vision+X Multi-Modal AI. He received his Ph.D. degree from Georgia Tech, advised by Prof. Ghassan AlRegib and in collaboration with Prof. Zsolt Kira.

Beyond CPO: A Motivation and Approach for Bringing Optics onto the Silicon Interposer

Co-packaged optics (CPO) technology is well positioned to break through the bottlenecks that impede efficient bandwidth scaling in key near-term commercial integrated circuits. In this paper, we begin by providing some historical context for this important sea change in the optical communications industry. Then, motivated by GPU-based accelerated computing requirements, we investigate the next pain points that are poised to constrain bandwidth and efficiency in future CPO-based systems.

Elucidating the Design Space of Diffusion-Based Generative Models

We argue that the theory and practice of diffusion-based generative models are currently unnecessarily convoluted and seek to remedy the situation by presenting a design space that clearly separates the concrete design choices. This lets us identify several changes to both the sampling and training processes, as well as preconditioning of the score networks. Together, our improvements yield new state-of-the-art FID of 1.79 for CIFAR-10 in a class-conditional setting and 1.97 in an unconditional setting, with much faster sampling (35 network evaluations per image) than prior designs.

Robust Trajectory Prediction against Adversarial Attacks

Trajectory prediction using deep neural networks (DNNs) is an essential component of autonomous driving (AD) systems. However, these methods are vulnerable to adversarial attacks, leading to serious consequences such as collisions. In this work, we identify two key ingredients to defend trajectory prediction models against adversarial attacks including (1) designing effective adversarial training methods and (2) adding domain-specific data augmentation to mitigate the performance degradation on clean data.

DiffStack: A Differentiable and Modular Control Stack for Autonomous Vehicles

Autonomous vehicle (AV) stacks are typically built in a modular fashion, with explicit components performing detection, tracking, prediction, planning, control, etc. While modularity improves reusability, interpretability, and generalizability, it also suffers from compounding errors, information bottlenecks, and integration challenges. To overcome these challenges, a prominent approach is to convert the AV stack into an end-to-end neural network and train it with data.

Task-Relevant Failure Detection for Trajectory Predictors in Autonomous Vehicles

In modern autonomy stacks, prediction modules are paramount to planning motions in the presence of other mobile agents. However, failures in prediction modules can mislead the downstream planner into making unsafe decisions. Indeed, the high uncertainty inherent to the task of trajectory forecasting ensures that such mispredictions occur frequently. Motivated by the need to improve safety of autonomous vehicles without compromising on their performance, we develop a probabilistic run-time monitor that detects when a harmful prediction failure occurs, i.e., a task-relevant failure detector.