Ali Hatamizadeh

Ali Hatamizadeh is a senior research scientist at NVIDIA Research. His research focuses on Large Language Models (LLMs), with an emphasis on the principled design of LLMs and advancing their reasoning capabilities. He received his PhD and MSc in Computer Science from the University of California, Los Angeles (UCLA), where he was honored with the 2018 UCLA Henry Samueli School of Engineering and Applied Science Edward K. Rice Outstanding Masters Student Award.

Vishwesh Nath

Dr. Vishwesh Nath is an Applied Research Scientist at Nvidia. He works with the Clara DLMED Research Team and his research is focused towards medical imaging with sub-domains including AI-Assisted Annotation (DeepGrow 2D & 3D), Neural Architecture Search and Federated Learning.

Yoshi Nishi

Yoshinori Nishi received the B.S. and M.S. degrees in low-temperature physics from Waseda University, Tokyo, Japan, in 1997 and 1999, respectively. From 1999 to 2003, he was with NTT Electronics, Inc., Atsugi, Japan, as a member of the High-Speed Device Development Group, where he was a Chief Designer for the 50-Gb/s InP HEMT logic family, first 50-Gb/s product in the market in 2001.

Accelerating Chip Design with Machine Learning

Recent advancements in machine learning provide an opportunity to transform chip design workflows. We review recent research applying techniques such as deep convolutional neural networks and graph-based neural networks in the areas of automatic design space exploration, power analysis, VLSI physical design, and analog design. We also present a future vision of an AI-assisted automated chip design workflow to aid designer productivity and automate optimization tasks.

DREAMPlace: Deep Learning Toolkit-Enabled GPU Acceleration for Modern VLSI Placement

Placement for very-large-scale integrated (VLSI) circuits is one of the most important steps for design closure. We propose a novel GPU-accelerated placement framework DREAMPlace, by casting the analytical placement problem equivalently to training a neural network. Implemented on top of a widely-adopted deep learning toolkit PyTorch, with customized key kernels for wirelength and density computations, DREAMPlace can achieve around 40× speedup in global placement without quality degradation compared to the state-of-the-art multi-threaded placer RePlAce.

GRANNITE: Graph Neural Network Inference for Transferable Power Estimation

This paper introduces GRANNITE, a GPU-accelerated novel graph neural network (GNN) model for fast, accurate, and transferable vector-based average power estimation. During training, GRANNITE learns how to propagate average toggle rates through combinational logic: a netlist is represented as a graph, register states and unit inputs from RTL simulation are used as features, and combinational gate toggle rates are used as labels. A trained GNN model can then infer average toggle rates on a new workload of interest or new netlists from RTL simulation results in a few seconds.