Instant Neural Graphics Primitives with a Multiresolution Hash Encoding

Neural graphics primitives, parameterized by fully connected neural networks, can be costly to train and evaluate. We reduce this cost with a versatile new input encoding that permits the use of a smaller network without sacrificing quality, thus significantly reducing the number of floating point and memory access operations: a small neural network is augmented by a multiresolution hash table of trainable feature vectors whose values are optimized through stochastic gradient descent. The multiresolution structure allows the network to disambiguate hash collisions, making for

Saving PAM4 Bus Energy with SMOREs: Sparse Multi-level Opportunistic Restricted Encodings

Pulse Amplitude Modulation (PAM) uses multiple voltage levels as different data symbols, transferring multiple bits of data simultaneously, thereby enabling higher communication bandwidth without increased operating frequencies. However, dividing the voltage into more symbols leads to a smaller voltage difference between adjacent symbols, making the interface more vulnerable to crosstalk and power noise. GDDR6X adopts four-level symbols (PAM4) with Maximum Transition Avoidance (MTA) coding, which reduces the effects of crosstalk.

Slang Shading Language Advances

In this talk, Yong He, a Senior Researcher at NVIDIA, shares recent advances and new features in the Slang shading language.

Generic Lithography Modeling with Dual-band Optics-Inspired Neural Networks

Lithography simulation is a critical step in VLSI design and optimization for manufacturability. Existing solutions for highly accurate lithography simulation with rigorous models are computationally expensive and slow, even when equipped with various approximation techniques. Recently, machine learning has provided alternative solutions for lithography simulation tasks such as coarse-grained edge placement error regression and complete contour prediction. However, the impact of these learning-based methods has been limited due to restrictive usage scenarios or low simulation accuracy.

GATSPI: GPU Accelerated Gate-Level Simulation for Power Improvement

(DAC 2022 preprint)

In this paper, we present GATSPI, a novel GPU accelerated logic gate simulator that enables ultra-fast power estimation for industry sized ASIC designs with millions of gates. GATSPI is written in PyTorch with custom CUDA kernels for ease of coding and maintainability. It achieves simulation kernel speedup of up to 1668X on a single-GPU system and up to 7412X on a multiple-GPU system when compared to a commercial gate-level simulator running on a single CPU core.

Harini Muthukrishnan

Harini joined NVIDIA in February 2022 as a part of the System Architecture Research Group. Her research focuses on developing GPU and interconnect solutions for scalable multi-GPU systems. 

Yufan He

Yufan He joined NVIDIA as an applied research scientist in 2021. His current research interests are AutoML, computer vision, and medical image analysis.

Before that, Yufan received his M.S. and Ph.D. in ECE from the Johns Hopkins University in 2021 and B.S. in EE from Tsinghua University in 2016.

Seungjun Nah

Seungjun Nah is a research scientist at NVIDIA.

He received his Ph.D. in Electrical and Computer Engineering at Seoul National University, advised by Kyoung Mu Lee.

His research interests include computer vision with focus on high-quality image generation and the fast execution of such algorithms. 

For more information, please visit: https://seungjunnah.github.io/