Reference-Noise Compensation Scheme for Single- Ended Package-to-Package Links

We present a method for tracking and compensating the reference noise in single-ended links, enabling energy-efficient package-to-package communication in systems with noisy environment. The 25Gb/s/pin serial link with compensation loop is fabricated in 16nm FinFET process operating over PCB channel with >200mV reference noise. The maximum measured bandwidth of the compensation loop is 30MHz, and its energy and area penalty are negligible.

Yuval Atzmon

My main fields of interest are computer vision and machine learning, focusing on compositional generalization from few or zero examples and reasoning across different data modalities (like natural-language, images and action). Recently, I have focused on integrating these ideas, with the emergence of generative and discriminative vision-and-language foundation models.

I obtained my Ph.D. under the guidance of Prof. Gal Chechik at Bar-Ilan University. My M.Sc. and B.Sc. are in Electrical Engineering from the Technion, Israel Institute of Technology.

Neural Temporal Adaptive Sampling and Denoising

Despite recent advances in Monte Carlo path tracing at interactive rates, denoised image sequences generated with few samples per-pixel often yield temporally unstable results and loss of high-frequency details. We present a novel adaptive rendering method that increases temporal stability and image fidelity of low sample count path tracing by distributing samples via spatio-temporal joint optimization of sampling and denoising.

Self-Supervised Viewpoint Learning From Image Collections

Training deep neural networks to estimate the viewpoint of objects requires large labeled training datasets. However, manually labeling viewpoints is notoriously hard, error-prone, and time-consuming. On the other hand, it is relatively easy to mine many unlabeled images of an object category from the internet, e.g., of cars or faces. We seek to answer the research question of whether such unlabeled collections of in-the-wild images can be successfully utilized to train viewpoint estimation networks for general object categories purely via self-supervision.

Tucker Hermans

Tucker Hermans is a senior research scientist at NVIDIA working on robotic manipulation focusing on the interaction of multi-sensory perception, learning, and control. Tucker is also an assosciate professor in the School of Computing at the University of Utah where he is a member of the Utah Robotics Center. Tucker earned his Ph.D. in Robotics from the Georgia Institute of Technology.

Convolutional Tensor-Train LSTM for Spatio-Temporal Learning

Learning from spatio-temporal data has numerous applications such as human-behavior analysis, object tracking, video compression, and physics simulation. However, existing methods still perform poorly on challenging video tasks suchas long-term forecasting. The gap partially is because these kinds of challenging tasks require learning long-term spatio-temporal correlations in the video sequence. We propose a higher-order convolutional LSTM model that can efficiently learn these correlations with a succinct representation of the history.

Toward Sim-to-Real Directional Semantic Grasping

We address the problem of directional semantic grasping, that is, grasping a specific object from a specific direction. We approach the problem using deep reinforcement learning via a double deep Q-network (DDQN) that learns to map downsampled RGB input images from a wrist-mounted camera to Q-values, which are then translated into Cartesian robot control commands via the cross-entropy method (CEM). The network is learned entirely on simulated data generated by a custom robot simulator that models both physical reality (contacts) and perceptual quality (high-quality rendering).