Estimates of Temporal Edge Detection Filters in Human Vision

Edge detection is an important process in human visual processing. However, as far as we know, few attempts have been made to map the temporal edge detection filters in human vision. To that end, we devised a user study and collected data from which we derived estimates of human temporal edge detection filters based on three different models, including the derivative of the infinite symmetric exponential function and temporal contrast sensitivity function.

Qianli Ma

Qianli Ma is research scientist at NVIDIA Research. He received his PhD from ETH Zürich and Max-Planck-Institute for Intelligent Systems (Tübingen), advised by Professor Michael Black and Professor Siyu Tang. He has also interned at Meta Reality Labs in Pittsburgh. He has been developing new representations and methods for reconstructing, generating and modeling digital humans. His research interests span generative models, 3D computer vision and graphics, with a current focus on dynamic 3D content generation.

Global Context Vision Transformers

We propose global context vision transformer (GC ViT), a novel architecture that enhances parameter and compute utilization for computer vision. Our method leverages global context self-attention modules, joint with standard local self-attention, to effectively and efficiently model both long and short-range spatial interactions, without the need for expensive operations such as computing attention masks or shifting local windows. In addition, we address the lack of the inductive bias in ViTs, and propose to leverage a modified fused inverted residual blocks in our architecture.