Scene motion, multiple reflections, and sensor noise introduce artifacts in the depth reconstruction performed by time-of-flight (ToF) cameras.
We propose a two-stage, deep-learning approach to address all of these sources of artifacts simultaneously.
We also introduce FLAT, a synthetic dataset of 2000 ToF measurements that capture all of these nonidealities, and allows to simulate different camera hardware. Using the Kinect 2 camera as a baseline, we show improved reconstruction errors over state-of-the-art methods, on both simulated and real data.