Super resolved time of flight imaging via FRI sampling theory

Ayash Bhandari, Andrew Michael Wallace, Ramesh Raskar

Research output: Chapter in Book/Report/Conference proceedingConference contribution

12 Citations (Scopus)

Abstract

Optical time-of-flight (ToF) sensors can measure scene depth accurately by projection and reception of an optical signal. The range to a surface in the path of the emitted signal is proportional to the delay time of the light echo or the reflected signal. In practice, a diverging beam may be subject to multi-echo backscatter, and all these echoes must be resolved to estimate the multiple depths. In this paper, we propose a method for super-resolution of optical ToF signals. Our contributions are twofold. Starting with a general image formation model common to most ToF sensors, we draw a striking analogy of ToF systems with sampling theory. Based on our model, we reformulate the ToF super-resolution problem as a parameter estimation problem pivoted around the finite-rate-of-innovation framework. In particular, we show that super-resolution of multi-echo backscattered signal amounts to recovery of Dirac impulses from low-pass measurements. Our theory is corroborated by analysis of data collected from a photon counting, LiDAR sensor, showing the effectiveness of our non-iterative and computationally efficient algorithm.
Original languageEnglish
Title of host publication2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
PublisherIEEE
Pages4009-4013
Number of pages5
ISBN (Print)9781479999880
DOIs
Publication statusPublished - Mar 2016

Fingerprint Dive into the research topics of 'Super resolved time of flight imaging via FRI sampling theory'. Together they form a unique fingerprint.

  • Cite this

    Bhandari, A., Wallace, A. M., & Raskar, R. (2016). Super resolved time of flight imaging via FRI sampling theory. In 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 4009-4013). IEEE. https://doi.org/10.1109/ICASSP.2016.7472430