Three-dimensional scene reconstruction from a two-dimensional image
Abstract
We propose and simulate a method of reconstructing a three-dimensional scene from a two-dimensional image for developing and augmenting world models for autonomous navigation. This is an extension of the Perspective-n-Point (PnP) method which uses a sampling of the 3D scene, 2D image point parings, and Random Sampling Consensus (RANSAC) to infer the pose of the object and produce a 3D mesh of the original scene. Using object recognition and segmentation, we simulate the implementation on a scene of 3D objects with an eye to implementation on embeddable hardware. The final solution will be deployed on the NVIDIA Tegra platform.
Publication Title
Proceedings of SPIE - The International Society for Optical Engineering
Recommended Citation
Parkins, F., & Jacobs, E. (2017). Three-dimensional scene reconstruction from a two-dimensional image. Proceedings of SPIE - The International Society for Optical Engineering https://doi.org/10.1117/12.2266411