High-Fidelity Depth Map Reconstruction System With RGB-Guided Super Resolution CNN and Cross-Calibrated Chaos LiDAR
High-fidelity 3D models are essential for immersive virtual and augmented reality (VR/AR) applications. However, the performance of current 3D recording devices is limited in several scenarios, such as dim light environments, long-distance measurements, and large-scale objects. Therefore, their appl...
Saved in:
Main Authors: | , , , , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2025-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10849544/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | High-fidelity 3D models are essential for immersive virtual and augmented reality (VR/AR) applications. However, the performance of current 3D recording devices is limited in several scenarios, such as dim light environments, long-distance measurements, and large-scale objects. Therefore, their applicability to indoor scenes is hindered. In this work, we propose a depth map reconstruction system that integrates an RGB-guided depth map super-resolution convolutional neural network (CNN) into a stand-alone Chaos LiDAR depth sensor. This system provides highly accurate depth estimates in various scenarios, particularly for indoor scenes with dim lighting or long distances ranging from 4 m to 6 m. We address two design challenges to maximize the quality of the reconstructed depth map of the system. First, the misalignment across RGB-depth sensors is addressed using a two-stage calibration pipeline. Second, the lack of large-scale real-world LiDAR datasets is addressed by generating a large-scale synthetic dataset and adopting transfer learning. Experimental results show that our proposed system significantly outperforms the commercial RGB-D recording device RealSense D435i in terms of subjective visual perception, precision, and density of depth estimates, making it a promising solution for general indoor scene recording. |
---|---|
ISSN: | 2169-3536 |