Abstract
General scene reconstruction refers to the task of estimating the full 3D geometry and texture of a scene containing previously unseen objects. In many practical applications such as AR/VR, autonomous navigation, and robotics, only a single view of the scene may be available, making the scene reconstruction task challenging. In this paper, we present a method for scene reconstruction by structurally breaking the problem into two steps: rendering novel views via inpainting and 2D to 3D scene lifting. Specifically, we leverage the generalization capability of large visual language models (DALL•E 2) to inpaint the missing areas of scene color images rendered from different views. Next, we lift these inpainted images to 3D by predicting normals of the inpainted image and solving for the missing depth values. By predicting for normals instead of depth directly, our method allows for robustness to changes in depth distributions and scale. With rigorous quantitative evaluation, we show that our method outperforms multiple baselines while providing generalization to novel objects and scenes. Code and data is available at https://samsunglabs.github.io/RIC-project-page/.
Original language | English (US) |
---|---|
Title of host publication | 2024 IEEE International Conference on Robotics and Automation, ICRA 2024 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 2713-2720 |
Number of pages | 8 |
ISBN (Electronic) | 9798350384574 |
DOIs | |
State | Published - 2024 |
Event | 2024 IEEE International Conference on Robotics and Automation, ICRA 2024 - Yokohama, Japan Duration: May 13 2024 → May 17 2024 |
Publication series
Name | Proceedings - IEEE International Conference on Robotics and Automation |
---|---|
ISSN (Print) | 1050-4729 |
Conference
Conference | 2024 IEEE International Conference on Robotics and Automation, ICRA 2024 |
---|---|
Country/Territory | Japan |
City | Yokohama |
Period | 5/13/24 → 5/17/24 |
Bibliographical note
Publisher Copyright:© 2024 IEEE.