Eyeing3D: Perceiving 3D from 2D Images

Tai Peng, Kang Yang Huang, Si Yu Lu, Rou An Chen, Jianlong Fu, Hong Han Shuai, Wen Huang Cheng

Research output: Contribution to journalConference articlepeer-review

Abstract

The recent vision foundation models, e.g. Segment Anything Model (SAM), have shown great potential in various downstream 2D tasks. However, their adaptability to 3D vision remains an unexplored area. In this paper, we propose a novel generative framework, namely Eyeing3D, by integrating generative vision models of multiple purposes (including SAM and Neural Radiance Fields) to achieve human's uncanny capability to perceive and interpret the 3D structure of a visual object, even when it is represented in a single 2D image. Particularly, a user is granted the ability to select any visual object of interest in the input 2D image with a simple click or bounding box, facilitating the reconstruction of its 3D model, with the added ability to manipulate the visual style and viewing angle. In the experiments, the effectiveness of our proposed Eyeing3D is demonstrated, showcasing improved performance in image-based 3D reconstruction tasks.

Original languageEnglish
Pages (from-to)120-121
Number of pages2
JournalIET Conference Proceedings
Volume2023
Issue number35
DOIs
StatePublished - 2023
Event2023 IET International Conference on Engineering Technologies and Applications, ICETA 2023 - Yunlin, Taiwan
Duration: 21 Oct 202323 Oct 2023

Keywords

  • 3D Reconstruction
  • Generative AI
  • Segment anything
  • Style manipulation
  • Vision foundation models

Fingerprint

Dive into the research topics of 'Eyeing3D: Perceiving 3D from 2D Images'. Together they form a unique fingerprint.

Cite this