Planar Abstraction and Inverse Rendering of 3D Indoor Environment
- Planar Abstraction and Inverse Rendering of 3D Indoor Environment
- 김익재; Young Min Kim; Sangwoo Ryu
- indoor modeling; texture generation; 3D contents generation; inverse rendering
- Issue Date
- IEEE transactions on visualization and computer graphics
- VOL 게재예정, 게재예정
- Scanning and acquiring a 3D indoor environment suffers from complex occlusions and misalignment errors. The reconstruction obtained from an RGB-D scanner contains holes in geometry and ghosting in texture. These are easily noticeable and cannot be considered as visually compelling VR content without further processing. On the other hand, the well-known Manhattan World priors successfully recreate relatively simple structures. In this paper, we would like to push the limit of planar representation in indoor environments. Given an initial 3D reconstruction captured by an RGB-D sensor, we use planes not only to represent the environment geometrically but also to solve an inverse rendering problem considering texture and light. The complex process of shape inference and intrinsic imaging is greatly simplified with the help of detected planes and yet produces a realistic 3D indoor environment. The generated content can adequately represent the spatial arrangements for various AR/VR applications and can be readily composited with virtual objects possessing plausible lighting and texture.
- Appears in Collections:
- KIST Publication > Article
- Files in This Item:
There are no files associated with this item.
- RIS (EndNote)
- XLS (Excel)
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.