Abstract
With advances in extended reality technologies, there is an increasing demand for three-dimensional (3D)
scene reconstruction from captured videos. With the advent of Neural Radiance Fields, research on 3D reconstruction has
made rapid progress, leading to environments where 3D reconstruction can be easily achieved only with smartphones. However,
editing reconstructed 3D scenes or specific objects remains challenging and is still in its early stages. This study aims to
semiautomatically edit 3D scenes reconstructed by Gaussian Splatting. We propose a method to estimate and extract foreground
objects automatically from user input, and to cut them out from the scene. By applying to 3D scenes reconstructed from existing
trained scenes and captured videos, we confirmed the ability of the proposed method to perform visually plausible cutout of
foreground objects.