AuraFusion360 icon

Augmented Unseen Region Alignment for Reference-based 360° Unbounded Scene Inpainting

arXiv 2025

1National Yang Ming Chiao Tung University, 2NVIDIA
Teaser

AuraFusion360 can perform inpainting in a 360° unbounded scene..

Inpainting Results of Our 360-USID Dataset

Inpainting Results of Other Collected 360 Scene.

News

[2025.2.10] icon Release project page, arXiv paper, dataset, and evaluation results!

Abstract

Three-dimensional scene inpainting is crucial for applications from virtual reality to architectural visualization, yet existing methods struggle with view consistency and geometric accuracy in 360° unbounded scenes. We present AuraFusion360, a novel reference-based method that enables high-quality object removal and hole filling in 3D scenes represented by Gaussian Splatting. Our approach introduces (1) depth-aware unseen mask generation for accurate occlusion identification, (2) Adaptive Guided Depth Diffusion, a zero-shot method for accurate initial point placement without requiring additional training, and (3) SDEdit-based detail enhancement for multi-view coherence. We also introduce 360-USID, the first comprehensive dataset for 360° unbounded scene inpainting with ground truth. Extensive experiments demonstrate that AuraFusion360 significantly outperforms existing methods, achieving superior perceptual quality while maintaining geometric accuracy across dramatic viewpoint changes.

Pipeline

Pipeline

Our approach takes multi-view RGB images and corresponding object masks as input and outputs a Gaussian representation with the masked objects removed. The pipeline consists of three main stages: (a) Depth-Aware Unseen Masks Generation to identify truly occluded areas, referred to as the “unseen region”, (b) Depth-Aligned Gaussian Initialization on Reference View to fill unseen regions with initialized Gaussian containing reference RGB information after object removal, and (c) SDEdit-Based RGB Guidance for Detail Enhancement, which enhances fine details using an inpainting model while preserving reference view information. Instead of applying SDEdit with random noise, we use DDIM Inversion on the rendered initial Gaussians to generate noise that retains the structure of the reference view, ensuring multi-view consistency across all RGB Guidance.

Depth-Aware Unseen Mask Generation

Unseen

Our depth-aware unseen mask generation method operates in two steps. First, it generates a depth-aware unseen contour through depth warping to identify truly occluded regions. Then, it uses this contour as a bounding box prompt for SAM2 to segment the unseen region.

Adaptive Guided Depth Diffusion (AGDD)

Agdd

Our AGDD is a zero-shot method that effectively aligns the estimated monocular depth with the existing geometric structure without requiring additional training.

Visualization Results

AuraFusion360 outperforms other methods in unbounded 360° scene inpainting.

Inpainting Results Unseen Mask Depth Initilaized Points

3 Try selecting different methods and scenes!

360USID/carton 360USID/cone 360USID/skateboard 360USID/newcone 360USID/sunflower 360USID/plant 360USID/cookie
Other360/kitchen Other360/bear Other360/bonsai Other360/room Other360/vasedeck Other360/pinecone

Quantitative Results

Quantitative comparison of 360° inpainting methods on our 360-USID dataset. Red text indicates the best, and Blue text indicates the second-best performing method.

Quantitative

Citation