The increasing demand for virtual reality applications has highlighted the significance of crafting immersive 3D assets.
We present a text-to-3D 360
Beginning with a concise text prompt, we employ a diffusion model to generate a 360
Unlike previous works that generate 3D scenes through a time-consuming score distillation process or progressive inpainting, our work uses Panoramic Gaussian Splatting to achieve ultra-fast “one-click” 3D scene generation. We integrate GPT-4V to facilitate iterative user input refinement and quality evaluation during the generation process, a feature that was challenging to implement in previous methods due to the absence of global 2D image representations.
In each row, from left to right, displays novel views as the camera undergoes clockwise rotation in yaw,
accompanied by slight random rotations in pitch and random translations.
LucidDreamer hallucinates novel views from a conditioned image (indicated by a red bounding box) but lacks global semantic,
stylized, and geometric consistency. In contract, our method provides complete 360
Our generated 3D scenes are diverse in style, consistent in geometry, and highly matched with the simple text inputs.
@inproceedings{zhou2024dreamscene360,
title={Dreamscene360: Unconstrained text-to-3d scene generation with panoramic gaussian splatting},
author={Zhou, Shijie and Fan, Zhiwen and Xu, Dejia and Chang, Haoran and Chari, Pradyumna and Bharadwaj, Tejas and You, Suya and Wang, Zhangyang and Kadambi, Achuta},
booktitle={European Conference on Computer Vision},
pages={324--342},
year={2024},
organization={Springer}
}