Abstract
The ability to reconstruct immersive and realistic three-dimensional scenes plays a fundamental role in advancing virtual reality, digital twins, and related fields. With the rapid development of differentiable rendering frameworks, the reconstruction quality of static scenes has been significantly improved. However, we observe that the challenge of insufficient initialization has been largely overlooked in existing studies, while at the same time heavily relying on dense multi-view imagery that is difficult to obtain. To address these challenges, we propose a pipeline for text driven 3D scene generation, which employs panoramic images as an intermediate representation and integrates with 3D Gaussian Splatting to enhance reconstruction quality and efficiency. Our method introduces an improved point cloud initialization using Fibonacci lattice sampling of panoramic images, combined with a dense perspective pseudo label strategy for teacher–student distillation supervision, enabling more accurate scene geometry and robust feature learning without requiring explicit multi-view ground truth. Extensive experiments validate the effectiveness of our method, consistently outperforming state-of-the-art methods across standard reconstruction metrics.
| Original language | English |
|---|---|
| Article number | 6840 |
| Journal | Sensors |
| Volume | 25 |
| Issue number | 22 |
| DOIs | |
| State | Published - Nov 2025 |
Bibliographical note
Publisher Copyright:© 2025 by the authors.
Keywords
- 3D reconstruction
- panoramic image
- point cloud initialization
ASJC Scopus subject areas
- Analytical Chemistry
- Information Systems
- Atomic and Molecular Physics, and Optics
- Biochemistry
- Instrumentation
- Electrical and Electronic Engineering