Accurate reconstruction of indoor environments is crucial for applications in augmented reality, virtual reality, and robotics. However, existing indoor datasets are often limited in scale, lack ground-truth point clouds, and provide insufficient viewpoints, which impedes the development of robust novel view synthesis (NVS) techniques. To address these limitations, we introduce a new large-scale indoor dataset that features diverse and challenging scenes, including basements and long corridors. This dataset offers panoramic image sequences for comprehensive coverage, high-resolution point clouds, meshes, and textures as ground truth, and a novel benchmark specifically designed to evaluate NVS algorithms in complex indoor environments. Our dataset and benchmark aim to advance indoor scene reconstruction and facilitate the creation of more effective NVS solutions for real-world applications.
The IVGM dataset encompasses a diverse array of environments, meticulously captured by our custom-designed data acquisition vehicle across three distinct scenes. This collection includes two segments from school office floors and one scene from underground garages. It provides:
Sequence Name | Area Size(m2) | Point Number | Insta Images | Titan Images |
---|---|---|---|---|
Office Area1 | 2,989.63 | 76,488,066 | 1,610 | 12,872 |
Office Area2 | 2,651.00 | 86,233,513 | 2,669 | 21,608 |
Underground Garage | 3,797.11 | 153,185,271 | 1,816 | 14,528 |
To evaluate the applicability, versatility, and performance of our dataset on novel view synthesis algorithms, we tested several popular methods developed in recent years.
Key results show: