NYC-Indoor-VPR Dataset Website

This is the official website for paper NYC-Indoor-VPR

Introduction


Our dataset is named NYC-Indoor-VPR. It is composed of images recorded in New York City from April 2022 to April 2023. Footage was captured using hand-held Insta360 one x2 spherical cameras, generating videos with a resolution of 1920x960. On the basis of raw images, we use MSeg, a semantic segmentation method, to replace moving objects such as people and cars with white pixels. Fig. 1 compares anonymized and raw images.

non_anony
Fig. 1 - Illustration of raw images and anonymized images

We recorded images of 13 different floors/scenes within the six buildings. We chose buildings with varied utilities and appearances: the Oculus, New York University Silver Center for Arts and Science, Elmer Holmes Bobst Library, Morton Williams Supermarket, and Metropolitan Museum of Art. These settings represent a broad range of indoor spaces, including shopping malls, teaching buildings, libraries, supermarkets, and museums. Fig. 2 shows the trajectories and example images of certain scenes.

data_vis
Fig. 2 - Trajectories annotated by our semi-automatic method and example images of 12 scenes in NYC-Indoor-VPR.

For each building, we selected one or multiple floors as scenes. For each scene, we fixed the trajectory and captured videos along the same route at different times throughout the year. Fig. 3 shows the time distribution of visits. The videos were recorded from April to July 2022 and from March to April 2023. Therefore, it contains various changes in illumination and appearance. As shown in Fig. 4, we can see image changes at the same location over a year.

challenge_case
Fig. 3
time_freq
Fig. 4

Detail Description



01

Dataset Detail

dataset_detail

02

Uniqueness

Our dataset stands out in two ways. First, NYC-Indoor-VPR images were captured in buildings such as The Oculus and the Bobst Library, which typically have a large flow of pedestrians. We anonymized these pedestrians in the images to reduce their exposure to personally identifiable information. These anonymized images not only enhance data privacy but also allow VPR algorithms to focus more on invariant or environmental features rather than transient features, such as moving people. Second, NYC-Indoor-VPR spans a year and includes images captured in buildings that undergo significant visual changes over time. For instance, goods in the supermarket vary and storefronts in the shopping mall are subject to change. This variability in the dataset allows us to test the performance of the VPR algorithms with fewer invariant features in the images.

Download


Files are organized as a zip file contains images. The image name includes annotated topometric location and its index.

Hugging Face Link: Please Click Here

Below is the Github repo of code used to perform experiments described in the paper.

Code Link: Please Click Here

Contact Us


Chen Feng - cfeng@nyu.edu

Diwei Sheng - ds5725@nyu.edu