Accurate camera relocalisation is a fundamental technology for extended reality (XR), facilitating the seamless integration and persistence of digital content within the real world. Benchmark datasets that measure camera pose accuracy have driven progress in visual re-localisation research. Despite notable progress in this field, there is a limited availability of datasets incorporating Visual Inertial Odometry (VIO) data from typical mobile AR frameworks such as ARKit or ARCore. This paper presents a new dataset, MARViN, comprising diverse indoor and outdoor scenes captured using het- erogeneous mobile consumer devices. The dataset includes camera images, ARCore or ARKit VIO data, and raw sensor data for several mobile devices, together with the corresponding ground-truth poses. MARViN allows us to demonstrate the capability of ARKit and ARCore to provide relative pose estimates that closely approximate ground truth within a short timeframe. Subsequently, we evaluate the performance of mobile VIO data in enhancing absolute pose estimations in both desktop simulation and user study.
@INPROCEEDINGS{10536574,
author={Liu, Changkun and Zhao, Yukun and Braud, Tristan},
booktitle={2024 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)},
title={MARViN: Mobile AR Dataset with Visual-Inertial Data},
year={2024},
volume={},
number={},
pages={532-538},
keywords={Performance evaluation;Visualization;Solid modeling;Three-dimensional displays;Pose estimation;User interfaces;Cameras;Visual localisation Dataset—Camera pose regression—Visual-Inertial Odometry—Visual positioning system;Computer Vision—Augmented Reality},
doi={10.1109/VRW62533.2024.00103}}