The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences
Publications Copernicus
Articles | Volume XL-4/W5
11 May 2015
 | 11 May 2015


T. Fuse and K. Matsumoto

Keywords: Sensor Fusion, Self-Localization, Mobile Devices, Augmented Reality, Visualization, Navigation

Abstract. Recently, development of high performance CPU, cameras and other sensors on mobile devices have been used for wide variety of applications. Most of the applications require self-localization of the mobile device. Since the self-localization is based on GPS, gyro sensor, acceleration meter and magnetic field sensor (called as POS) of low accuracy, the applications are limited. On the other hand, self-localization method using images have been developed, and the accuracy of the method is increasing. This paper develops the self-localization method by integrating sensors, such as POS and cameras, on mobile devices simultaneously. The proposed method mainly consists of two parts: one is the accuracy improvement of POS data filtering, and another is development of self-localization method by integrating POS and camera. The POS data filtering combines all POS data by using Kalman filter in order to improve the accuracy of exterior orientation factors. The exterior orientation factors with POS filtering are used as initial value of ones in image-based self-localization method. The image-based self-localization method consists of feature points extraction and tracking, relative orientation, coordinates estimation of the feature points, and orientation factors updates of the mobile device. The proposed method is applied to POS data and images taken in urban area. Through experiments with real data, the accuracy improvement by POS data filtering is confirmed. The proposed self-localization method with POS and camera make the accuracy more sophisticated by comparing with only POS data filtering.