Currently, Pedestrian Dead Reckoning (PDR) systems are becoming an important tool in indoor navigation. This is mainly due to the development of affordable and portable Micro Electro-Mechanical Systems (MEMS) on smartphones and decreased requirement of additional infrastructures in indoor areas. The main drawback to this technology remains the problem of drift accumulation and the need for support from external positioning systems. Vision-aided inertial navigation is one possible solution to that problem. This solution has become more popular in indoor localization with improved satisfaction compared to individual PDR system. Previous studies have used fixed platforms and visual tracking employed feature-extraction-based methods. This paper proposes a distributed implementation of positioning system and uses deep learning for visual tracking. Meanwhile, as both inertial navigation and optical systems can provide only relative positioning information, this paper proposes a method to integrate digital maps with real geographical coordinates to supply absolute location. This hybrid system has been tested on two common operation systems of smartphones using iOS and Android systems, based on corresponding data collection apps respectively, in order to test the robustness of method. It also uses two different methods for calibration, by time synchronization of positions and heading calibration based on time steps. Results demonstrate that localization information collected from both operating systems can be significantly improved after integrating with visual tracking data.