Acta Geodaetica et Cartographica Sinica ›› 2016, Vol. 45 ›› Issue (11): 1328-1334.doi: 10.11947/j.AGCS.2016.20160046

Previous Articles     Next Articles

A Low-Sampling-Rate Trajectory Matching Algorithm in Combination of History Trajectory and Reinforcement Learning

SUN Wenbin, XIONG Ting   

  1. College of Geosciences and Surveying Engineering, China University of Mining and Technology(Beijing), Beijing 100083, China
  • Received:2016-02-01 Revised:2016-10-01 Online:2016-11-20 Published:2016-12-03
  • Supported by:
    The National Natural Science Foundation of China (No.41671383)

Abstract: In order to improve the accuracy of low frequency (sampling interval greater than 1 minute) trajectory data matching algorithm, this paper proposed a novel matching algorithm termed HMDP-Q (History Markov Decision Processes Q-learning). The new algorithm is based on reinforced learning on historic trajectory. First, we extract historic trajectory data according to incremental matching algorithm as historical reference, and filter the trajectory dataset through the historic reference, the shortest trajectory and the reachability. Then we model the map matching process as the Markov decision process, and build up reward function using deflected distance between trajectory points and historic trajectories. The largest reward value of the Markov decision process was calculated by using the reinforced learning algorithm, which is the optimal matching result of trajectory and road. Finally we calibrate the algorithm by utilizing city's floating cars data to experiment. The results show that this algorithm can improve the accuracy between trajectory data and road. The matching accuracy is 89.2% within 1 minute low-frequency sampling interval, and the matching accuracy is 61.4% when the sampling frequency is 16 minutes. Compared with IVVM (Interactive Voting-based Map Matching), HMDP-Q has a higher matching accuracy and computing efficiency. Especially, when the sampling frequency is 16 minutes, HMDP-Q improves the matching accuracy by 26%.

Key words: low-sampling-rate floating car data, trajectory matching, Markov decision process, reinforcement learning

CLC Number: