Acta Geodaetica et Cartographica Sinica ›› 2021, Vol. 50 ›› Issue (11): 1558-1573.doi: 10.11947/j.AGCS.2021.20210246

• Environment Perception for Intelligent Driving • Previous Articles     Next Articles

A joint network of point cloud and multiple views for roadside objects recognition from mobile laser point clouds

FANG Lina1, SHEN Guixi1, YOU Zhilong1, GUO Yingya2, FU Huasheng3, ZHAO Zhiyuan1, CHEN Chongcheng1   

  1. 1. Academy of Digital China(Fujian), Fuzhou University, Fuzhou 350002, China;
    2. College of Computer and Data Science, Fuzhou University, Fuzhou 350002, China;
    3. Fujian Provincial Investigation, Design & Research Institute of Water Conservancy & Hydropower, Fuzhou 350002, China
  • Received:2021-05-11 Revised:2021-09-27 Published:2021-12-07
  • Supported by:
    The National Natural Science Foundation of China (Nos. 42071446;61801121);Foreign Cooperation Projects of Fujian Province (No. 202010007)

Abstract: Accurately identifying roadside objects like trees, cars, and traffic poles from mobile LiDAR point clouds is of great significance for some applications such as intelligent traffic systems, navigation and location services, autonomous driving, and high precision map. In the paper, we proposed a point-group-view network (PGVNet) to classify the roadside objects into trees, cars, traffic poles, and others, which utilize and fuse the advanced global features of multi-view images and the spatial geometry information of point cloud. To reduce redundant information between similar views and highlight salient view features, the PGVNet model employs a hierarchical view-group-shape architecture to split all views into different groups according to their discriminative level, which uses the pre-trained VGG network as the bone network. In view-group-shape architecture, global-level significant features are further generated from group descriptors with their weights. Moreover, an attention-guided fusion network is used to fuse the global features from multi-view images and local geometric features from point clouds. In particular, the global advanced features from multi-view images are quantified and leveraged as the attention mask to further refine the intrinsic correlation and discriminability of the local geometric features from point clouds, which contributions to recognize the roadside objects. We have evaluated the proposed method on five different mobile LiDAR point cloud data. Five test datasets of different urban scenes by different mobile laser scanning systems are used to evaluate the validities of the proposed method. Four accuracy evaluation metrics precision, recall, quality and Fscore of trees, cars and traffic poles on the selected testing datasets achieve (99.19%,94.27%,93.58%,96.63%),(94.20%,97.56%,92.02%,95.68%),(91.48%,98.61%,90.39%,94.87%), respectively. Experimental results and comparisons with state-of-the-art methods demonstrate that the PGVNet model is available to effectively identify roadside objects from the mobile LiDAR point cloud, which can provide data support for elements construction and vectorization in high precision map applications.

Key words: mobile laser scanning systems, point cloud classification, multi-view, deep learning, attention mechanism

CLC Number: