In this paper we address the problem of automatic laser scan registration in urban environments. This represents a challenging problem for two major reasons. First, two individual laser scans might be captured at significantly changed viewpoints (wide baseline) and have very little overlap. Second, man-made buildings usually contain many structures of similar appearances. This will result in considerable aliasing in the matching process. By sensor fusion of laser data with camera images, we propose a novel improvement to the existing 2D feature techniques to enable automatic 3D alignment between two widely separated scans. The key idea consists of extracting dominant planar structures from 3D point clouds and then utilizing the recovered 3D geometry to improve the performance of 2D image feature for wide baseline matching. The resulting feature descriptors become more robust to camera viewpoint changes after the procedure of viewpoint normalization. Moreover, the viewpoint normalized 2D features provide robust local feature information including patch scale and dominant orientation for effective repetitive structure matching in man-made environments. Comprehensive experimental evaluations with real data demonstrate the potential of the proposed method for automatic wide baseline 3D scan alignment in urban environments.
|Number of pages||10|
|Journal||ISPRS journal of photogrammetry and remote sensing|
|Publication status||Published - 1 Dec 2011|
Yang, M. Y., Cao, Y., & Mcdonald, J. (2011). Fusion of camera images and laser scans for wide baseline 3D scene alignment in urban environments. ISPRS journal of photogrammetry and remote sensing, 66(6), S52-S61. https://doi.org/10.1016/j.isprsjprs.2011.09.004