Error propagation and uncertainty analysis between 3D laser scanner and camera

Angel Ivá García-Moreno, Denis Eduardo Hernandez-García, José Joel Gonzalez-Barbosa, Alfonso Ramírez-Pedraza, Juan B. Hurtado-Ramos, Francisco Javier Ornelas-Rodriguez

Research output: Contribution to journalArticlepeer-review

11 Scopus citations

Abstract

In this work we present an in-situ method to compute the calibration of two sensors, a LIDAR (Light Detection and Ranging) and a spherical camera. Both sensors are used in urban environment reconstruction tasks. In this scenario the speed at which the various sensors acquire and merge the information is very important; however reconstruction accuracy, which depends on sensors calibration, is also of high relevance. Here, a new calibration pattern, visible to both sensors is proposed. By this means, the correspondence between each laser point and its position in the camera image is obtained so that the texture and color of each LIDAR point can be known. Experimental results for the calibration and uncertainty analysis are presented for data collected by the platform integrated with a LIDAR and a spherical camera.

Original languageEnglish
Pages (from-to)782-793
Number of pages12
JournalRobotics and Autonomous Systems
Volume62
Issue number6
DOIs
StatePublished - Jun 2014

Keywords

  • Camera calibration
  • Effect of noise
  • LIDAR calibration
  • Sensors calibration
  • Uncertainty analysis

Fingerprint

Dive into the research topics of 'Error propagation and uncertainty analysis between 3D laser scanner and camera'. Together they form a unique fingerprint.

Cite this