AUTOMATED MOSAICKING OF MULTIPLE 3D POINT CLOUDS GENERATED FROM A DEPTH CAMERA
In this paper, we propose a method for automated mosaicking of multiple 3D point clouds generated from a depth camera. A depth camera generates depth data by using ToF (Time of Flight) method and intensity data by using intensity of returned signal. The depth camera used in this paper was a SR4000...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Copernicus Publications
2016-06-01
|
Series: | The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences |
Online Access: | https://www.int-arch-photogramm-remote-sens-spatial-inf-sci.net/XLI-B3/269/2016/isprs-archives-XLI-B3-269-2016.pdf |
Summary: | In this paper, we propose a method for automated mosaicking of multiple 3D point clouds generated from a depth camera. A depth
camera generates depth data by using ToF (Time of Flight) method and intensity data by using intensity of returned signal. The depth
camera used in this paper was a SR4000 from MESA Imaging. This camera generates a depth map and intensity map of 176 x 44
pixels. Generated depth map saves physical depth data with mm of precision. Generated intensity map contains texture data with
many noises. We used texture maps for extracting tiepoints and depth maps for assigning z coordinates to tiepoints and point cloud
mosaicking. There are four steps in the proposed mosaicking method. In the first step, we acquired multiple 3D point clouds by
rotating depth camera and capturing data per rotation. In the second step, we estimated 3D-3D transformation relationships between
subsequent point clouds. For this, 2D tiepoints were extracted automatically from the corresponding two intensity maps. They were
converted into 3D tiepoints using depth maps. We used a 3D similarity transformation model for estimating the 3D-3D
transformation relationships. In the third step, we converted local 3D-3D transformations into a global transformation for all point
clouds with respect to a reference one. In the last step, the extent of single depth map mosaic was calculated and depth values per
mosaic pixel were determined by a ray tracing method. For experiments, 8 depth maps and intensity maps were used. After the four
steps, an output mosaicked depth map of 454x144 was generated. It is expected that the proposed method would be useful for
developing an effective 3D indoor mapping method in future. |
---|---|
ISSN: | 1682-1750 2194-9034 |