Instead of using 3D scanners, which yield consistent 3D scans in
the first place, some groups have attempted to build 3D
volumetric representations of environments with 2D laser range
finders. Thrun et al. [15,28], Früh et
al. [12] and Zhao et al. [30] use two
2D laser range finders for acquiring 3D data. One laser scanner is
mounted horizontally, the other vertically. The latter one grabs
a vertical scan line, which is transformed into 3D points based on
the current robot pose. Since the vertical scanner is not able to
scan sides of objects, Zhao et al. use two additional, vertically
mounted 2D scanners, shifted by to reduce occlusions
[30]. The horizontal scanner is used to compute the
robot pose. The precision of 3D data points depends on that pose
and on the precision of the scanner.
A few other groups use highly accurate, expensive 3D laser scanners [23,1,13]. The RESOLV project aimed at modeling interiors for virtual reality and tele-presence [23]. They used a RIEGL laser range finder on robots and the ICP algorithm for scan matching [7,9]. The AVENUE project develops a robot for modeling urban environments [1], using a CYRAX laser scanner and a feature-based scan matching approach for registering of the 3D scans in a common coordinate system [25]. Nevertheless, in their recent work they do not use data of the laser scanner in the robot control architecture for localization [13]. The research group of M. Hebert has reconstructed environments using the Zoller+Fröhlich laser scanner and aims at building 3D models without initial position estimates, i.e., without odometry information [16].
Recently, different groups employ rotating SICK scanners for acquiring 3D data. Wulf et al. let the scanner rotate around the vertical axis. They acquire 3D data while moving, thus the quality of the resulting map cruicially depends on the pose estimate that is given by inertial sensors, i.e., gyros [29]. In addition, their SLAM algorithms do not consider all six degrees of freedom. Nevado et al. present novel algorithms for post processing 3D scans/scenes and extracting planar models [18].
Other approaches use information of CCD-cameras that provide a view of the robot's environment [22,8]. Yet, cameras are difficult to use in natural environments with changing light conditions. Camera-based approaches to 3D robot vision, e.g., stereo cameras and structure from motion, have difficulties providing reliable navigation and mapping information for a mobile robot in real-time. Thus some groups try to solve 3D modeling by using a planar SLAM methods and cameras, e.g., in [8].