You are on page 1of 2

2014 ICCE-Taiwan

,QWHJUDWLRQRI/L'$5DQG&DPHUD'DWDIRU'UHFRQVWUXFWLRQ
1

Tsung-I Chen1, Yu-Xiang Zhang1, Chia-Yen Chen1, Member, IEEE and Chia-Hung Yeh2, Member, IEEE
1
Dept. of Computer Science and Information Engineering, National University of Kaohsiung, Kaohsiung, Taiwan
2
Dept. of Electrical Engineering, National Sun Yat-sen University, Kaohsiung, Taiwan

(SfM) problem, which recovers the shape of objects based on


Abstract-- In this work, we propose an integration of the images of the object taken at different viewpoints. Similar
LiDAR device and a pair of binocular cameras for egomotion calculations are also seen in visual odometry (VO), which
estimation and 3D reconstruction. The proposed system estimates the movement of a body based on consecutive
simultaneously acquires both range data and image data as input
images [5]. In simultaneous localization and mapping (SLAM)
and calibrates the LiDAR and the cameras to determine the
correspondences between them. Egomotion is determined as the or visual SLAM (V-SLAM), the task is to have a robot or a
system moves within the environment and finally the egomotion vehicle build a map of an environment using sensory
is used to register and integrate the data from the LiDAR device feedbacks, while keeping track of its own location within the
and the cameras to construct a 3D model of the environment. 3D environment [6].
reconstructions for both indoor and outdoor environments are
performed to demonstrate the effectiveness of the proposed
III. PROPOSED METHOD
method.
The proposed method works as follows. In the first stage,
I. INTRODUCTION the laser scan of a calibration target is obtained, and the IR
In recent years, 3D related technologies have received images of the target are recorded simultaneously. The pixels
increased attention. 3D applications which were not possible belonging to the laser trails of the scanned object are identified
previously have become more and more popular. For example, from the captures images. Each of these trails is then mapped
the self-driving car developed by Google [1] uses a light to a laser rangefinder according to its position in the image.
detection and ranging (LiDAR) device to provide information Once mapped, we associate each laser point to a pixel in the
of the vehicle's 3D surroundings in real-time. The particular image which denotes the expected 2-D location where the 3-D
device gained recognition for its usage in mobile sensing since point is supposed to be projected onto. Finally, the calibration
several of the winning teams in the 2007 US DARPA Urban data is constructed. Fig. 1 shows the 3D reconstruction device
Challenge incorporated it on their vehicles [2]. used in our work.
In this work, our ultimate goal is to perform large scale 3D
reconstruction of the environment, using a 3D reconstruction
system that we have constructed. The system consists of a
device and two cameras, all of which mounted rigidly upon a
mobile platform. To achieve this goal, we first perform
calibrations for the LiDAR device and the cameras to
determine their correspondences; then we need to perform
spatial feature detection and matching on the images to obtain
a sparse point cloud. As the system moves with time, temporal
feature tracking is also performed on the images sequence to
calculate intra-frame motion. We use an auto-calibration
method to calculate the extrinsic parameters of the LiDAR and
the cameras, which are used to align the point clouds scanned Fig. 1. The LiDAR and stereo camera system for 3D reconstruction.
by the LiDAR along the path. Experimental results have A. Device calibrations
shown that the proposed system and method is able to achieve
integrated dense reconstruction for large scale scene. Conventional sensor calibration approaches utilize a set of
correspondences to adjust parameters, so that the deviation
between the expected results and the observations is
II. RELATED WORK
minimized. We use automatic calibration method proposed in
Early research on egomotion started in the 1980's when the [7], the method is able to correctly detect laser trails from the
Stanford AI Lab wanted a cart equipped with cameras to be infrared images and match them to the depth data acquired by
able to deduce its own motion [3]. Egomotion is often defined the multi-layer laser. Fig. 2 shows the results after calibration,
as the calculation of sensor motion in 3D space from 2D where the coloured points represent range data acquired by the
images [4], and has since become increasingly important in LiDAR device. The points are transformed based on the
robotics, autonomous navigation, 3D reconstruction, and so on. calibrated correspondences between the LiDAR device and the
Egomotion estimation is similar to the structure from motion cameras and overlaid on one of the binocular stereo images.

978-1-4799-4851-2/14/$31.00 2014 IEEE 93


2014 ICCE-Taiwan

(a) (b)
Fig. 5. 3D reconstruction for (a) indoor and (b) outdoor.
Fig. 2. Projection of calibrated depth data onto images

V. CONCLUSIONS
A. Egomotion estimation
In this work, we integrated a multi-beam LiDAR device and
We perform egomotion estimation using images obtained
a pair of binocular cameras to achieve egomotion estimation
from stereo cameras. The method proposed in [8] is used to
and large scale 3D reconstruction. The system simultaneously
determine the path of the 3D reconstruction system as it
acquires both range data and image data, calibrates the LiDAR
moves along a path. Fig. 3 shows the path obtained using the
device and the cameras to determine the correspondences, and
selected egomotion estimation method. To demonstrate the
calculates egomotion as the system moves through the
performance of egomotion estimation, we moved the system
environment. The calibrated device parameters and egomotion
forward and back along the same straight path. As Fig. 3
are used to register and integrate the range data from the
shows, the estimated paths overlap almost perfectly.
LiDAR device and the image data from the cameras to
construct a 3D model of the environment. Both indoor and
outdoor experiments have been performed to demonstrate the
practicality of the proposed 3D reconstruction system. In
future work, we will continue to improve the system with
respect to its reconstruction accuracy and efficiency.

ACKNOWLEDGMENT
The authors would like to thank the National Science
Council of Taiwan for sponsoring this work under grant
Fig. 3. Paths showing egomotion estimated using the method in [8], the units numbers NSC 102-2221-E-390-021, NSC 103-2218-E-110-
are in mm. 006, NSC102-2221-E-110-032-MY3 and NSC101-2221-E-
B. 3D reconstruction 110-093-MY2.
In the reconstruction step, the range and image data at each
REFERENCES
point of acquisition are registered using the correspondences
[1] G. Erico, "How Google's Self-Driving Car Works," [online] Available
calculated from the calibration process, and the egomotion from: http://spectrum.ieee.org/automaton/robotics/artificial-intelligence
estimated along the path of reconstruction is used to align the /how-google-self-driving-car-works [Accessed 18 July 2013].
data acquired at different points on the path, such that over the [2] P. Stone, P. Beeson, T. Merili and R. Madigan, "DARPA Urban
entire course, a 3D model of the environment can be Challenge Technical Report," technical report, Austin Robot
Technology, 2007.
reconstructed. We selected two locations for reconstruction, an [3] H. Moravec, "Obstacle Avoidance and Navigation in the Real World by
indoor location as shown in Fig. 4(a) and an outdoor location a Seeing Robot Rover," Ph.D. thesis, Univ. of Stanford, CA, 1980.
spanning over 10m in range as shown in Fig. 4(b). [4] M. Irani, B. Rousso and S. Peleg, "Recovery of Ego-Motion using Image
Stabilization," in Proceedings of the IEEE Computer Society Conference
on Computer Vision and Pattern Recognition 1994, pp. 21-23, 1994.
[5] D. Nister, O. Naroditsky and J. Bergen, "Visual Odometry," in
Proceedings of the IEEE Computer Society Conference on Computer
Vision and Pattern Recognition 2004, Vol.1, pp. 652-659, 2004.
[6] B. Williams and I. Reid, "On Combining Visual SLAM and Visual
Odometry," in Proceedings of the IEEE International Conference on
Robotics and Automation 2010, pp. 3494-3500, 2010.
(a) (b) [7] C.-Y. Chen, H.-J. Chien, "Geometric Calibration of a Multi-layer
Fig. 4. 3D reconstruction locations (a) indoor and (b) outdoor. LiDAR System and Image Sensors using Plane-based Implicit Laser
Parameters for Textured 3-D Depth Reconstruction", J. Visual
IV. EXPERIMENT RESULT Communication and Image Representation,
DOI:10.1016/j.jvcir.2013.08.005, Aug. 2013.
The indoor and outdoor 3D reconstructions are shown in [8] C.-Y. Chen, J.-H. Zhang, T.-I Chen and C.-F. Chen, "3D Egomotion
Fig. 5(a) and (b), respectively. As can be seen from the results, from Stereo Cameras Using Constrained Search Window and Bundle
Adjustment," In Proc. IVCNZ '13, Wellington, New Zealand, 2013.
the 3D reconstructions at different points along the paths are
integrated quite well to produce very dense 3D models of the
tested environments.

94

You might also like