Professional Documents
Culture Documents
Abstract: This paper addresses the issue of window estimation of a micro Air Vehicle (MAV)
in unknown urban environments. The MAV is required to navigate from an initial and outdoor
position to a final position inside a building. This paper develops two vision-based methods using
the information provided by the onboard vision system. To effectively identify the target and
estimate the distance between the camera carrier and target, firstly a stereo camera system is
applied. Besides, we propose another approach using point cloud captured by a RGB-D camera.
Keywords: Computer vision; Window estimation; Stereo vision; Kinect; Point cloud
1. INTRODUCTION
Recently micro air vehicles (MAV) are encountered in an
increasing number of applications due to their capabilities
to fly in indoor/outdoor environments. The abilities of
flying at low speeds, hovering, flying laterally, exploring
terrains and acquiring visual information in narrow spaces
make the MAV a highly suitable platform for several tasks
such as surveillance, reconnaissance, traffic monitoring and
inspection in complex and dangerous environments, where
manned or regular-sized aerial vehicles are not able to
accomplish these missions, even with their full operational
capabilities.
Speaking of accomplishing an efficient exploratory navigation and leading through objects avoiding collisions with
obstacles in cluttered environments, relative positioning is
necessary for precise flight in relation to objects of interest.
Vision-based positioning is well-suited for this type of
positioning tasks as it is able to free the system from
relying on external positioning devices such as satellitebased GPS. Vision also allows autonomous helicopters to
serve as intelligent eyes in the sky and provides a natural
sensing modality for object detection and tracking.
As one of the applications of vision-based state estimation,
the topic of vision guided object detection and navigation of MAV has been widely investigated. Relative pose
estimation of an unmanned helicopter has been investigated in Xu et al. (2006), where the UAV must land on
a moving target. Object positioning by using a visual
odometer for an autonomous helicopter is presented in
Amidi et al. (1998). The odometer estimates helicopter
position by visually locking on to and tracking ground
objects. The work on window tracking for one-dimensional
visual control of an unmanned autonomous helicopter is
reported in Mejias et al. (2005). Then the vision-based
This work is partially supported by the Institute for Science &
Technology of Mexico City (ICyTDF).
978-3-902823-57-1/2013 IFAC
107
feature tracking system has been combined with GPSpositioning references to navigate the helicopter towards
these features and track them, see Mejias et al. (2006).
In this paper, we require that the MAV accomplishes the
task of identifying a window and fly through it, in order to
access into a building. The fulfillment of this objective will
be quite significant for various military and civil missions.
This work presents two vision approaches developed for
the real-time identification of a window model. The first
vision-based strategy is performed by using the stereo
vision techniques. The advantage of applying stereo vision
is that it can reconstruct the depth information of the
scene more precisely than single camera, which guarantees
the safety navigation. The second vision algorithm is based
on structured light RGB-D camera. Although based on
stereo techniques and share many properties with stereo
cameras, RGB-D cameras achieve a better performance in
the spatial density of depth data. Since RGB-D cameras
illuminate a scene with a structured light pattern, they
can estimate depth in areas with poor visual texture.
The reminder of this paper is organized as follows: Section
2 addresses the window estimation problem. Section 3
presents the window estimation strategy based on stereo
vision techniques. The vision-based window detection algorithm using the RGB-D camera is introduced in Section 4.
Section 5 gives the comparison of two approaches. Finally,
Section 6 draws a conclusion and gives a perspective on
future work of the related research.
2. PROBLEM DEFINITION
Suppose that the MAV can arrive at the intermediate point
outside the target window which is provided by the GPS
system. This scenario is shown in Fig.1. Once the MAV
has achieved the aforementioned point, the parameters of
the target window model must be estimated so that the
flying control system can minimize the distance between
10.3182/20131120-3-FR-4045.00026
Before applying the vision strategy in the real-time navigation system, some simulations are conducted. To simplify
the proposed problem, the design of the window model
must be simple and distinctive for the location, feature
extraction and matching. We consider a designed pattern
as a square with black edge on a white wall, in which the
target window frame is designed as the black border.
(a)
Based on the light coding, Kinect projects a known infrared pattern onto the scene and determines depth based
on the patterns deformation captured by the infrared
CMOS imager, see Xiang et al. (2011). Functioning in this
way, Kinect can provide a 320X240 depth image at 30fps
and a 640X480 RGB color image at 30fps. When stripped
down to its essential components, the Kinect weighs 115g
- light enough to be carried by a MAV.
(a)
(b)