Professional Documents
Culture Documents
CHAPTER 1
INTRODUCTION
Video games have been entertaining us for nearly 30 years, ever since Pong was
introduced to arcades in the early 1970s. Computer graphics have become much more
sophisticated since then, and game graphics are pushing the barriers of photorealism.
Now, researchers and engineers are pulling graphics out of your television screen or
computer display and integrating them into real-world environments. This new
technology, called augmented reality, blurs the line between what's real and what's
computer-generated by enhancing what we see, hear, feel and smell.
On the spectrum between virtual reality, which creates immersive, computergenerated environments, and the real world, augmented reality is closer to the real world.
Augmented reality adds graphics, sounds, haptic feedback and smell to the natural world
as it exists. Both video games and cell phones are driving the development of augmented
reality.
Augmented reality is changing the way we view the world - or at least the way its
users see the world. Picture yourself walking or driving down the street. With augmentedreality displays, which will eventually look much like a normal pair of glasses,
informative graphics will appear in your field of view, and audio will coincide with
whatever you see.
Figure 1.1 The Sixth Sense augmented reality system lets you project a phone pad onto your hand
and phone a friend -- without removing the phone from your pocket.
Page 1
Augmented Reality
These enhancements will be refreshed continually to reflect the movements of your head.
Similar devices and applications already exist, particularly on smartphones like the
iPhone. Picture shows The SixthSense augmented reality system lets you project a phone
pad onto your hand and phone a friend -- without removing the phone from your pocket.
As computers increase in power and decrease in size, new mobile, wearable, and
pervasive computing applications are rapidly becoming feasible, providing people access
to online resources always and everywhere. This new flexibility makes possible new kind
of applications that exploit the person's surrounding context. Augmented reality (AR)
presents a particularly powerful user interface (UI) to context-aware computing
environments. AR systems integrate virtual information into a person's physical
environment so that he or she will perceive that information as existing in their
surroundings. Mobile augmented reality systems (MARS) provide this service without
constraining the individuals whereabouts to a specially equipped area. Ideally, they work
virtually anywhere, adding a palpable layer of information to any environment whenever
desired. By doing so, they hold the potential to revolutionize the way in which
information is presented to people.
Page 2
Augmented Reality
CHAPTER 2
Page 3
Augmented Reality
While it may be some time before you buy a device like SixthSense, more
primitive versions of augmented reality are already here on some cell phones, particularly
in applications for the iPhone and phones with the Android operating system. In the
Netherlands, cell phone owners can download an application called Layar that uses the
phone's camera and GPS capabilities to gather information about the surrounding area.
Layar then shows information about restaurants or other sites in the area, overlaying this
information on the phone's screen. You can even point the phone at a building, and Layar
will tell you if any companies in that building are hiring, or it might be able to find photos
of the building on Flickr or to locate its history on Wikipedia.
Layar isn't the only application of its type. In August 2009, some iPhone users were
surprised to find an augmented-reality "easter egg" hidden within the Yelp application.
Yelp is known for its user reviews of restaurants and other businesses, but its hidden
augmented-reality component, called Monocle, takes things one step further. Just start up
the Yelp app, shake your iPhone 3GS three times and Monocle activates. Using your
phone's GPS and compass, Monocle will display information about local restaurants,
including ratings and reviews, on your cell phone screen. You can touch one of the
listings to find out more about a particular restaurant.
Page 4
Augmented Reality
There are other augmented reality apps out there for the iPhone and other similar
phones -- and many more in development. Urbanspoon has much of the same
functionality as Yelp's Monocle. Then there's Wikitude, which finds information from
Wikipedia about sites in the area. Underlying most of these applications are a phone's
GPS and compass; by knowing where you are, these applications can make sure to offer
information relevant to you. We're still not quite at the stage of full-on image recognition,
but trust us, people are working on it. We've looked at some of the existing forms of
augmented reality. On the next page, we'll examine some of the other applications of the
technology, such as in video games and military hardware.
Page 5
Augmented Reality
Camera
Small projector
Smartphone
Mirror
These components are strung together in a lanyard like apparatus that the user
wears around his neck. The user also wears four colored caps on the fingers, and these
caps are used to manipulate the images that the projector emits. SixthSense is remarkable
because it uses these simple, off-the-shelf components that cost around $350. It is also
notable because the projector essentially turns any surface into an interactive screen.
Essentially, the device works by using the camera and mirror to examine the surrounding
world, feeding that image to the phone (which processes the image, gathers GPS
coordinates and pulls data from the Internet), and then projecting information from the
projector onto the surface in front of the user, whether it's a wrist, a wall, or even a
person. Because the user is wearing the camera on his chest, SixthSense will augment
whatever he looks at; for example, if he picks up a can of soup in a grocery store,
SixthSense can find and project onto the soup information about its ingredients, price,
nutritional value -- even customer reviews.
Department of ECE, BGSIT
Page 6
Augmented Reality
CHAPTER 3
On the spectrum between virtual reality, which creates immersive, computergenerated environments, and the real world, augmented reality is closer to the real world.
Augmented reality adds graphics, sounds, haptic feedback and smell to the natural world
as it exists. Both video games and cell phones are driving the development of augmented
reality. Everyone from tourists, to soldiers, to someone looking for the closest subway
stop can now benefit from the ability to place computer-generated graphics in their field
of vision.
Augmented reality is changing the way we view the world -- or at least the way its
users see the world. Picture yourself walking or driving down the street. With augmentedreality displays, which will eventually look much like a normal pair of glasses,
informative graphics will appear in your field of view, and audio will coincide with
whatever you see. These enhancements will be refreshed continually to reflect the
movements of your head.
3.1 Hardware
The main hardware components for augmented reality are: processor, display,
sensors and input devices. These elements, specifically CPU, display, camera and MEMS
sensors such as accelerometer, GPS, solid state compass are often present in modern
smartphones, which make them prospective AR platforms.
3.1.1 Computer
The computer analyzes the sensed visual and other data to synthesize and position
augmentations. Camera based systems require powerful CPU and considerable amount of
RAM for processing camera images. Wearable computing systems employ a laptop in a
Department of ECE, BGSIT
Page 7
Augmented Reality
backpack configuration. For stationary systems a traditional workstation with a powerful
graphics card. Sound processing hardware could be included in augmented reality
systems.
3.1.3 Display
There are three major display techniques for Augmented Reality: head-mounted
displays, handheld displays and spatial displays. Some examples of spatial augmented
reality displays include shader lamps, mobile projectors, virtual tables, and smart
projectors, described by O. Bimber and R. Raskar in 2005. Shader lamps, developed by
Raskar et al. in 1999, mimic and augment reality by projecting imagery onto neutral
objects, providing the opportunity to enhance the objects appearance. This can be
accomplished with materials of a simple unit- a projector, camera, and sensor. Handheld
projectors further this goal by enabling cluster configurations of environment sensing,
reducing the need for additional environmental sensing.
Other tangible applications include table and wall projections. One such innovation,
the Extended Virtual Table, separates the virtual from the real by including beam-splitter
mirrors attached to the ceiling at an adjustable angle. Virtual showcases, which employ
beam-splitter mirrors together with multiple graphics displays, provide an interactive
means of simultaneously engaging with the virtual and the real. Altogether, augmented
reality display technology can be applied to improve design and visualization, or function
as scientific simulations, and tools for education or entertainment.
Page 8
Augmented Reality
itself could be used as 3D pointing device, with 3D position of the phone restored from
the camera images.
Page 9
Augmented Reality
CHAPTER 4
The biggest challenge facing developers of augmented reality is the need to know where
the user is located in reference to his or her surroundings. There's also the additional
problem of tracking the movement of users' eyes and heads. A tracking system has to
recognize these movements and project the graphics related to the real world environment
the user is seeing at any given moment. Currently, both video see through and optical seethrough displays typically have lag in the overlaid material due to the tracking
technologies currently available.
Page 10
Augmented Reality
External view of the columbian printer maintainence application. Note that all the
components must be tracked. External view of the columbian printer maintainence
application. Note that all the components must be tracked. The system uses the known
location of the LEDs, the known geometry of the usermounted optical sensors and a
special algorithm to compute and report the user's position and orientation. The system
resolves linear motion of less than .2 millimeters, and angular motions less than .03
degrees. It has an update rate of more than 1500 Hz, and latency is kept at about one
millisecond.
In everyday life, people rely on several senses--including what they see, cues from
their inner ears and gravity's pull on their bodies--to maintain their bearings. In a
similarfashion, "hybrid trackers" draw on several sources of sensory information. For
example, thewearer of an AR display can be equipped with inertial sensors (gyroscopes
and accelerometers) to record changes in head orientation. Combining this information
with data from the optical, video or ultrasonic devices greatly improves the accuracy of
the tracking.
Page 11
Augmented Reality
time kinematic GPS, a more sophisticated form of differential GPS that also compares the
phases of the signals at the fixed and mobile receivers.
Unfortunately, GPS is not the ultimate answer to position tracking. The satellite
signals are relatively weak and easily blocked by buildings or even foliage. This rules out
useful tracking indoors or in places like midtown Manhattan, where rows of tall buildings
block most of the sky. GPS tracking works well in wide open spaces and relatively low
buildings. GPS provides far too few updates per second and is too inaccurate to support
the precise overlaying of graphics on nearby objects.
Augmented-reality systems place extraordinarily high demands on the accuracy,
resolution, repeatability and speed of tracking technologies. Hardware and software
delays introduce a lag between the user's movement and the update of the display. As a
result, virtual objects will not remain in their proper positions as the user moves about or
turns his or her head. One technique for combating such errors is to equip AR systems
with software that makes short-term predictions about the user's future motions by
extrapolating from previous movements. And in the long run, hybrid trackers that include
computer vision technologies may be able to trigger appropriate graphics overlays when
the devices recognize certain objects in the user's view.
Page 12
Augmented Reality
CHAPTER 5
CLASSIFICATION OF MOBILE AR
Control: Unlike a stand-alone desktop UI, where the only way the user can
interact with the presented environment is through a set of well defined
techniques,the MARS UI needs to take into account the unpredictability of the
real world. Forexample, a UI technique might rely on a certain object being in the
user's field of view and not occluded by other information. Neither of the
properties can be guaranteed: the user is free to look away, and other information
could easily get in the way, triggered by the user's own movement or an
unforeseen event (such as another user entering the field of view). Thus, to be
effective, the UI technique either has to relax the non-occlusion requirement, or
has to somehow guarantee on-occlusion in spite of possible contingencies.
Consistency: People have internalized many of the laws of the physical world.
When using a computer, a person can learn the logic of a new UI. As long as these
two worlds are decoupled (as they are in the desktop setting), inconsistencies
between them are often understandable. In the case of MARS, however, we need
to be very careful to design UIs in which the physical and virtual world are
consistent with each other. Need for embedded semantic information: In MARS,
virtual material is overlaid on top of the real world. Thus we need to establish
concrete semantic relationships between virtual and physical objects to
characterize UI behavior. In fact, since many virtual objects are designed to
Page 13
Augmented Reality
annotate the real world, these virtual objects need to store information about the
physical objects to which they refer (or at least have to know how to access that
information). Display space: In terms of the available display space and its best
use, MARS UIs have to deal with a much more complicated task compared to
traditional 2D UIs. Instead of one area of focus (e.g., one desktop display), we
have to deal with a potentially unlimited display space surrounding the user, only
a portion of which is visible at any time. The representation of that portion of
augmented space depends on the user's position, head orientation, personal
preferences (e.g., filter settings) and ongoing interactions with the augmented
world, among other things. Management of virtual information in this space is
made even more difficult by constraints that other pieces of information may
impose. Certain virtual or physical objects may, for example, need to be visible
under all circumstances, and thus place restrictions on the display space that other
elements are allowed to obstruct. The display management problem is further
complicated by the possibility of taking into account multiple displays. MARS, as
a nonexclusive UI to the augmented world, may seamlessly make use of other
kinds of displays, ranging from wall-sized, to desk-top, to hand-held. If such
display devices are available and Telegeoinformatics: Location-Based Computing
and Services 33 accessible to the MARS, questions arise as to which display to
use for what kind of information and how to let the user know about that decision.
Scene dynamics: In a head-tracked UI, the scene will be much more dynamic than
in a stationary UI. In MARS, this is especially true, since in addition to all the
dynamics due to head motion, the system has to consider moving objects in the
real world that might interact visually or audibly with the UI presented on the
headworn display. Also, we have to contend with a potentially large variability in
tracking accuracy over time. Because of these unpredictable dynamics, the spatial
composition of the UI needs to be flexible and the arrangement of UI elements
may need to be changed. On the other hand, traditional UI design wisdom
suggests minimizing dynamic changes in the UI composition (Shneiderman,
1998). One possible solution to this dilemma lies in the careful application of
automated UI management techniques.
Page 14
Augmented Reality
Page 15
Augmented Reality
Page 16
Augmented Reality
Page 17
Augmented Reality
Page 18
Augmented Reality
from the video of the real world. This has the effect of superimposing the virtual objects
over real world. A more sophisticated composition would use depth information. If the
system had depth information at each pixel for the real world images, it could combine
the real and virtual images by a pixel-by-pixel depth comparison. This would allow real
objects to cover virtual objects and vice-versa.
Page 19
Augmented Reality
Page 20
Augmented Reality
CHAPTER 6
Mobile augmented realitycontact lenses may do more that improve your sight.
Someday they could replace your mobile phone and let you communicate visually
anywhere in the world, improve your health and make virtual reality real. Perhaps your
ophthalmologist could perform Lasik surgery, burning a wireless circuit into your cornea?
6.1.1 Mobile Augmented Reality Hits Contact Lens Technology
Babak Parviz at the University of Washington in Seattle, is working on a contact
lens technology that could revolutionize wireless health monitoring and mobile
applications for your iPhone. But dont stop there
Babak Parviz lenses become biosensors that monitor internal body functions. While the
prototype version of the lens is powered by radio waves beaming electricity to a loop
antenna embedded in the contact lens, Parviz thinks a mobile phone or solar cells
(wireless electricity) could generate power for the lenses. Mobile augmented reality could
be just around the corner.
Page 21
Augmented Reality
The starting point for this article came from an unconventional place a furniture
app from Harveys that allows you to place the sofa of your dreams (possibly the most
unnecessary use of hyperbole ever there) into a room in your house without actually
having to buy it and lug it into your house. Very simply, the app allows users to see what
particular furniture will look like in their rooms, using an innovative augmented reality
camera, in a hypothetical try-before-you-buy approach to pre-shopping. A great app, by
the way, and one that scratches an itch you never thought you knew existed, which is
always key to anything like this being successful, but hardly one of the great
advancements of the 21st century.
Page 22
Augmented Reality
Imagine an exit or turn arrow that doesnt just point to the right but has the same
angle as the turn youre approaching and expands as you reach the turn. Youll interact
with the displays by arm gestures, voice input, or traditional dashboard buttons and
knobs. Cost will be an issue since current HUDs run more than $1,000.
The Mercedes demo on the show floor took the form of a virtual drive in a
simulator through San Francisco with points of interest circled, the idea being you could
gesture to get more information. The gesture youd use in a busy urban area might be a
raised middle finger because of the potential information overload: restaurant, bar,
jewelry shops, tour bus stops, bridge and tunnel congestion. And thats even before you
wonder whos providing the POI information and is it there because its the best, or
because it pays the automaker the best. This is an issue for the whole industry, not just
Mercedes, whose motto is The Best or Nothing.
Department of ECE, BGSIT
Page 23
Augmented Reality
The term augmented reality, or AR, is used to describe technology that enables
normal sight to be modified or enhanced by a computer, allowing data on an object in the
field of view to be displayed alongside it, or to add objects or menus that arent really
there to augment your normal senses. If this all sounds a long way off to you, you need
look no further than your smartphones app store to find a host of smartphone apps
currently in use that use augmented reality to some extent, using the phones camera.
The Nintendo 3DS uses cards for its augmented reality content. If you place an
AR card on a table the 3DSs camera recognises the card and its orientation and inserts
Nintendo characters onto your screen to do battle on the table top. This works just like
QR codes for your smartphone, as a kind of visual barcode for your devices camera- each
card represents a different character or action. This kind of content will become a
standard feature of handheld gaming in 2012.
Page 24
Augmented Reality
objects. The technology itself is not exactly new and is being used in a varied of
applications from GPS systems to fitness apps.
Future
The future of augmented reality seems to be more inclined towards haptics and
tactile feedback (which uses sense of touch).
Earlier, Senseg, which claims mastery over such technology, had demonstrated at
the Consumer Electronics Show, haptic technology that allowed users to feel their apps on
the touchscreen by manipulating an electric field. With the technology, users will now be
able to feel bumps and ridges, and also figure out which areas are more rough than others.
The company believes if users are able to get the feel of anything other than glass, it
would be a better experience. The company demonstrated an Android tablet with a
touchscreen, which had different textures on it. Users get used to such a touchscreen
quickly and may not want to go back to a regular touchscreen. Senseg has deliberately
made the effect subtle so it doesn't distract the users while making its presence clearly
felt. The company is still working on different kind of sensations. Another app from
application developer CrowdOptic may point towards a new trend in augmented reality
apps. The new technology of CrowdOptic focuses on crowds, such as in concert or sports
events. When the camera of the smartphone is pointed at a player during a sporting event,
Department of ECE, BGSIT
Page 25
Augmented Reality
it displays real time information about the player and the game. The details and context
can also be shared through different social networks. So far, getting information on
moving objects through augmented reality apps was not possible.
Page 26
Augmented Reality
Page 27
Augmented Reality
Despite these concerns, imagine the possibilities: you may learn things about the
city you've lived in for years just by pointing your AR-enabled phone at a nearby park or
building. If you work in construction, you can save on materials by using virtual markers
to designate where a beam should go or which structural support to inspect.
Paleontologists working in shifts to assemble a dinosaur skeleton could leave virtual
"notes" to team members on the bones themselves, artists could produce virtual graffiti
and doctors could overlay a digital image of a patient's X-rays onto a mannequin for
added realism.
The future of augmented reality is clearly bright, even as it already has found its
way into our cell phones and video game systems.
Page 28
Augmented Reality
REFERENCES
[01] M. Uenohara and T. Kanade. Vision-based object registration for real-time image
overlay. In Proc. of the IEEE Conference on Computer Vision, Virtual Reality and
Robotics in Medicine (CVRMed95), pages 1322, 1995.
[02] B. Walter, G. Alppay, E. Lafortune, S. Fernandez, and D. Greenberg. Fitting virtual
lights for non-diffuse walkthroughs. In Proceedings of SIGGRAPH 97, 1997.
[03] P. Wellner. Interacting with paper on the digital desk. Communications of the ACM,
36(7):8796, 1993.
[04] J. Weng, P. Cohen, and N. Rebibo. Motion and structure estimation from stereo
image sequences. IEEE Transactions on Robotics and Automation, 8:362382, June 1992.
[05] J. Weng, T. S. Huang, and N. Ahuja. Motion and structure from two perspective
views: Algorithms, error analysis, and error estimation. IEEE Transactions on Pattern
Analysis and Machine Intelligence, 11(5):451476, 1989. This is a computer vision paper
that has geometry extraction from motion and stereo.
[06] R. Whitaker, C. Crampton, D. Breen, M. Tuceryan, and E. Rose. Object calibration
for augmented reality. In Proceedings of Eurographics95, pages 1527, Maastricht, The
Netherlands, August 1995.
[07] M. Wloka and B. Anderson. Resolving occlusion in augmented reality. In
Proceedings of the ACM Symposium on Interactive 3D Graphics, pages 512, 1995.
[08] Yizhou Yu and Jitendra Malik. Recovering photometric properties of architectural
scenes from photographs. In Proceedings of the SIGGRAPH 98 conference, Annual
Conference Series, pages 207217, 1998.
Page 29