Augmented reality

From Wikipedia, the free encyclopedia
  (Redirected from Augmented Reality)
Jump to: navigation, search
Samsung SARI AR SDK marker less tracker used in the AR EdiBear game (Android OS)
AR Tower Defense game on the Nokia N95 smartphone (Symbian OS) uses fiduciary markers
NASA X38 display showing video map overlays including runways and obstacles during flight test in 2000.

Augmented reality (AR) is a live direct or indirect view of a physical, real-world environment whose elements are augmented (or supplemented) by computer-generated sensory input such as sound, video, graphics or GPS data. It is related to a more general concept called mediated reality, in which a view of reality is modified (possibly even diminished rather than augmented) by a computer. As a result, the technology functions by enhancing one’s current perception of reality.[1] By contrast, virtual reality replaces the real world with a simulated one.[2][3] Augmentation is conventionally in real-time and in semantic context with environmental elements, such as sports scores on TV during a match. With the help of advanced AR technology (e.g. adding computer vision and object recognition) the information about the surrounding real world of the user becomes interactive and digitally manipulable. Artificial information about the environment and its objects can be overlaid on the real world.[4][5][6][7][8]

Technology[edit]

Hardware[edit]

Hardware components for augmented reality are: processor, display, sensors and input devices. Modern mobile computing devices like smartphones and tablet computers contain these elements which often include a camera and MEMS sensors such as accelerometer, GPS, and solid state compass, making them suitable AR platforms.[9]

Display[edit]

Various technologies are used in Augmented Reality rendering including optical projection systems, monitors, hand held devices, and display systems worn on one's person.

Head-mounted[edit]

A head-mounted display (HMD) is a display device paired to a headset such as a harness or helmet. HMDs place images of both the physical world and virtual objects over the user's field of view. Modern HMDs often employ sensors for six degrees of freedom monitoring that allow the system to align virtual information to the physical world and adjust accordingly with the user's head movements.[10][11][12] HMDs can provide users immersive, mobile and collaborative AR experiences.[13]

Eyeglasses[edit]

AR displays can be rendered on devices resembling eyeglasses. Versions include eyewear that employ cameras to intercept the real world view and re-display its augmented view through the eye pieces[14] and devices in which the AR imagery is projected through or reflected off the surfaces of the eyewear lens pieces.[15][16][17]

HUD[edit]
See also: Head-up display

Devices that can augment only part of ones field of view like Google Glass is are intended for an AR experience. After the debut of Google Glass many other HUD devices emerged as alternatives.[18][19][20]

CrowdOptic, an existing app for smartphones, applies algorithms and triangulation techniques to photo metadata including GPS position, compass heading, and a time stamp to arrive at a relative significance value for photo objects.[21] CrowdOptic technology can be used by Google Glass users to learn where to look at a given point in time.[22]

Contact lenses[edit]

Contact lenses that display AR imaging are in development. These bionic contact lenses might contain the elements for display embedded into the lens including integrated circuitry, LEDs and an antenna for wireless communication.[23][24][25][26] Another version of contact lenses, in development for the U.S. Military, is designed to function with AR spectacles, allowing soldiers to focus on close-to-the-eye AR images on the spectacles and distant real world objects at the same time.[27][28] In 2013, at the Augmented World Expo Conference,[29] a futuristic video named Sight[30] featuring the potential of having augmented reality through contact lenses received the best futuristic augmented reality video award.

Virtual retinal display[edit]

A virtual retinal display (VRD) is a personal display device under development at the University of Washington's Human Interface Technology Laboratory. With this technology, a display is scanned directly onto the retina of a viewer's eye. The viewer sees what appears to be a conventional display floating in space in front of them.[31]

EyeTap[edit]

The EyeTap (also known as Generation-2 Glass[32]) captures rays of light that would otherwise pass through the center of a lens of an eye of the wearer, and substitutes synthetic computer-controlled light for each ray of real light. The Generation-4 Glass[32] (Laser EyeTap) is similar to the VRD (i.e. it uses a computer controlled laser light source) except that it also has infinite depth of focus and causes the eye itself to, in effect, function as both a camera and a display, by way of exact alignment with the eye, and resynthesis (in laser light) of rays of light entering the eye.[33]

Handheld[edit]

Handheld displays employ a small display that fits in a user's hand. All handheld AR solutions to date opt for video see-through. Initially handheld AR employed fiduciary markers,[34] and later GPS units and MEMS sensors such as digital compasses and six degrees of freedom accelerometergyroscope. Today SLAM markerless trackers such as PTAM are starting to come into use. Handheld display AR promises to be the first commercial success for AR technologies. The two main advantages of handheld AR is the portable nature of handheld devices and ubiquitous nature of camera phones. The disadvantages are the physical constraints of the user having to hold the handheld device out in front of them at all times as well as distorting effect of classically wide-angled mobile phone cameras when compared to the real world as viewed through the eye.[35]

Spatial[edit]

Spatial Augmented Reality (SAR) augments real world objects and scenes without the use of special displays such as monitors, head mounted displays or hand-held devices. SAR makes use of digital projectors to display graphical information onto physical objects. The key difference in SAR is that the display is separated from the users of the system. Because the displays are not associated with each user, SAR scales naturally up to groups of users, thus allowing for collocated collaboration between users.

Examples include shader lamps, mobile projectors, virtual tables, and smart projectors. Shader lamps mimic and augment reality by projecting imagery onto neutral objects, providing the opportunity to enhance the object’s appearance with materials of a simple unit- a projector, camera, and sensor.

Other applications include table and wall projections. One innovation, the Extended Virtual Table, separates the virtual from the real by including beam-splitter mirrors attached to the ceiling at an adjustable angle.[36] Virtual showcases, which employ beam-splitter mirrors together with multiple graphics displays, provide an interactive means of simultaneously engaging with the virtual and the real. Many more implementations and configurations make spatial augmented reality display an increasingly attractive interactive alternative.

A SAR system can display on any number of surfaces of an indoor setting at once. SAR supports both a graphical visualisation and passive haptic sensation for the end users. Users are able to touch physical objects in a process that provides passive haptic sensation.[7][37][38][39]

Tracking[edit]

Modern mobile augmented reality systems use one or more of the following tracking technologies: digital cameras and/or other optical sensors, accelerometers, GPS, gyroscopes, solid state compasses, RFID and wireless sensors. These technologies offer varying levels of accuracy and precision. Most important is the position and orientation of the user's head. Tracking the user's hand(s) or a handheld input device can provide a 6DOF interaction technique.[40]

Input devices[edit]

Techniques include speech recognition systems that translate a user's spoken words into computer instructions and gesture recognition systems that can interpret a user's body movements by visual detection or from sensors embedded in a peripheral device such as a wand, stylus, pointer, glove or other body wear.[41][42][43][44]

Computer[edit]

The computer analyzes the sensed visual and other data to synthesize and position augmentations.

Software and algorithms[edit]

A key measure of AR systems is how realistically they integrate augmentations with the real world. The software must derive real world coordinates, independent from the camera, from camera images. That process is called image registration which uses different methods of computer vision, mostly related to video tracking.[45][46] Many computer vision methods of augmented reality are inherited from visual odometry. Usually those methods consist of two parts.

First detect interest points, or fiduciary markers, or optical flow in the camera images. First stage can use feature detection methods like corner detection, blob detection, edge detection or thresholding and/or other image processing methods.[47][48] The second stage restores a real world coordinate system from the data obtained in the first stage. Some methods assume objects with known geometry (or fiduciary markers) present in the scene. In some of those cases the scene 3D structure should be precalculated beforehand. If part of the scene is unknown simultaneous localization and mapping (SLAM) can map relative positions. If no information about scene geometry is available, structure from motion methods like bundle adjustment are used. Mathematical methods used in the second stage include projective (epipolar) geometry, geometric algebra, rotation representation with exponential map, kalman and particle filters, nonlinear optimization, robust statistics.

Augmented Reality Markup Language (ARML) is a data standard developed within the Open Geospatial Consortium (OGC),[49] which consists of an XML grammar to describe the location and appearance of virtual objects in the scene, as well as ECMAScript bindings to allow dynamic access to properties of virtual objects.

To enable rapid development of Augmented Reality Application, some software development kits (SDK) have emerged.[50][51] A few SDK such as CloudRidAR [52] leverage cloud computing for performance improvement. Some of the well known AR SDKs are offered by Metaio,[53] Vuforia,[54] Mobinett AR.[55] Wikitude[56] and Layar.[57]

Applications[edit]

Augmented reality has many applications. First used for military, industrial, and medical applications, it has also been applied to commercial and entertainment areas.[58]

Archaeology[edit]

AR can be used to aid archaeological research, by augmenting archaeological features onto the modern landscape, enabling archaeologists to formulate conclusions about site placement and configuration.[59]

Another application given to AR in this field is the possibility for users to rebuild ruins, buildings, or even landscapes as they formerly existed.[60]

Architecture[edit]

AR can aid in visualizing building projects. Computer-generated images of a structure can be superimposed into a real life local view of a property before the physical building is constructed there. AR can also be employed within an architect's work space, rendering into their view animated 3D visualizations of their 2D drawings. Architecture sight-seeing can be enhanced with AR applications allowing users viewing a building's exterior to virtually see through its walls, viewing its interior objects and layout.[61][62]

Art[edit]

AR technology has helped disabled individuals create art by using eye tracking to translate a user's eye movements into drawings on a screen.[63] An item such as a commemorative coin can be designed so that when scanned by an AR-enabled device it displays additional objects and layers of information that were not visible in a real world view of it.[64][65] In 2013, L'Oreal used CrowdOptic technology to create an augmented reality at the seventh annual Luminato Festival in Toronto, Canada.[22]

Commerce[edit]

View Description image 1

AR can enhance product previews such as allowing a customer to view what's inside a product's packaging without opening it.[66] AR can also be used as an aid in selecting products from a catalog or through a kiosk. Scanned images of products can activate views of additional content such as customization options and additional images of the product in its use.[67][68] AR is used to integrate print and video marketing. Printed marketing material can be designed with certain "trigger" images that, when scanned by an AR enabled device using image recognition, activate a video version of the promotional material. A major difference between Augmented Reality and straight forward image recognition is that you can overlay multiple media at the same time in the view screen, such as social media share buttons, in-page video even audio and 3D objects. Traditional print only publications are using Augmented Reality to connect many different types of media.[69][70][71]

Construction[edit]

With the continual improvements to GPS accuracy, businesses are able to use augmented reality to visualize georeferenced models of construction sites, underground structures, cables and pipes using mobile devices.[72] Following the Christchurch earthquake, the University of Canterbury released, CityViewAR, which enabled city planners and engineers to visualize buildings that were destroyed in the earthquake.[73] Not only did this provide planners with tools to reference the previous cityscape, but it also served as a reminder to the magnitude of the devastation caused, as entire buildings were demolished.

Education[edit]

App iSkull, an augmented human skull for education (iOS OS)

Augmented reality applications can complement a standard curriculum. Text, graphics, video and audio can be superimposed into a student’s real time environment. Textbooks, flashcards and other educational reading material can contain embedded “markers” that, when scanned by an AR device, produce supplementary information to the student rendered in a multimedia format.[74][75][76] Students can participate interactively with computer generated simulations of historical events, exploring and learning details of each significant area of the event site.[77] On higher education, there are some applications that can be used. For instance, Construct3D, a Studierstube system, allows students to learn mechanical engineering concepts, math or geometry. This is an active learning process in which students learn to learn with technology.[78] AR can aid students in understanding chemistry by allowing them to visualize the spatial structure of a molecule and interact with a virtual model of it that appears, in a camera image, positioned at a marker held in their hand.[79] It can also enable students of physiology to visualize different systems of the human body in three dimensions.[80] Augmented reality technology also permits learning via remote collaboration, in which students and instructors not at the same physical location can share a common virtual learning environment populated by virtual objects and learning materials and interact with another within that setting.[81]

This resource could also take of advantage in Primary School. Students learn through experiences, besides when children are so young, they need see to learn. For instance, they can learn new knowledge about Astronomy, which is usually difficult to acquire to them, with this device children can understand better The Solar System because they would see it in 3D; even children under 6 years old could understand it following that method. In addition, learners could change the pictures of their Science Book for using this resource. On the other hand to teach bones or organs, they could also stick one paper on their body and that paper contains an embedded “markers” about a bones or an organ that existed under the paper, and the teacher would only need to press a button when children would change the place of the paper, in this way, we would use the same embedded “markers” in order to teach another part of the body.[citation needed]

Emergency Management / Search and Rescue[edit]

LandForm+ is a geographic augmented reality system used for search and rescue, and emergency management.

Augmented reality systems are used in public safety situations - from super storms to suspects at large. Two interesting articles from Emergency Management magazine discuss the power of the technology for emergency management. The first is "Augmented Reality--Emerging Technology for Emergency Management" by Gerald Baron.[82] Per Adam Crowe: "Technologies like augmented reality (ex: Google Glass) and the growing expectation of the public will continue to force professional emergency managers to radically shift when, where, and how technology is deployed before, during, and after disasters.".[83]

Another example, a search aircraft is looking for a lost hiker in rugged mountain terrain. Augmented reality systems provide aerial camera operators with a geographic awareness of forest road names and locations blended with the camera video. As a result, the camera operator is better able to search for the hiker knowing the geographic context of the camera image. Once found, the operator can more efficiently direct rescuers to the hiker's location.[84]

Everyday[edit]

30 years of Augmediated Reality in everyday life.

Since the 1970s and early 1980s, Steve Mann has been developing technologies meant for everyday use i.e. "horizontal" across all applications rather than a specific "vertical" market. Examples include Mann's "EyeTap Digital Eye Glass", a general-purpose seeing aid that does dynamic-range management (HDR vision) and overlays, underlays, simultaneous augmentation and diminishment (e.g. diminishing the electric arc while looking at a welding torch).[85]

Gaming[edit]

Merchlar's mobile game Get On Target uses a trigger image as fiduciary marker

Augmented reality allows gamers to experience digital game play in a real world environment. In the last 10 years there has been a lot of improvements of technology, resulting in better movement detection and the possibility for the Wii to exist, but also direct detection of the player's movements.[86]

Industrial design[edit]

AR can help industrial designers experience a product's design and operation before completion. Volkswagen uses AR for comparing calculated and actual crash test imagery.[87] AR can be used to visualize and modify a car body structure and engine layout. AR can also be used to compare digital mock-ups with physical mock-ups for finding discrepancies between them.[88][89]

Medical[edit]

Augmented Reality can provide the surgeon with information, which are otherwise hidden, such as showing the heartbeat rate, the blood pressure, the state of the patient’s organ, etc. AR can be used to let a doctor look inside a patient by combining one source of images such as an X-ray with another such as video.

Examples include a virtual X-ray view based on prior tomography or on real time images from ultrasound and confocal microscopy probes[90] or visualizing the position of a tumor in the video of an endoscope.[91] AR can enhance viewing a fetus inside a mother's womb.[92] Also, patients wearing Google Glass can be reminded to take medications.[93]

Spatial immersion and interaction[edit]

Augmented reality applications, running on handheld devices utilised as virtual reality headsets, can also digitalise human presence in space and provide a computer generated model of them, in a virtual space where they can interact and perform various actions [6].

See also[edit]

Mixed reality

Military[edit]

Rockwell WorldView Console showing space surveillance telescope video map overlay of satellite flight tracks from a 1993 paper.

In combat, AR can serve as a networked communication system that renders useful battlefield data onto a soldier's goggles in real time. From the soldier's viewpoint, people and various objects can be marked with special indicators to warn of potential dangers. Virtual maps and 360° view camera imaging can also be rendered to aid a soldier's navigation and battlefield perspective, and this can be transmitted to military leaders at a remote command center.[94]

An interesting application of AR occurred when Rockwell International created video map overlays of satellite and orbital debris tracks to aid in space observations at Air Force Maui Optical System. In their 1993 paper "Debris Correlation Using the Rockwell WorldView System" the authors describe the use of map overlays applied to video from space surveillance telescopes. The map overlays indicated the trajectories of various objects in geographic coordinates. This allowed telescope operators to identify satellites, and also to identify - and catalog - potentially dangerous space debris.[95]

Screen capture of SmartCam3D in picture in picture (PIP) mode. This helps sensor operators maintain a broader situation awareness than a telescopic camera "soda-straw". It was shown to essentially double the speed at which points can be located on the ground.

Starting in 2003 the US Army integrated the SmartCam3D augmented reality system into the Shadow Unmanned Aerial System to aid sensor operators using telescopic cameras to locate people or points of interest. The system combined both fixed geographic information including street names, points of interest, airports and railroads with live video from the camera system. The system offered "picture in picture" mode that allows the system to show a synthetic view of the area surrounding the camera's field of view. This helps solve a problem in which the field of view is so narrow that it excludes important context, as if "looking through a soda straw". The system displays real-time friend/foe/neutral location markers blended with live video, providing the operator with improved situation awareness.

Researchers at USAF Research Lab (Calhoun, Draper et al.) found an approximately two-fold increase in the speed at which UAV sensor operators found points of interest using this technology.[96] This ability to maintain geographic awareness quantitatively enhances mission efficiency. The system is in use on the US Army RQ-7 Shadow and the MQ-1C Gray Eagle Unmanned Aerial Systems.

Navigation[edit]

LandForm video map overlay marking runways, road, and buildings during 1999 helicopter flight test.

AR can augment the effectiveness of navigation devices. Information can be displayed on an automobile's windshield indicating destination directions and meter, weather, terrain, road conditions and traffic information as well as alerts to potential hazards in their path.[97][98][99] Aboard maritime vessels, AR can allow bridge watch-standers to continuously monitor important information such as a ship's heading and speed while moving throughout the bridge or performing other tasks.[100]

The NASA X-38 was flown using a Hybrid Synthetic Vision system that overlaid map data on video to provide enhanced navigation for the spacecraft during flight tests from 1998 to 2002. It used the LandForm software and was useful for times of limited visibility, including an instance when the video camera window frosted over leaving astronauts to rely on the map overlays.[101] The LandForm software was also test flown at the Army Yuma Proving Ground in 1999. In the photo at right one can see the map markers indicating runways, air traffic control tower, taxiways, and hangars overlaid on the video.[102]

Office workplace[edit]

AR can help facilitate collaboration among distributed team members in a work force via conferences with real and virtual participants. AR tasks can include brainstorming and discussion meetings utilizing common visualization via touch screen tables, interactive digital whiteboards, shared design spaces, and distributed control rooms.[103][104][105]

Sports and entertainment[edit]

AR has become common in sports telecasting. Sports and entertainment venues are provided with see-through and overlay augmentation through tracked camera feeds for enhanced viewing by the audience. Examples include the yellow "first down" line seen in television broadcasts of American football games showing the line the offensive team must cross to receive a first down. AR is also used in association with football and other sporting events to show commercial advertisements overlaid onto the view of the playing area. Sections of rugby fields and cricket pitches also display sponsored images. Swimming telecasts often add a line across the lanes to indicate the position of the current record holder as a race proceeds to allow viewers to compare the current race to the best performance. Other examples include hockey puck tracking and annotations of racing car performance and snooker ball trajectories. [45][106]

AR can enhance concert and theater performances. For example, artists can allow listeners to augment their listening experience by adding their performance to that of other bands/groups of users.[107][108][109]

The gaming industry has benefited a lot from the development of this technology. A number of games have been developed for prepared indoor environments. Early AR games also include AR air hockey, collaborative combat against virtual enemies, and an AR-enhanced pool games. A significant number of games incorporate AR in them and the introduction of the smartphone has made a bigger impact.[110][111]

Task support[edit]

Complex tasks such as assembly, maintenance, and surgery can be simplified by inserting additional information into the field of view. For example, labels can be displayed on parts of a system to clarify operating instructions for a mechanic who is performing maintenance on the system.[112][113] Assembly lines gain many benefits from the usage of AR. In addition to Boeing, BMW and Volkswagen are known for incorporating this technology in their assembly line to improve their manufacturing and assembly processes.[114][115][116] Big machines are difficult to maintain because of the multiple layers or structures they have. With the use of AR the workers can complete their job in a much easier way because AR permits them to look through the machine as if it was with x-ray, pointing them to the problem right away.[117]

Television[edit]

Weather visualizations were the first application of Augmented Reality to television. It has now become common in weathercasting to display full motion video of images captured in real-time from multiple cameras and other imaging devices. Coupled with 3D graphics symbols and mapped to a common virtual geospace model, these animated visualizations constitute the first true application of AR to TV.

Augmented reality has also become common in sports telecasting. Sports and entertainment venues are provided with see-through and overlay augmentation through tracked camera feeds for enhanced viewing by the audience. Examples include the yellow "first down" line seen in television broadcasts of American football games showing the line the offensive team must cross to receive a first down. AR is also used in association with football and other sporting events to show commercial advertisements overlaid onto the view of the playing area. Sections of rugby fields and cricket pitches also display sponsored images. Swimming telecasts often add a line across the lanes to indicate the position of the current record holder as a race proceeds to allow viewers to compare the current race to the best performance. Other examples include hockey puck tracking and annotations of racing car performance and snooker ball trajectories.[118][119]

Augmented reality is starting to allow Next Generation TV viewers to interact with the programs they are watching. They can place objects into an existing program and interact with these objects, such as moving them around. Avatars of real persons in real time who are also watching the same program.[120]

Tourism and sightseeing[edit]

Augmented reality applications can enhance a user's experience when traveling by providing real time informational displays regarding a location and its features, including comments made by previous visitors of the site. AR applications allow tourists to experience simulations of historical events, places and objects by rendering them into their current view of a landscape.[121][122][123] AR applications can also present location information by audio, announcing features of interest at a particular site as they become visible to the user.[124][125]

Translation[edit]

AR systems can interpret foreign text on signs and menus and, in a user's augmented view, re-display the text in the user's language. Spoken words of a foreign language can be translated and displayed in a user's view as printed subtitles.[126][127][128]

Notable researchers[edit]

  • Ivan Sutherland invented the first AR head-mounted display at Harvard University.
  • Steven Feiner, Professor at Columbia University, is a leading pioneer of augmented reality, and author of the first paper on an AR system prototype, KARMA (the Knowledge-based Augmented Reality Maintenance Assistant), along with Blair MacIntyre and Doree Seligmann.[129]
  • Steve Mann formulated an earlier concept of Mediated reality in the 1970s and 1980s, using cameras, processors, and display systems to modify visual reality to help people see better (dynamic range management), building computerized welding helmets, as well as "Augmediated Reality" vision systems for use in everyday life.[130]
  • Louis Rosenberg developed one of the first known AR systems, called Virtual Fixtures, while working at the U.S. Air Force Armstrong Labs in 1991, and published the first study of how an AR system can enhance human performance.[131] Rosenberg's subsequent work at Stanford University in the early 90's, was the first proof that virtual overlays, when registered and presented over a user's direct view of the real physical world, could significantly enhance human performance.[132][133][134]
  • Mike Abernathy pioneered one of the first successful augmented reality applications of video overlay using map data for space debris in 1993,[95] while at Rockwell International. He co-founded Rapid Imaging Software, Inc. and was the primary author of the LandForm system in 1995, and the SmartCam3D system.[102][101] LandForm augmented reality was successfully flight tested in 1999 aboard a helicopter and SmartCam3D was used to fly the NASA X-38 from 1999-2002. He and NASA colleague Francisco Delgado received the National Defense Industries Association Top5 awards in 2004.[135]
  • Francisco “Frank” Delgado is a NASA engineer and project manager specializing in human interface research and development. Starting 1998 he conducted research into displays that combined video with synthetic vision systems (called hybrid synthetic vision at the time) that we recognize today as augmented reality systems for the control of aircraft and spacecraft. In 1999 he and colleague Mike Abernathy flight-tested the LandForm system aboard a US Army helicopter. Delgado oversaw integration of the LandForm and SmartCam3D systems into the X-38 Crew Return Vehicle.[102][101] In 2001, Aviation Week reported NASA astronaut’s successful use of hybrid synthetic vision (augmented reality) to fly the X-38 during a flight test at Dryden Flight Research Center. The technology was used in all subsequent flights of the X-38. Delgado was co-recipient of the National Defense Industries Association 2004 Top 5 software of the year award for SmartCam3D.[135]
  • Dieter Schmalstieg and Daniel Wagner jump started the field of AR on mobile phones. They developed the first marker tracking systems for mobile phones and PDAs.[136]
  • Bruce H. Thomas and Wayne Piekarski develop the Tinmith system in 1998.[137] They along with Steve Feiner with his MARS system pioneer outdoor augmented reality.
  • Dr. Mark Billinghurst is one of the world’s leading augmented reality researchers, focusing on innovative computer interfaces that explore how virtual and real worlds can be merged. Director of the HIT Lab New Zealand (HIT Lab NZ) at the University of Canterbury in New Zealand, he has produced over 250 technical publications and presented demonstrations and courses at a wide variety of conferences.
  • Reinhold Behringer performed important early work in image registration for augmented reality, and prototype wearable testbeds for augmented reality. He also co-organized the First IEEE International Symposium on Augmented Reality in 1998 (IWAR'98), and co-edited one of the first books on augmented reality.[138][139][140]

History[edit]

  • 1901: L. Frank Baum, an author, first mentions the idea of an electronic display/spectacles that overlays data onto real life (in this case 'people'), it is named a 'character marker'.[141]
  • 1957–62: Morton Heilig, a cinematographer, creates and patents a simulator called Sensorama with visuals, sound, vibration, and smell.[142]
  • 1968: Ivan Sutherland invents the head-mounted display and positions it as a window into a virtual world.[143]
  • 1975: Myron Krueger creates Videoplace to allow users to interact with virtual objects for the first time.
  • 1980: Steve Mann creates the first wearable computer, a computer vision system with text and graphical overlays on a photographically mediated reality, or Augmediated Reality.[144] See EyeTap.
  • 1981: Dan Reitan geospatially maps multiple weather radar images and space-based and studio cameras to virtual reality Earth maps and abstract symbols for television weather broadcasts, bringing Augmented Reality to TV.[145]
  • 1989: Jaron Lanier coins the phrase Virtual Reality and creates the first commercial business around virtual worlds.
  • 1990: The term "'Augmented Reality'" is believed to be attributed to Tom Caudell, a former Boeing researcher.[146]
  • 1992: Louis Rosenberg develops one of the first functioning AR systems, called Virtual Fixtures, at the U.S. Air Force Research Laboratory—Armstrong, and demonstrates benefits to human performance.[131][134][147]
  • 1992: Steven Feiner, Blair MacIntyre and Doree Seligmann present the first major paper on an AR system prototype, KARMA, at the Graphics Interface conference.
  • 1993: Mike Abernathy, et al., report the first use of augmented reality in identifying space debris using Rockwell WorldView by overlaying satellite geographic trajectories on live telescope video.[95]
  • 1993 A widely cited version of the paper above is published in Communications of the ACM – Special issue on computer augmented environments, edited by Pierre Wellner, Wendy Mackay, and Rich Gold.[148]
  • 1993: Loral WDL, with sponsorship from STRICOM, performed the first demonstration combining live AR-equipped vehicles and manned simulators. Unpublished paper, J. Barrilleaux, "Experiences and Observations in Applying Augmented Reality to Live Training", 1999.[149]
  • 1994: Julie Martin creates first 'Augmented Reality Theater production', Dancing In Cyberspace, funded by the Australia Council for the Arts, features dancers and acrobats manipulating body–sized virtual object in real time, projected into the same physical space and performance plane. The acrobats appeared immersed within the virtual object and environments. The installation used Silicon Graphics computers and Polhemus sensing system.
  • 1998: Spatial Augmented Reality introduced at University of North Carolina at Chapel Hill by Ramesh Raskar, Welch, Henry Fuchs.[37]
  • 1999: Frank Delgado, Mike Abernathy et al. report successful flight test of LandForm software video map overlay from a helicopter at Army Yuma Proving Ground overlaying video with runways, taxiways, roads and road names.[102][101]
  • 1999: The US Naval Research Laboratory engage on a decade long research program called the Battlefield Augmented Reality System (BARS) to prototype some of the early wearable systems for dismounted soldier operating in urban environment for situation awareness and training NRL BARS Web page
  • 1999: Hirokazu Kato (加藤 博一) created ARToolKit at HITLab, where AR later was further developed by other HITLab scientists, demonstrating it at SIGGRAPH.
  • 2000: Bruce H. Thomas develops ARQuake, the first outdoor mobile AR game, demonstrating it in the International Symposium on Wearable Computers.
  • 2001: NASA X-38 flown using LandForm software video map overlays at Dryden Flight Research Center.[150]
  • 2008: Wikitude AR Travel Guide launches on 20 Oct 2008 with the G1 Android phone.[151]
  • 2009: ARToolkit was ported to Adobe Flash (FLARToolkit) by Saqoosha, bringing augmented reality to the web browser.[152]
  • 2013: Google announces an open beta test of its Google Glass augmented reality glasses. The glasses reach the Internet through Bluetooth, which connects to the wireless service on a user’s cellphone. The glasses respond when a user speaks, touches the frame or moves the head.[153]

See also[edit]

References[edit]

  1. ^ Graham, M., Zook, M., and Boulton, A. "Augmented reality in urban places: contested content and the duplicity of code." Transactions of the Institute of British Geographers, DOI: 10.1111/j.1475-5661.2012.00539.x 2012.
  2. ^ Steuer, Jonathan. Defining Virtual Reality: Dimensions Determining Telepresence, Department of Communication, Stanford University. 15 October 1993.
  3. ^ Introducing Virtual Environments National Center for Supercomputing Applications, University of Illinois.
  4. ^ Chen, Brian X. If You’re Not Seeing Data, You’re Not Seeing, Wired, 25 August 2009.
  5. ^ Maxwell, Kerry. Augmented Reality, Macmillan Dictionary Buzzword.
  6. ^ Augmented reality-Everything about AR, Augmented Reality On.
  7. ^ a b Azuma, Ronald. A Survey of Augmented Reality Presence: Teleoperators and Virtual Environments, pp. 355–385, August 1997.
  8. ^ Zhanpeng, H.,Pan H., et al. Mobile augmented reality survey: a bottom-up approach.
  9. ^ Metz, Rachel. Augmented Reality Is Finally Getting Real Technology Review, 2 August 2012.
  10. ^ Fleet Week: Office of Naval Research Technology- Virtual Reality Welder Training, eweek, 28 May 2012.
  11. ^ Rolland, Jannick; Baillott, Yohan; Goon, Alexei.A Survey of Tracking Technology for Virtual Environments, Center for Research and Education in Optics and Lasers, University of Central Florida.
  12. ^ Klepper, Sebastian.Augmented Reality – Display Systems.
  13. ^ Rolland, J; Biocca F; Hamza-Lup F; Yanggang H; Martins R (October 2005). "Development of Head-Mounted Projection Displays for Distributed, Collaborative, Augmented Reality Applications". Presence: Teleoperators & Virtual Environments 14 (5): 528–549. 
  14. ^ Grifatini, Kristina. Augmented Reality Goggles, Technology Review 10 November 2010.
  15. ^ Arthur, Charles. UK company's 'augmented reality' glasses could be better than Google's, The Guardian, 10 September 2012.
  16. ^ Gannes, Liz. "Google Unveils Project Glass: Wearable Augmented-Reality Glasses". http://allthingsd.com. Retrieved 2012-04-04. , All Things D.
  17. ^ Benedetti, Winda. Xbox leak reveals Kinect 2, augmented reality glasses NBC News.
  18. ^ "Google Glass Alternative". Digital Trends. Retrieved 15 November 2013. 
  19. ^ "Some of Google Glass Alternative". Premier Logic. Retrieved 15 November 2013. 
  20. ^ "5 Google Glass Alternatives". NBC News. Retrieved 15 November 2013. 
  21. ^ "How Crowdoptic’s big data technology reveals the world’s most popular photo objects". VentureBeat. Retrieved 6 June 2013. 
  22. ^ a b "CrowdOptic and L'Oreal To Make History By Demonstrating How Augmented Reality Can Be A Shared Experience". Forbes. Retrieved 6 June 2013. 
  23. ^ Greenemeier, Larry. Computerized Contact Lenses Could Enable In-Eye Augmented Reality. Scientific American, 23 November 2011.
  24. ^ Yoneda, Yuka. Solar Powered Augmented Contact Lenses Cover Your Eye with 100s of LEDs. inhabitat, 17 March 2010.
  25. ^ Rosen, Kenneth. "Contact Lenses Can Display Your Text Messages". Mashable.com. Mashable.com. Retrieved 2012-12-13. 
  26. ^ O'Neil, Lauren. "LCD contact lenses could display text messages in your eye". CBC. Retrieved 2012-12-12. 
  27. ^ Anthony, Sebastian. US military developing multi-focus augmented reality contact lenses. ExtremeTech, 13 April 2012.
  28. ^ Bernstein, Joseph. 2012 Invention Awards: Augmented-Reality Contact Lenses Popular Science, 5 June 2012.
  29. ^ Augmented World Expo Conference. [1]. AR Conference, 15 November 2013.
  30. ^ A Futuristic Short Film: by Sight Systems. [2]. Sight, 15 November 2013.
  31. ^ Tidwell, Michael; Johnson, Richard S.; Melville, David; Furness, Thomas A.The Virtual Retinal Display – A Retinal Scanning Imaging System, Human Interface Technology Laboratory, University of Washington.
  32. ^ a b “GlassEyes”: The Theory of EyeTap Digital Eye Glass, supplemental material for IEEE Technology and Society, Volume Vol. 31, Number 3, 2012, pp. 10-14.
  33. ^ "Intelligent Image Processing", John Wiley and Sons, 2001, ISBN 0-471-40637-6, 384 p.
  34. ^ Marker vs Markerless AR, Dartmouth College Library.
  35. ^ Feiner, Steve. "Augmented reality: a long way off?". AR Week. Pocket-lint. Retrieved 2011-03-03. 
  36. ^ Bimber, Oliver; Encarnação, Miguel; Branco, Pedro. The Extended Virtual Table: An Optical Extension for Table-Like Projection Systems, MIT Press Journal Vol. 10, No. 6, Pages 613–631, March 13, 2006.
  37. ^ a b Ramesh Raskar, Greg Welch, Henry Fuchs Spatially Augmented Reality, First International Workshop on Augmented Reality, Sept 1998.
  38. ^ Knight, Will. Augmented reality brings maps to life 19 July 2005.
  39. ^ Sung, Dan. Augmented reality in action – maintenance and repair. Pocket-lint, 1 March 2011.
  40. ^ Stationary systems can employ 6DOF track systems such as Polhemus, ViCON, A.R.T, or Ascension.
  41. ^ Marshall, Gary.Beyond the mouse: how input is evolving, Touch,voice and gesture recognition and augmented realitytechradar.computing\PC Plus 23 August 2009.
  42. ^ Simonite, Tom. Augmented Reality Meets Gesture Recognition, Technology Review, 15 September 2011.
  43. ^ Chaves, Thiago; Figueiredo, Lucas; Da Gama, Alana; de Araujo, Christiano; Teichrieb, Veronica. Human Body Motion and Gestures Recognition Based on Checkpoints. SVR '12 Proceedings of the 2012 14th Symposium on Virtual and Augmented Reality pp. 271–278.
  44. ^ Barrie, Peter; Komninos, Andreas; Mandrychenko, Oleksii.A Pervasive Gesture-Driven Augmented Reality Prototype using Wireless Sensor Body Area Networks.
  45. ^ a b Azuma, Ronald; Balliot, Yohan; Behringer, Reinhold; Feiner, Steven; Julier, Simon; MacIntyre, Blair. Recent Advances in Augmented Reality Computers & Graphics, November 2001.
  46. ^ Maida, James; Bowen, Charles; Montpool, Andrew; Pace, John. Dynamic registration correction in augmented-reality systems, Space Life Sciences, NASA.
  47. ^ State, Andrei; Hirota, Gentaro; Chen,David T; Garrett, William; Livingston, Mark. Superior Augmented Reality Registration by Integrating Landmark Tracking and Magnetic Tracking, Department of Computer ScienceUniversity of North Carolina at Chapel Hill.
  48. ^ Bajura, Michael; Neumann, Ulrich. Dynamic Registration Correction in Augmented-Reality Systems University of North Carolina, University of Southern California.
  49. ^ "ARML 2.0 SWG". Open Geospatial Consortium website. Open Geospatial Consortium. Retrieved 12 November 2013. 
  50. ^ "Top 5 AR SDKs". Augmented Reality News. Retrieved 15 November 2013. 
  51. ^ "Top 10 AR SDKs". Augmented World Expo. Retrieved 15 November 2013. 
  52. ^ Zhanpeng, H., WeiKai, L., Pan, H., Christoph, P. [3]. CloudRidAR: A Cloud-based Architecture for Mobile Augmented Reality Proceeding of MARS'14, July 2014.
  53. ^ "Metaio AR SDK". Metaio. Retrieved 15 November 2013. 
  54. ^ "Vuforia AR SDK". Vuforia. Retrieved 15 November 2013. 
  55. ^ "Mobinett AR SDK". Mobinett. Retrieved 15 November 2014. 
  56. ^ "Wikitude AR SDK". Wikitude. Retrieved 15 November 2013. 
  57. ^ "Layar AR SDK". Layar. Retrieved 15 November 2013. 
  58. ^ Augmented Reality Landscape 11 August 2012.
  59. ^ Stuart Eve. "Augmenting Phenomenology: Using Augmented Reality to Aid Archaeological Phenomenology in the Landscape". Retrieved 2012-09-25. 
  60. ^ Dähne, Patrick; Karigiannis, John N. "Archeoguide: System Architecture of a Mobile Outdoor Augmented Reality System". Retrieved 2010-01-06. 
  61. ^ Divecha, Devina.Augmented Reality (AR) used in architecture and design. designMENA 8 September 2011.
  62. ^ Architectural dreams in augmented reality. University News, University of Western Australia. 5 March 2012.
  63. ^ Webley, Kayla. The 50 Best Inventions of 2010 – EyeWriter Time, 11 November 2010.
  64. ^ Alexander, Michael.Arbua Shoco Owl Silver Coin with Augmented Reality, Coin Update July 20, 2012.
  65. ^ Royal Mint produces revolutionary commemorative coin for Aruba, Today August 7, 2012.
  66. ^ Humphries, Mathew.[4].Geek.com 19 September 2011.
  67. ^ Netburn, Deborah.Ikea introduces augmented reality app for 2013 catalog. Los Angeles Times, 23 July 2012.
  68. ^ Saenz, Aaron.Virtual Mirror Brings Augmented Reality to Makeup Counters. singularityHub, 15 June 2010.
  69. ^ Katts, Rima. Elizabeth Arden brings new fragrance to life with augmented reality Mobile Marketer, 19 September 2012.
  70. ^ Meyer, David. Telefónica bets on augmented reality with Aurasma tie-in gigaom, 17 September 2012.
  71. ^ Mardle, Pamela.Video becomes reality for Stuprint.com. Printweek, 3 October 2012.
  72. ^ Churcher, Jason. "Internal accuracy vs external accuracy". Retrieved 7 May 2013. 
  73. ^ Lee, Gun (2012). CityViewAR outdoor AR visualization. ACM. p. 97. ISBN 978-1-4503-1474-9. 
  74. ^ Groundbreaking Augmented Reality-Based Reading Curriculum Launches, ‘’PRweb’’, 23 October 2011.
  75. ^ Stewart-Smith, Hanna. Education with Augmented Reality: AR textbooks released in Japan, ‘’ZDnet’’, 4 April 2012.
  76. ^ Augmented reality in education smarter learning.
  77. ^ Lubrecht, Anna. Augmented Reality for Education ‘’Digital Union’’, The Ohio State University 24 April 2012.
  78. ^ http://acdc.sav.us.es/pixelbit/images/stories/p41/15.pdf
  79. ^ Maier, Patrick; Tönnis, Marcus; Klinker, Gudron. Augmented Reality for teaching spatial relations, Conference of the International Journal of Arts & Sciences (Toronto 2009).
  80. ^ Vuforia Case Study: Anatomy 4D
  81. ^ Kaufmann, Hannes. Collaborative Augmented Reality in Education, Institute of Software Technology and Interactive Systems, Vienna University of Technology.
  82. ^ "Augmented Reality--Emerging Technology for Emergency Management", Emergency Management Magazine, September 24, 2009
  83. ^ "What Does the Future Hold for Emergency Management?", Emergency Management Magazine, November 8, 2013
  84. ^ Cooper, J., “SUPPORTING FLIGHT CONTROL FOR UAV-ASSISTED WILDERNESS SEARCH AND RESCUE THROUGH HUMAN CENTERED INTERFACE DESIGN”, Thesis, Brigham Young University, DEC 2007
  85. ^ Davies, Chris (2012-09-12). "Quantigraphic camera promises HDR eyesight from Father of AR". SlashGear. Retrieved 2012-12-30. 
  86. ^ "YOUR THOUGHTS ABOUT AUGMENTED REALITY IN VIDEO GAMES". 2013-05-01. Retrieved 2013-05-07. 
  87. ^ Noelle, S. (2002). "Stereo augmentation of simulation results on a projection wall". Mixed and Augmented Reality, 2002. ISMAR 2002. Proceedings.: 271–322. Retrieved 2012-10-07. 
  88. ^ Verlinden, Jouke; Horvath, Imre. "Augmented Prototyping as Design Means in Industrial Design Engineering". Delft University of Technology. Retrieved 2012-10-07. 
  89. ^ Pang, Y; Nee, A; Youcef-Toumie, Kamal; Ong, S.K; Yuan, M.L (November 18, 2004). "Assembly Design and Evaluation in an Augmented Reality Environment". National University of Singapore, M.I.T. Retrieved 2012-10-07. 
  90. ^ Mountney P, Giannarou S, Elson D, Yang GZ (2009). "Optical biopsy mapping for minimally invasive cancer screening". Medical Image Computing and Computer-assisted Intervention 12 (Pt 1): 483–90. PMID 20426023. 
  91. ^ Scopis Augmented Reality: Path guidance to craniopharyngioma on YouTube
  92. ^ "UNC Ultrasound/Medical Augmented Reality Research". Archived from the original on 12 February 2010. Retrieved 2010-01-06. 
  93. ^ "Augmented Reality Revolutionizing Medicine". Health Tech Event. Retrieved 9 October 2014. 
  94. ^ Cameron, Chris. Military-Grade Augmented Reality Could Redefine Modern Warfare ReadWriteWeb June 11, 2010.
  95. ^ a b c Abernathy, M., Houchard, J., Puccetti, M., and Lambert, J,"Debris Correlation Using the Rockwell WorldView System",Proceedings of 1993 Space Surveillance Workshop 30 March to 1 April 1993,pages 189-195
  96. ^ Calhoun, G. L., Draper, M. H., Abernathy, M. F., Delgado, F., and Patzek, M. “Synthetic Vision System for Improving Unmanned Aerial Vehicle Operator Situation Awareness,” 2005 Proceedings of SPIE Enhanced and Synthetic Vision, Vol. 5802, pp. 219–230.
  97. ^ GM's Enhanced Vision System. Techcrunch.com (17 March 2010). Retrieved 9 June 2012.
  98. ^ Couts, Andrew. New augmented reality system shows 3D GPS navigation through your windshield Digital Trens,27 October 2011.
  99. ^ Griggs, Brandon. Augmented-reality' windshields and the future of driving CNN Tech, 13 January 2012.
  100. ^ Cheney-Peters, Scott (12 April 2012). "CIMSEC: Google's AR Goggles". Retrieved 2012-04-20. 
  101. ^ a b c d Delgado, F., Abernathy, M., White J., and Lowrey, B. Real-Time 3-D Flight Guidance with Terrain for the X-38,SPIE Enhanced and Synthetic Vision 1999, Orlando Florida, April 1999, Proceedings of the SPIE Vol. 3691, pages 149-156
  102. ^ a b c d Delgado, F., Altman, S., Abernathy, M., White, J. Virtual Cockpit Window for the X-38,SPIE Enhanced and Synthetic Vision 2000, Orlando Florida, Proceedings of the SPIE Vol. 4023, pages 63-70
  103. ^ Stafford, Aaron; Piekarski, Wayne; Thomas, Bruce H. "Hand of God". Archived from the original on 2009-12-07. Retrieved 2009-12-18. 
  104. ^ Benford, S, Greenhalgh, C, Reynard, G, Brown, C and Koleva, B. Understanding and constructing shared spaces with mixed-reality boundaries. ACM Trans. Computer-Human Interaction, 5(3):185–223, Sep. 1998.
  105. ^ Office of Tomorrow Media Interaction Lab.
  106. ^ Marlow, Chris. Hey, hockey puck! NHL PrePlay adds a second-screen experience to live games, digitalmediawire April 27, 2012.
  107. ^ Pair, J.; Wilson, J.; Chastine, J.; Gandy, M. "The Duran Duran Project: The Augmented Reality Toolkit in Live Performance". The First IEEE International Augmented Reality Toolkit Workshop, 2002.
  108. ^ Broughall, Nick. Sydney Band Uses Augmented Reality For Video Clip. Gizmodo, 19 October 2009.
  109. ^ Pendlebury, Ty. Augmented reality in Aussie film clip. c|net 19 October 2009.
  110. ^ Hawkins, Mathew. Augmented Reality Used To Enhance Both Pool And Air Hockey Game Set WatchOctober 15, 2011.
  111. ^ One Week Only – Augmented Reality Project Combat-HELO Dev Blog July 31, 2012.
  112. ^ The big idea:Augmented Reality. Ngm.nationalgeographic.com (15 May 2012). Retrieved 2012-06-09.
  113. ^ Henderson, Steve; Feiner, Steven. "Augmented Reality for Maintenance and Repair (ARMAR)". Retrieved 2010-01-06. 
  114. ^ Sandgren, Jeffrey. The Augmented Eye of the Beholder, BrandTech News January 8, 2011.
  115. ^ Cameron, Chris. Augmented Reality for Marketers and Developers, ReadWriteWeb.
  116. ^ Dillow, Clay BMW Augmented Reality Glasses Help Average Joes Make Repairs, Popular Science September 2009.
  117. ^ King, Rachael. Augmented Reality Goes Mobile, Bloomberg Business Week Technology November 3, 2009.
  118. ^ [42]
  119. ^ [81]
  120. ^ [93]
  121. ^ Saenz, Aaron Augmented Reality Does Time Travel Tourism SingularityHUB November 19, 2009.
  122. ^ Sung, Dan Augmented reality in action – travel and tourism Pocket-lint March 2, 2011.
  123. ^ Dawson, Jim Augmented Reality Reveals History to Tourists Life Science August 16, 2009.
  124. ^ Bartie, P and Mackaness, W.Development of a speech-based augmented reality system to support exploration of cityscape. Trans. GIS, 10(1):63–86, 2006.
  125. ^ Benderson, Bejamin B. Audio Augmented Reality: A Prototype Automated Tour Guide Bell Communications Research,, ACM Human Computer in Computing Systems conference, pp. 210–211.
  126. ^ Tsotsis, Alexia. Word Lens Translates Words Inside of Images. Yes Really. TechCrunch (16 December 2010).
  127. ^ N.B. Word Lens: This changes everything The Economist: Gulliver blog 18 December 2010.
  128. ^ Borghino, Dario Augmented reality glasses perform real-time language translation. gizmag, 29 July 2012.
  129. ^ "Knowledge-based augmented reality". ACM. July 1993. 
  130. ^ "Wearable Computing: A first step towards personal imaging", IEEE Computer, pp. 25–32, Vol. 30, Issue 2, Feb. 1997 link.
  131. ^ a b L. B. Rosenberg. The Use of Virtual Fixtures As Perceptual Overlays to Enhance Operator Performance in Remote Environments. Technical Report AL-TR-0089, USAF Armstrong Laboratory, Wright-Patterson AFB OH, 1992.
  132. ^ Rosenberg, L., "Virtual fixtures as tools to enhance operator performance in telepresence environments," SPIE Manipulator Technology, 1993.
  133. ^ Rosenberg, "Virtual Haptic Overlays Enhance Performance in Telepresence Tasks," Dept. of Mech. Eng., Stanford Univ., 1994.
  134. ^ a b Rosenberg, "Virtual Fixtures: Perceptual Overlays Enhance Operator Performance in Telepresence Tasks," Ph.D. Dissertation, Stanford University.
  135. ^ a b C. Segura E. George F. Doherty J. H. Lindley M. W. Evans "SmartCam3D Provides New Levels of Situation Awareness", CrossTalk: The Journal of Defense Software Engineering. Volume 18, Number 9, Page 10-11.
  136. ^ Wagner, Daniel (29 September 2009). "First Steps Towards Handheld Augmented Reality". ACM. Retrieved 2009-09-29. 
  137. ^ Piekarski, William; Thomas, Bruce. Tinmith-Metro: New Outdoor Techniques for Creating City Models with an Augmented Reality Wearable Computer Fifth International Symposium on Wearable Computers (ISWC'01), 2001, pp. 31.
  138. ^ Behringer, R.;Improving the Registration Precision by Visual Horizon Silhouette Matching. Rockwell Science Center.
  139. ^ Behringer, R.;Tam, C; McGee, J.; Sundareswaran, V.; Vassiliou, Marius. Two Wearable Testbeds for Augmented Reality: itWARNS and WIMMIS. ISWC 2000, Atlanta, 16–17 October 2000.
  140. ^ R. Behringer, G. Klinker,. D. Mizell. Augmented Reality – Placing Artificial Objects in Real Scenes. Proceedings of IWAR '98. A.K.Peters, Natick, 1999. ISBN 1-56881-098-9.
  141. ^ Johnson, Joel. “The Master Key”: L. Frank Baum envisions augmented reality glasses in 1901 Mote & Beam 10 September 2012.
  142. ^ http://www.google.com/patents?q=3050870
  143. ^ http://90.146.8.18/en/archiv_files/19902/E1990b_123.pdf
  144. ^ Mann, Steve (2012-11-02). "Eye Am a Camera: Surveillance and Sousveillance in the Glassage". Techland.time.com. Retrieved 2013-10-14. 
  145. ^ http://www.etceter.com/c-news/p-google-glasses-project/
  146. ^ Lee, Kangdon (March 2012). "Augmented Reality in Education and Training". Techtrends: Linking Research & Practice To Improve Learning 56 (2). Retrieved 2014-05-15. 
  147. ^ L. B. Rosenberg, "The Use of Virtual Fixtures to Enhance Operator Performance in Telepresence Environments" SPIE Telemanipulator Technology, 1993.
  148. ^ Wellner, Pierre. "Computer Augmented Environments: back to the real world". ACM. Retrieved 2012-07-28. 
  149. ^ Barrilleaux, Jon. Experiences and Observations in Applying Augmented Reality to Live Training. Jmbaai.com. Retrieved 2012-06-09.
  150. ^ AviationNow.com Staff, "X-38 Test Features Use Of Hybrid Synthetic Vision" AviationNow.com, December 11, 2001
  151. ^ Wikitude AR Travel Guide. Youtube.com. Retrieved 2012-06-09.
  152. ^ Cameron, Chris. Flash-based AR Gets High-Quality Markerless Upgrade, ReadWriteWeb 9 July 2010.
  153. ^ Miller, Claire. [5], New York Times 20 February 2013.

External links                                              [edit]

Augmented reality at DMOZ Media related to Augmented reality at Wikimedia Commons