Forty years after the first pioneering step in this field, we now have people that are crafting augmented reality applications for daily use, not only for special moments, like seeing a movie or listening to a song. The augmented reality applications augment and modify the reality using the object recognition technology.
You’d be surprised to find out that augmented reality apps for Android actually have a practical use. Are you a football lover? An augmented reality application can mark different areas so that you could understand easier who’s breaking and who’s following the rules. Do you remember Terminator and how he used to scan people? Well, an augmented reality application is trying to do just that.
A Canadian publisher has launched a campaign targeting 12-17 year old girls using an augmented reality app to enhance the traditional experience of reading books. The campaign was launched with an in store promotion that allowed users to scan a bookcase to find hidden treasure (follows the book’s story line as well) and if they didn’t have a smart phone with them there was one there in the store to use. We are seeing lots of these sorts of interactive campaigns pop up from both book publishers and traditional print media as they aim to bridge the gap between traditional media and the new wave of technology coming along. Ultimately people are going to end up reading the books on their phones and tablets but it’s refreshing to see companies trying to come up with an innovative alternative that appeals to a whole new audience.
Projection-based input seems to be all the buzz of Microsoft Research this year as a number of such projects were shown off not too long ago at the 20th anniversary of the research group. Although KinectFusion, OmniTouch and HoloDesk got the most press coverage, another research project does a really good job at bringing all the different technologies together.
If I may use a cartoon analogy, “Augmented Projector” is like the Captain Planet equivalent of projection-based human-computer interaction. In essence it seems to combine the results of other experimental projects – Kinect-based environment modelling, projection-based gesture interaction and spatial-aware projections into one super project.