Monday 29 October 2012

Paste augmented-reality video graffiti on the streets

LOOK closely and you can find digital graffiti right under your noseMovie Camera: people have daubed videos, animations and comments over buildings and streets around the world. Soon there could be a new type of tag in our cities - cut-and-pasted people - thanks to a technique for editing augmented reality (AR) videos.

Using the AR apps available for smartphones or tablets, anybody can overlay digital text, video and graphics onto the physical world for others to see later. Most major cities are teeming with these digital annotations. You just need to identify a tagged location using your smartphone's map, and watch through the camera using an AR app. Hey presto, a video or animation will then be overlaid on the scene.

Yet if somebody wants to annotate a place with video that they've filmed themselves, today's apps are constrained. They can only overlay a YouTube clip, say, in its original rectangular shape. Now Tobias Langlotz of Graz University of Technology, Austria, and colleagues have designed software that can cut a person or an object out of a video, so that they alone can be pasted as a digital overlay. The idea is to make virtual human guides that could offer city tours or how-to demos, as well as enhancing AR games.

Langlotz and colleagues used a computer-imaging technique called foreground-background segmentation to identify the required foreground object - usually a person. So a user would film a video, then simply point to the object they wanted to extract. The software would do the rest. In a demo, they filmed a skateboarder doing a jump, and showed how he could be pasted onto a street scene. When the app "sees" the environment, it can replay the person in the right place, skating along the ground, for example.

Although they demonstrated the concept using a PC, the researchers say it would be possible to put the software into a smartphone or tablet app. Their work will be presented at OzCHI, the Australian conference on computer-human interaction in Melbourne next month.

Matt Mills of AR company Aurasma in Cambridge, UK, agrees that the tool could be handy for making AR-driven how-to guides. And while it's at an early stage, he adds, it may go on to find wider uses too.

If you would like to reuse any content from New Scientist, either in print or online, please contact the syndication department first for permission. New Scientist does not own rights to photos, but there are a variety of licensing options available for use of articles and graphics we own the copyright to.

Have your say

Only subscribers may leave comments on this article. Please log in.

Only personal subscribers may leave comments on this article

Subscribe now to comment.

All comments should respect the New Scientist House Rules. If you think a particular comment breaks these rules then please use the "Report" link in that comment to report it to us.

If you are having a technical problem posting a comment, please contact technical support.

Source: http://feeds.newscientist.com/c/749/f/10897/s/24f17798/l/0L0Snewscientist0N0Carticle0Cmg216288860B0A0A0A0Epaste0Eaugmentedreality0Evideo0Egraffiti0Eon0Ethe0Estreets0Bhtml0Dcmpid0FRSS0QNSNS0Q20A120EGLOBAL0Qonline0Enews/story01.htm

sag awards 2012 nominees sag awards pro bowl 2012 roster yamaguchi road house occupy oakland occupy oakland

No comments:

Post a Comment