Here’s a fascinating project that started with a great idea and piled on a remarkable amount of innovation. Graffiti Analysis is a project that captures gestures used to create graffiti art and codifies them through a data-type called Graffiti Markup Language (GML). After the break you can watch a video showing the data capture method used in version 2.0 of the project. A marker taped to a light source is used to draw a graffiti tag on a piece of paper. The paper rests on a plexiglass drawing surface with a webcam tracking the underside in order to capture each motion.
The newest iteration, version 3.0, has some unbelievable features. The addition of audio input means that the markup can be projected and animated based on sound, with the example of graffiti interacting with a fireworks show. The 3D tools are quite amazing too, allowing not only for stereoscopic video playback, but for printing out graffiti markup using a 3D printer. The collection of new features is so vast, and produces such amazing results it’s hard to put into words. So we’ve also embedded the demo of the freshly released version after the break.
Continue reading “Graffiti Analysis”
This two handed glove input setup, by [Sean Chen] and [Evan Levine], is one step closer to achieving that [Tony Stark] like workstation; IE, interacting with software in 3D with simple hand gestures. Dubbed the Mister Gloves, the system incorporates accelerometer, push button, and flex sensor data over RF where an MCU converts it to a standard USB device, meaning no drivers are needed and a windows PC can recognize it as a standard keyboard and mouse. Catch a video of Mister Gloves playing portal after the jump.
While amazing, we’re left wondering if gesture setups are really viable options considering one’s arm(s) surly would get tired? Continue reading “Mister Gloves, gesture input”
MIT is debuting their latest advancement in technology, a multitouch screen that also functions as a gestural interface. The multitouch aspect is nothing new, the team explains how traditional interfaces using LEDs or camera systems do work, but fail to recognize gestures off-screen.
Gestures are a relatively recent highlight with the introduction of projects like Natal or perspective tracking, but fail to work at closer distances to the screen. MIT has done what seems the impossible by combining and modifying the two to produce the first ever multitouch close proximity gestural display.
And to think, just a couple of months ago the same school was playing with pop-up books.
This new video about [Pranav Mistry’s] SixthSense project doesn’t bring us much that we haven’t seen before. At least, not on that project. What really caught our eye was the device he shows off at the beginning of the video. Using two old ball mice, he constructed a grip style input device. It is simple and elegant and we can definitely see using this in future hacks. Not only is it cheap and apparently effective, it seems as though it could be constructed in a very short amount of time. all you need are the wheels that spin when the ball moves, 4 springs and some string. Why didn’t we think of that?
We’re filing this one under “best interface implementation”. This robot is controlled by finger gestures on the surface of an iPod Touch. It can walk forward, turn, sidestep, jump, and kick a ball based on the input it receives from your sweaty digits. Unlike vehicles controlled by an iPhone (or by Power Wheels), this has some potential. Especially considering the inevitable proliferation of multi-touch devices in our everyday lives.
[Benjamin] submitted this slick project. It’s a gesture based control unit for the ipod and iphone. It plugs into the dock port and allows you to control the track and volume with simple gestures. While accellerometer equipped units can already “shake to shuffle”, they lack the ability to simply skip tracks forward or backward. He notes that with an accellerometer, simple gestures can be harder to decipher than with a gyro. The gyro gives the ability to tell which direction you are twisting it, so it’s easier to utilize. [Benjamin] was previously covered when he released the iPodGPS.
Some grad students at Duke University have been working on a new tool for cell phones equipped with accellerometers. The software called Phonepoint Pen, allows you to write with your phone in the air. Though we don’t find the applications they mention very practical, we could see this being very nice for application navigation. If you could program a 3 dimentional gesture to load certain apps, that would be nice.