Multitouch Pinball Instrument

[youtube=http://www.youtube.com/watch?v=VpDltKLZSrM]

This is an interesting twist on multi touch input. [s9lucas] is using a simple multitouch system as an instrument that uses solenoids and pinballs to make music.  He has taken the basic multitouch tutorial and combined it with an Arduino bell tower. The end result is possibly more fun to watch than it is to listen to, but that’s fine.

Nintendo DS OSC Support

dsmi

OpenSound Control protocol is an emerging standard for communication between musical programs. It’s meant to replace MIDI. The DSMI, DS Music Interface, team has just added support for OSC. You can now use your DS as generic OSC music controller over WiFi. OSC has TCP/IP support built in, so there is no need to run a host sever to talk to DSMI like you did when they only supported MIDI. We’ve seen OSC used in other projects like the monome. It’s also the basis for the multitouch communication protocol TUIO.

[via CDM]

Tbeta, Cross-platform Multitouch

[vimeo=http://vimeo.com/2034557]

tbeta is a new tool developed by the NUI Group Community. tbeta acts as an image processing layer to take in image data and output tracking data for multitouch applications. Whether FTIR or DI, scratch built multitouch systems generate IR video streams that need to be processed to find fingertips. tbeta can take this or any arbitrary video stream and run it through a series of filters to generate the touch data. This data is sent as OSC TUIO, a standard protocol for touch events. Along with the camera and input switcher, tbeta also aids in system calibration. I works on Windows, OSX, and Linux. Have a look at the getting started guide for a better idea of how it works.

[via CDM]

TISCH, Multitouch Framework

multitouch

[floe] wrote in to tell us about his multitouch based thesis work. While many projects have focused on the hardware side of multitouch, TISCH is designed to promote the software side. TISCH is a multiplatform library that features hardware abstraction and gesture recognition. This takes a lot of weight off of widget developers since they can specify known library gestures instead of writing the exact motions from scratch. Using TISCH also means a standard set of gestures across multiple widgets, so the learning curve will be much easier when a user tries out a new app. If you’re researching multitouch, check out this project and help improve the codebase.

G1 Multitouch Proof Of Concept

[youtube=http://www.youtube.com/watch?v=pSBYqmWVqeM]

[RyeBrye] has been trying to get multitouch working on the Android based T-Mobile G1. He hacked the Synaptics touchscreen driver so that it would dump raw event info to a character device. The demo above is using example code from Google for a fingerpaint program. Polling the device is not the fastest method, but [RyeBrye] just wanted to get a demo out there to prove it could be done.

G-speak Spatial Operating Environment

[vimeo 2229299]

Our fascination with multitouch is fairly well known, but it expands even further to cover all sorts of man machine interaction. Embedded above is a tech demo of g-speak, a spatial operating environment. The user combines gestures and spatial location to interact with on screen objects. If it seems familiar, it’s because one of the company’s founders advised on Minority Report. We doubt all this hand waving is going to catch on very quickly though. Our bet is on someone developing a multitouch Cintiq style device for people to use as a secondary monitor. It would bridge the gap between between our standard 2D interactions and gestures without making a full leap to 3D metaphors.

[via Create Digital Motion]

IR Glass Level Detection

[youtube=http://www.youtube.com/watch?v=DG3_2lvoL8A]

[Johnny Lee]’s colleague [Paul Dietz] has done some interesting work using interactive tables. He’s specifically researched how to determine how full a drink glass is. In the video above, he’s using Microsoft’s Surface, but this technique should work with any IR camera based multitouch table. Determining the drink level requires custom glassware that has a small prism inside. When the liquid level is above the prism, light passes through, but when it’s below the top it reflects more IR light back into the table. Using this information, restaurant staff could serve drinks in a more efficient manner.

[Paul] has worked on another project that uses RFID and capacitive sensing to a similar effect.