The basics of building a multitouch table

Here is a bare-bones multitouch table setup. We looked in on [Seth Sandler's] multitouch work a few years ago when he completed the MTmini build. He’s scaling up the size a bit with the MTbiggie, and showing you how easy it is to put together. The demo rig seen above is just a couple of chairs, a sheet of acrylic, a mirror, a projector, a computer, and a diy infrared webcam.

The rig uses ambient infrared light to detect the outlines of your fingers when they touch the acrylic surface. A webcam with an exposed camera film filter feeds an image of the infrared light received below the surface to the computer. The incoming video is processed using Community Core Vision, where each individual point is isolated and mapped. Once the data is available the sky’s the limit on what you can develop. [Seth's] demo packages include a mouse driver, some physics applications, an Angry Birds implementation, and a few others. See for yourself in the video after the break.

Continue reading “The basics of building a multitouch table”

Subcycles: Multitouch music controller

Subcycles is a sound controller application that [Christian] is using on the third multitouch display that he built. The screen is a sheet of acrylic in an aluminum frame. The image is rear projected onto an area covered with Digiline dispersion film. As with other projects that use the Community Core Vision package, a PS3 eye camera captures the touch information.

This build does a great job of including the audience in what the musician on stage is doing. [Chris] points out that the sight of artists staring at laptops on stage is becoming more and more common. The ‘Minority Report’-like interface that Subcycles uses makes not just for interesting music, but for an added visual reinforcement to the live part of the performance.

[Thanks Mark]