The Basics Of Building A Multitouch Table

Here is a bare-bones multitouch table setup. We looked in on [Seth Sandler’s] multitouch work a few years ago when he completed the MTmini build. He’s scaling up the size a bit with the MTbiggie, and showing you how easy it is to put together. The demo rig seen above is just a couple of chairs, a sheet of acrylic, a mirror, a projector, a computer, and a diy infrared webcam.

The rig uses ambient infrared light to detect the outlines of your fingers when they touch the acrylic surface. A webcam with an exposed camera film filter feeds an image of the infrared light received below the surface to the computer. The incoming video is processed using Community Core Vision, where each individual point is isolated and mapped. Once the data is available the sky’s the limit on what you can develop. [Seth’s] demo packages include a mouse driver, some physics applications, an Angry Birds implementation, and a few others. See for yourself in the video after the break.

Continue reading “The Basics Of Building A Multitouch Table”

Converting A Scanner Into A Touchscreen

[Sprite_TM] was cleaning up his hacking workbench when he came across an all-in-one device that had seen better days. After a bit of consideration he decided to tear down the scanner portion of the device and ended up turning it into a multi-touch display.

The scanner relies on a long PCB with a line CCD sensor. This sensor is read in a similar way that information is passed along a shift register. Tell it to take a reading, and then start a clock signal to pulse out each analog value from the pixels of the sensor. In order to scan color images it uses multicolored LEDs to take different readings under different illumination.

[Sprite_TM] takes advantage of this functionality to turn it into a multitouch sensor. The sensor board itself is mounted below an LCD display along with a shield with a slit in it to help filter out ambient light. Above the screen a series of LEDs shine down on the sensor. When you break the beams with your finger it casts a series of shadows which are picked up by the sensor and processed in software. Watch the clip after the break to see it for yourself. It has no problem detecting and tracking multiple contact points.

Continue reading “Converting A Scanner Into A Touchscreen”

EeePC Touchscreen Retrofit

Adding touch screen capabilities to your computer is really not very expensive, but it’s a huge amount of work to get everything looking the way that it should. [Deadbird] wrote up a step-by-step guide that will help you install touch screen hardware and get your netbook put back together just like new.

The hardware comes in two parts. There’s the transparent film that covers the screen and the driver board that reads the inputs. The film itself has an adhesive layer on the back that sticks to the LCD panel. But to install it you first must remove the panel from the bezel. You’re also going to need a place to house the driver board. [Deadbird] somehow found enough room inside the case for the controller, but he had to remove the keyboard and motherboard to set it in place. This translates to a complete disassembly of your eeePC. But if you’re used to touch-sensitive devices, and have ever found yourself touching an LCD monitor and wondering why the computer is not following the link, this may be worth it to you. You can see the final product in a clip after the break.

Continue reading “EeePC Touchscreen Retrofit”

3D Modeling Out Of Thin Air

kinect_3d_modeling

It seems that with each passing day, the Kinect hacks that we see become exponentially more impressive. Take for instance this little number that was sent to us today.

[sonsofsol] has combined several open source software packages and a little electronics know-how to create one of the more useful Kinect hacks we have seen lately. His project enables him to manipulate 3D models in GEM simply by moving his hands about in front of his Kinect sensor. Using OpenNI and Ubuntu, all of his actions are tracked by the computer and translated into actions within the GEM 3D engine.

To make things easier on himself, he also constructed a pair of electronic gloves that interface with the system. Using an Arduino, the gloves send different complex commands to the 3D modeling software, just by touching different pairs of fingers together.

You really need to take a look at the video embedded below to get a feel for how complex [sonsofsol’s] “simple” mesh modeler really is.

Looking for more Kinect fun? Check out these previously featured stories.

[Thanks, Jared]

Continue reading “3D Modeling Out Of Thin Air”

Kinect Home Theater Control

kinect_home_theater

[Harishankar] has posted a video on his blog demonstrating the ability to control devices using the Microsoft Kinect sensor via IR. While controlling devices with Kinect is nothing new, he is doing something a little different than you have seen before. The Kinect directly interfaces with his Mac Mini and tracks his movements via OpenNI. These movements are then compared to a list of predefined gestures, which have been mapped to specific IR functions for controlling his home theater.

Once the gestures have been acknowledged, they are then relayed from the Mac via a USB-UIRT to various home theater components. While there are not a lot of details fleshed out in the blog post, [Harishankar] says he will gladly forward his code to you if you request it via email.

Thanks to [Peter] for the tip.

Laptop LCD Reused In Beagleboard Project

This daughterboard lets [Matt Evans] drive a laptop LCD using a Beagleboard. Apparently the Beagleboard gained a VGA header when it moved to revision C but [Matt’s] working with revision B4 which is why he had to do all of that ninja soldering with the blue wires. The driver board itself is a thing of beauty, hosting a DS90C363 LVDS serialiser as well as some buffer chips that handle level conversion for it. He’s also included an ATmega48 so that he has some options for future improvements.

The LCD is mounted in a custom acrylic case, with Beagleboard and driver board taped to the back of it. There’s RS232 and a USB hub which opens up the possibility of using a WiFi dongle for communications. So far he doesn’t have much functionality other than displaying images on the screen but there is some talk about using a touchpad for control. We’d love to see a touchscreen overlay, transforming the build into a proper ARM-based tablet.

Projector Tricks Make Use Of Kinect 3D Mapping

[Don’t stop the clock] is doing some work with a projector, a camera, and the Kinect. What he’s accomplished is quite impressive, combining the three to manipulate light with your body. The image above is a safer rendition of the Hadouken from the Street Fighter video games, throwing light across the room instead of fire. This comes at the end of the video after the break, but first he’ll show off the core features of the system. You can hold up your hand and wave it to turn it into a light source. In other words, the projector will shine light on your hand, moving it, and manipulating the intensity based on hand location in 3D space. Since the Kinect is sending fairly precise data back to the computer the projected image is trimmed to match your hand and arm without overflowing onto the rest of the room until you touch your hand to a surface you want illuminated or throw the light source with a flick or the wrist. It may seem trivial at first glance, but we find the alignment of the projector and the speed at which the image updates to be quite impressive.

Continue reading “Projector Tricks Make Use Of Kinect 3D Mapping”