Wander through a well-funded museum these days and you’re likely to find interactive exhibits scattered around, such as this sleek 50″ projection-based multitouch table. The company responsible for this beauty, Ideum, has discontinued the MT-50 model in favor of an LCD version, and has released the plans for the old model as part of the Open Exhibits initiative. This is a good thing for… well, everyone!
The frame consists of aluminum struts that crisscross through an all-steel body, which sits on casters for mobility. The computer specs seem comparable to a modern gaming rig, and rely on IEEE1394 inputs for the cameras. The costs start to pile up with the multiple row of high-intensity infrared LED strips, which can run $200 per roll. The glass is a custom made, 10mm thick sheet with projection film on one side and is micro-etched to reduce reflections and increase the viewing angle to nearly 180 degrees. The projector is an InFocus IN-1503, which has an impressively short projection throw ratio, and a final resolution of 1280×720.
The estimated price tag mentioned in the comments is pretty steep: $12k-16k. Let us know with your own comment what alternative parts might cut the cost, and watch the video overview of the table below, plus a video demonstration of its durability. For another DIY museum build, check out Bill Porter’s “Reaction Time Challenge.”
Continue reading “50″ multitouch table is expensive, indestructable”
For over a hundred years, good typists didn’t ‘hunt and peck’ but instead relied on keeping their fingers on the home row. This technique relies on physical buttons, but with on-screen keyboards used on tablets and other touch screen devices touch typists have a very hard time. [Zach] is working on a new project to bring a chorded keyboard to these devices called ASETNIOP.
Instead of training a typist where to place their finger – the technique used in most other keyboard replacements, ASETNIOP trains the typist which fingers to press. For example, typing ‘H’ requires the typist to press the index and middle fingers of their right hand against the touchscreen. In addition to touchscreens, ASETNIOP can be used with projection systems, Nintendo Power Glove replicas, and extremely large touchpads that include repurposed nooks and Kindles.
If you’d like to try out ASENTNIOP, there’s a tutorial that allows you to try it out on a physical keyboard as well as one for the iPad. It’s a little weird to try out but surely no more difficult to learn than a Dvorak keyboard.
[Sprite_tm], a name many of you will recognize from these pages, has wasted no time in replicating the latest cool thing in a much simpler fashion. En Garde is a touch sensor that can detect up to 32 different points of contact on… whatever you use as the surface. He couldn’t sit idly by and let the Disney funded one from yesterday keep the spot light. As you can see in the video, it works pretty well. If he didn’t tell you that his can only detect up to 32 points as opposed to the 200 of the other, you probably wouldn’t even notice the difference. Of course, [Sprite_tm] also shares how you could easily beef his up to be even more precise. You can also download his source code an schematics from his site and give it a try yourself.
Continue reading “Replicating the fancy touch sensor that uses anything”
This year at the CHI conference in Austin, [Munehiko Sato], [Ivan Poupyrev], and [Chris Harrison] out of the Disney research lab in Pittsburgh demonstrated their way to make touch sensors out of anything. Not only to they suggest using the surface of your skin to control cell phones and MP3 players, they’re also able to recognize touch gestures, like poking or grasping an object. That sounds a little heady, so check out the video of the Touché tech in action.
Like the capacitive touch sensors in our phones and tablets, Touché measures the rise and fall of a capacitor’s charge over time. Unlike other touch sensors, Touché scans the capacitor at different rates, allowing for a ‘capacitive profile’ that is used to recognized touch gestures.
The applications for this tech are nearly innumerable; the team demonstrated scolding someone for eating cereal with chopsticks (yeah, we know…), an on-body music player interface, and gestures for an office doorknob that notifies passersby if you’ve stepped out for a minute or are gone for the day.
It’s a very interesting build, and we give it two weeks until someone replicates this build. We’ll be sure to post it then.
Continue reading “Turning anything into a touch sensor”
[Bastian] sent in a coffee table he built. This isn’t a place to set your drinks and copies of Make, though: it’s a multitouch table with a 3D display. Since no description can do this table justice, take a look at the video.
The build was inspired by the subject of this Hackaday post where [programming4fun] was able to build a ‘holographic display’ using a regular 2D projector and a Kinect. Both builds work on the principle of redrawing the 3D space in relation to the user’s head – as [Bastian] moves his head around the coffee table, the Kinect tracks his location and moves the 3 dimensional grid of boxes in the opposite direction. It’s extremely clever, and looks to be a promising user interface.
In addition to a Kinect, the coffee table uses a Microsoft Surface-like display; four infrared lasers are placed at the corner and detected with a camera next to the projector in the base.
After the break you can see the demo video and a gallery of the images [Bastion] put up on the NUI group forum.
Continue reading “Multitouch table uses a Kinect for a 3D display”
If you’re going to build a giant touch screen, why not use an OS that is designed for touch interfaces, like Android? [Colin] had the same idea, so he connected his phone to a projector and a Kinect.
Video is carried from [Colin]’s Galaxy Nexus to the projector via an MHL connection. Getting the Kinect to work was a little more challenging, though. The Kinect is connected to a PC running Simple Kinect Touch. The PC converts the data from the Kinect into TUIO commands that are received using TUIO for Android.
In order for the TUIO commands to be recognized as user input, [Colin] had to compile his own version of Android. It was a lot of work, but using an OS designed for touch interface seems much better than all the other touch screen hacks that start from the ground up.
You can check out [Colin]’s demo after the break. Sadly, there are no Angry Birds.
Continue reading “Control Android with a projector and Kinect”
Even though giant multouch display tables have been around for a few years now we have yet to see them being used in the wild. While the barrier to entry for a Microsoft Surface is very high, one of the biggest problems in implementing a touch table is one of interaction; how exactly should the display interpret multiple commands from multiple users? [Stephan], [Christian], and [Patrick] came up with an interesting solution to sorting out who is touching where by having a computer look at shoes.
The system uses a Kinect mounted on the edge of a table to extract users from the depth images. From there, interaction on the display can be pinned to a specific user based on hand and arm orientation. As an added bonus the computer can also recognize users from their shoes. If a user is wearing a pair of shoes the computer recognizes, they’ll just walk up to the table and the software will recognize them.
Continue reading “Nice shoes, wanna recognize some input?”