[Bastian] sent in a coffee table he built. This isn’t a place to set your drinks and copies of Make, though: it’s a multitouch table with a 3D display. Since no description can do this table justice, take a look at the video.
The build was inspired by the subject of this Hackaday post where [programming4fun] was able to build a ‘holographic display’ using a regular 2D projector and a Kinect. Both builds work on the principle of redrawing the 3D space in relation to the user’s head – as [Bastian] moves his head around the coffee table, the Kinect tracks his location and moves the 3 dimensional grid of boxes in the opposite direction. It’s extremely clever, and looks to be a promising user interface.
In addition to a Kinect, the coffee table uses a Microsoft Surface-like display; four infrared lasers are placed at the corner and detected with a camera next to the projector in the base.
After the break you can see the demo video and a gallery of the images [Bastion] put up on the NUI group forum.
Continue reading “Multitouch table uses a Kinect for a 3D display”
Here’s a bulky old CRT monitor used as a touch-screen without any alterations. It doesn’t use an overlay, but instead detects position using phototransistors in the fingertips of a glove.
Most LCD-based touch screens use some type overlay, like these resistive sensors. But cathode-ray-tube monitors function in a fundamentally different way from LCD screens, using an electron gun and ring of magnets to direct a beam across the screen. The inside of the screen is coated with phosphors which glow when excited by electrons. This project harness that property, using a photo transistor in both the pointer and middle finger of the glove. An FPGA drives the monitor and reads from the sensors. It can extrapolate the position of the phototransistors on the display based on the passing electron beam, and use that as cursor data.
Check out the video after the break to see this in action. It’s fairy accurate, but we’re sure the system can be tightened up a bit from this first prototype. There developers also mention that the system has a bit of trouble with darker shades.
Continue reading “Glove-based touch screen from a CRT monitor”
To us it makes a lot of sense to hold the tablet in one hand and type with the other. That’s exactly how [Adam Kumpf] has implemented this one-handed typing interface which was originally conceived by [Doug Engelbart].
As you can see, there’s a large contextual area for each finger on your right hand. Letters and navigational keystrokes are input through this interface based on single touches, or combinations up to and including all five digits. This offers up 32 possible combinations (including all on and all off) which is enough to cover the modern English alphabet.
[Adam’s] demo page works for most tablets so give it a whirl. Yes, it works with iDevices too which is a surprise as we would have thought this was using Flash. If you’re not near a touch-sensitive device you can get the gist of the operation from the demo video embedded after the break.
Now, who’s going to be the first to make this into a replacement keyboard on iOS 5?
Continue reading “Single hand keyboard for tablets”
This mirror has a large monitor behind it which can be operated using hand gestures. It’s the result of a team effort from [Daniel Burnham], [Anuj Patel], and [Sam Bell] to build a web-enabled mirror for their ECE 4180 class at the Georgia Institute of Technology.
So far they’ve implemented four widget for the system. You can see the icons which activate each in the column to the right of the mirror. From top to bottom they are Calendar, News, Traffic, and Weather. The video after the break shows the gestures used to control the display. First select the widget by holding your hand over the appropriate icon. Next, bring that widget to the main display area by swiping from right to left along the top of the mirror.
Hardware details are shared more freely in their presentation slides (PDF). A sonar distance sensor activated the device when a user is close enough to the screen. Seven IR reflectance sensors detect a hand placed in front of them. We like this input method, as it keep the ‘display’ area finger-print free. But we wonder if the IR sensors could be placed behind the glass instead of beside it?
Continue reading “Cloud Mirror adds Internet to your morning ritual”
If you’re tired of playing flash games with a mouse, perhaps you’ll draw inspiration from this project. Arthur built a multitouch interface that uses objects as part of the control scheme. In the image above you can see that the game board for a tower defense game is shown on the display. There is a frustum-shaped game piece resting on the surface. Just place that piece where you want to build your next tower, and then select the tower type from the list.
The controller itself is pretty straight-forward. The surface is a piece of acrylic topped with some light diffusing material. A projector shines through another acrylic window on the side of the unit, reflecting on a mirror positioned at a 45 degree angle. As for the multitouch detection, the hardware uses a series of UV LEDs along with a modified PS3 eye camera. [Arthur] chose the reacTIVision software package to process the input from the camera. Check out a couple of videos after the break to see the hardware, and some game play.
Continue reading “Multitouch tower defense uses physical towers”
[David Revoy] recently picked up a brand new Cintiq 21UX, and while he liked the drawing pad overall, he was less than impressed with the tablet’s buttons. He says that most 2D linux apps require a good bit of keyboard interaction, and the built-in buttons just were not cutting it.
After seeing a fellow artist use a joypad to augment his tablet, [David] thought that he might be able to do something similar, but he wanted to add a lot more buttons. He dug out an old Logitech game pad that was collecting dust, and disassembled it, rearranging some buttons in the process. Once he was happy with the layout, he built a cardboard enclosure for the PCB and hooked it up to the Wacom via USB.
He spent a few minutes mapping buttons to key presses using Qjoypad, and was up and running with an additional 14 buttons in short order. He says that the extra buttons make his job a ton easier, and add a little bit of comfort to his long drawing sessions. We like the fact that it is a non-permanent fixture, and that he was able to repurpose an old game pad in the process.
Check out the video below for a quick demonstration of his drawing pad hack.
[via Adafruit blog]
Continue reading “Adding extra buttons to a Cintiq drawing pad”
Fanboys may be in shock from seeing duct tape applied to the screen of an iPad, but we can assure you it’s in the name of science. [Michael Knuepfel] is working on his thesis for the ITP graduate program at the Tisch School for the Arts. He managed to augment the usability of touchscreen devices by adding hardware to them.
What he’s come up with are devices for both input and output. The output devices generally rely on light and color of light displayed on the screen itself which is picked up by a light sensor. The input devices use conductive material to complete a path between your hand and your screen. This lets the capacitive sensing screen detect the presence of your hand, through the conductor. Some of his example devices include gaming controller overlays, encoder rings, and multiple stylus designs.
After the break we’ve embedded [Michael’s] teaser trailer which jumps through several demonstrations. It’s plenty to get your mind rolling, but if you want to know more you must watch his thesis presentation. It’s available as an MP4 download on this page. Just search for his name, [Michael Knuepfel] for the proper link.
Continue reading “Extending the usability of touchscreens”