Sunday we saw robots playing pool and an augmented reality pool game. Today we’ll complete the pool trifecta: virtual pool using a real cue stick and ball in another vintage video from Hack a Day’s secret underground vault. The video is noteworthy for a couple of reasons:
First is the year it was made: 1990. There’s been much buzz lately over real-world gaming interfaces like the Nintendo Wii motion controller or Microsoft’s Project Natal. Here we’re seeing a much simpler but very effective physical interface nearly twenty years prior.
Second: the middle section of the video reveals the trick behind it all, and it turns out to be surprisingly simple. No complex sensors or computer vision algorithms; the ball’s speed and direction are calculated by an 8-bit processor and a clever arrangement of four infrared emitter/detector pairs.
The visuals may be dated, but the interface itself is ingenious and impressive even today, and the approach is easily within reach of the casual garage tinkerer. What could you make of this? Is it just a matter of time before we see a reader’s Mini-Golf Hero III game here?
[Tom Gerhardt] has made this very interesting mud interface for a computer. Follow the link to see a video of it in action. It appears as though he’s using a laser grid of some kind to establish elevation. We might be way off on that though, there aren’t any details on the construction. He does mention that it is an open source hardware and software project, so maybe the details are available on request. In the video you can see it running as a projection surface where people are interacting with items directly on the mud. You can also see it being used as an external input device. People play Tetris using it in that example.
UPDATE: [Moon] reports from the ITP show that the tub has a 16×12 grid of generic pressurs sensors on the bottom. These feed into a MacBook Pro which is projecting on the surface. Despite the sparse grid, [Tom] says he gets good resolution by interpolating between sensors; it can detect a resting hand pivoting on the surface.
Microchip’s MCP6S21/2/6/8 are programmable gain amplifiers that multiply an input voltage by a factor of 1, 2, 4, 5, 8, 10, 16, and 32. The MCP6S22/6/8 also have selectable input channels for working with different signal sources. The multiplication factor and input channel are configured through an SPI interface. This chip is useful for multiplying a small input signal, and selecting among several analog input sources. We demonstrate the six channel MCP6S26 below.
Continue reading “Parts: Programmable gain amplifier (MCP6S26)”
We use the Bus Pirate to interface a new chip without writing code or designing a PCB. Based on your feedback, and our experience using the original Bus Pirate to demonstrate various parts, we updated the design with new features and cheaper components.
There’s also a firmware update for both Bus Pirate hardware versions, with bug fixes, and a PC AT keyboard decoder. Check out the new Hack a Day Bus Pirate page, and browse the Bus Pirate source code in our Google code SVN repository.
We cover the design updates and interface a digital to analog converter below.
Continue reading “How-to: Bus Pirate v1, improved universal serial interface”
A few weeks ago we wrote about our Bus Pirate universal serial interface tool. We used the recent holiday to add some new features, like a JTAG programmer, macros, frequency measurement, and more. A major code reorganization makes everything easier to read and update.
Check out the a demonstration of the new features below. We’re compiling a roadmap and wish list, so share your ideas in the comments. You can also see how we used the Bus Pirate to read a smart card and test-drive an I2C crystal oscillator.
Continue reading “Bus Pirate firmware update (v.0c), JTAG and more”
[floe] wrote in to tell us about his multitouch based thesis work. While many projects have focused on the hardware side of multitouch, TISCH is designed to promote the software side. TISCH is a multiplatform library that features hardware abstraction and gesture recognition. This takes a lot of weight off of widget developers since they can specify known library gestures instead of writing the exact motions from scratch. Using TISCH also means a standard set of gestures across multiple widgets, so the learning curve will be much easier when a user tries out a new app. If you’re researching multitouch, check out this project and help improve the codebase.
Our fascination with multitouch is fairly well known, but it expands even further to cover all sorts of man machine interaction. Embedded above is a tech demo of g-speak, a spatial operating environment. The user combines gestures and spatial location to interact with on screen objects. If it seems familiar, it’s because one of the company’s founders advised on Minority Report. We doubt all this hand waving is going to catch on very quickly though. Our bet is on someone developing a multitouch Cintiq style device for people to use as a secondary monitor. It would bridge the gap between between our standard 2D interactions and gestures without making a full leap to 3D metaphors.
[via Create Digital Motion]