A Minority Report Arduino-Based Hand Controller

Movies love to show technology they can’t really build yet. Even in 2001: A Space Oddessy (released in 1968), for example, the computer screens were actually projected film.  The tablet they used to watch the news looks like something you could pick up at Best Buy this afternoon. [CircuitDigest] saw Iron Man and that inspired him to see if he could control his PC through gestures as they do on that film and so many others (including Minority Report). Although he calls it “virtual reality,” we think of VR as being visually immersed and this is really just the glove, but it is still cool.

The project uses an Arduino on the glove and Processing on the PC. The PC has a webcam which tracks the hand motion and the glove has two Hall effect sensors to simulate mouse clicks. Bluetooth links the glove and the PC. You can see a video of the thing in action, below.

Continue reading “A Minority Report Arduino-Based Hand Controller”

Drone Takes Off With a Flick of the Wrist

One of the companion technologies in the developing field of augmented reality is gesture tracking. It’s one thing to put someone in a virtual or augmented world, but without a natural way to interact inside of it the user experience is likely to be limited. Of course, gestures can be used to control things in the real world as well, and to that end [Sarah]’s latest project uses this interesting human interface device to control a drone.

The project uses a Leap Motion sensor to detect and gather the gesture data, and feeds all of that information into LabVIEW. A Parrot AR Drone was chosen for this project because of a robust API that works well with this particular software suite. It seems as though a lot of the grunt work of recognizing gestures and sending commands to the drone are taken care of behind-the-scenes in software, so if you’re looking to do this on your own there’s likely to be quite a bit more work involved. That being said, it’s no small feat to get this to work in the first place and the video below is worth a view.

To some, gestures might seem like a novelty technology with no real applications, but they do have real-world uses for people with disabilities or others with unusual workflow that require a hands-free approach. So far we’ve seen hand gesture technologies that drive cars, help people get around in the physical world, and even play tetris.

Continue reading “Drone Takes Off With a Flick of the Wrist”

Hand Waving Unlocks Door

Who doesn’t like the user interface in the movie Minority Report where [Tom Cruise] manipulates a giant computer screen by just waving his hands in front of it? [AdhamN] wanted to unlock his door with hand gestures. While it isn’t as seamless as [Tom’s] Hollywood interface, it manages to do the job. You just have to hold on to your smartphone while you gesture.

The project uses an Arduino and a servo motor to move a bolt back and forth. The gesture part requires a 1sheeld board. This is a board that interfaces to a phone and allows you to use its capabilities (in this case, the accelerometer) from your Arduino program.

The rest should be obvious. The 1sheeld reads the accelerometer data and when it sees the right gesture, it operates the servo. It would be interesting to do this with a smart watch, which would perhaps look a little less obvious.

We covered the 1sheeld board awhile back. Of course, you could also use NFC or some other sensor technology to trigger the mechanism. You can find a video that describes the 1sheeld below.

Continue reading “Hand Waving Unlocks Door”

Hand Gestures Play Tetris

There are reports of a Tetris movie with a sizable budget, and with it come a plentiful amount of questions about how that would work. Who would the characters be? What kind of lines would there be to clear? Whatever the answers, we can all still play the classic game in the meantime. And, thanks to some of the engineering students at Cornell, we could play it without using a controller.

This hack comes from [Bruce Land]’s FPGA design course. The group’s game uses a video camera which outputs a standard NTSC signal and also does some filtering to detect the user. From there, the user can move their hands to different regions of the screen, which controls the movement of the Tetris pieces. This information is sent across GPIO to another FPGA which uses that to then play the game.

This game is done entirely in hardware, making it rather unique. All game dynamics including block generation, movement, and boundary conditions are set in hardware and all of the skin recognition is done in hardware as well. Be sure to check out the video of the students playing the game, and if you’re really into hand gesture-driven fun, you aren’t just limited to Tetris, you can also drive a car.

Continue reading “Hand Gestures Play Tetris”

Hand Gestures Drive Car

There are a number of ways to control an automobile without using the pedals, and sometimes even without using the steering wheel. Most commonly these alternative control mechanisms are installed in vehicles whose owners are disabled in some way, but [Anurag] has taken this idea of alternative control one step further. He has built a car that can be driven by hand gestures alone.

On a remote controlled car, a Raspberry Pi 2 was installed that handles processing and communication. A wireless network is created on the Pi, and a laptop connects to the Pi over the network. The web camera on the laptop regularly captures frames at 15 fps to check for the driver’s hand gestures. The image is converted to gray scale, thresholded, contours are obtained, and the centroid and farthest points are obtained.

After some calculations are done, a movement decision is taken. The decision is passed to the Pi, which in turn, passed that to the internal chip of the car. All of the code is available on the project’s github page. [Anurag] hopes that this can be scaled up to full sized cars in the future. We’ve seen gesture-based remote controls before that rely on Sonar sensors, so it’s interesting to see one that relies strictly on image processing.

Continue reading “Hand Gestures Drive Car”

Impedance Tomography is the new X-Ray Machine

Seeing what’s going on inside a human body is pretty difficult. Unless you’re Superman and you have X-ray vision, you’ll need a large, expensive piece of medical equipment. And even then, X-rays are harmful part of the electromagnetic spectrum. Rather than using a large machine or questionable Kryptonian ionizing radiation vision, there’s another option now: electrical impedance tomography.

[Chris Harrison] and the rest of a research team at Carnegie Mellon University have come up with a way to use electrical excitation to view internal impedance cross-sections of an arm. While this doesn’t have the resolution of an X-ray or CT, there’s still a large amount of information that can be gathered from using this method. Different structures in the body, like bone, will have a different impedance than muscle or other tissues. Even flexed muscle changes its impedance from its resting state, and the team have used their sensor as proof-of-concept for hand gesture recognition.

This device is small, low power, and low-cost, and we could easily see it being the “next thing” in smart watch features. Gesture recognition at this level would open up a whole world of possibilities, especially if you don’t have to rely on any non-wearable hardware like ultrasound or LIDAR.