We’ve seen a few near-future sci-fi films recently where computers respond not just to touchscreen gestures but also to broad commands, like swiping a phone to throw its display onto a large flat panel display. It’s a nice metaphor, and if we’re going to see something like it soon, perhaps this wrist-mounted pointing device will be one way to get there.
The video below shows the finished product in action, with the cursor controlled by arm movements. Finger gestures that are very much like handling a real mouse’s buttons are interpreted as clicks. The wearable has a Nano, an MPU6050 IMU, and a nRF24L01 transceiver, all powered by some coin cells and tucked nicely into a 3D-printed case. To be honest, as cool as [Ronan Gaillard]’s wrist mouse is, the real story here is the reverse engineering he and his classmate did to pull this one off.
The road to the finished product was very interesting and more detail is shared in their final presentation (in French and heavy with memes). Our French is sufficient only to decipher “Le dongle Logitech,” but there are enough packet diagrams supporting into get the gist. They sniffed the packets going between a wireless keyboard and its dongle and figured out how to imitate mouse movements using an NRF24 module. Translating wrist and finger movements to cursor position via the 6-axis IMU involved some fairly fancy math, but it all seems to have worked in the end, and it makes for a very impressive project.
Is sniffing wireless packets in your future? Perhaps this guide to Wireshark and the nRF24L01 will prove useful.
Continue reading “Wireless Protocol Reverse Engineered to Create Wrist Wearable Mouse”
With pervasive smartphones and tablets, the touch interface is assumed for small LCD screens, and we’ve likely all poked and pinched at some screen, only to find it immune to our gestures. Manufacturers have noticed this and begun adding touch interfaces to instruments like digital oscilloscopes, but touch interfaces tend to be an upgrade feature. But thanks to this hybrid oscilloscope touchscreen interface, even the low-end scopes can get in on the action.
It only makes sense that [Matt Heinz] started with one of the most hackable scopes for this build, which was his Master’s thesis project. Using an Android tablet as an auxiliary interface, [Matt] is able to control most of the main functions of the scope remotely. Pinching and expanding gestures are interpreted as horizontal and vertical scaling, while dragging the displayed waveform changes its position and controls triggering. While it’s not a true touchscreen scope, the code is all open source, so can a true aftermarket Rigol touchscreen be far away?
Rigol hacks abound here — you can talk to them in Linux, increase the bandwidth, or just get a look at their guts.
Continue reading “Hybrid Interface Brings Touchscreen to Rigol Scope”
[B. Aswinth Raj] wanted to control a VLC player with hand gestures. He turned to two common ultrasonic sensors and Python to do the job. There is also, of course, an Arduino. You can see a video of the results, below.
The Arduino code reads the distance from both sensors — one for the left hand and the other for the right. This allows the device to react to single hand gestures that get closer or further away from one sensor as well as gestures involving both hands. For example, raising your left hand and moving it closer or further away will adjust the volume. The right hand controls rewind and fast forward. Raising both hands will start or stop playback.
Continue reading “Listening for Hand Gestures”
Movies love to show technology they can’t really build yet. Even in 2001: A Space Oddessy (released in 1968), for example, the computer screens were actually projected film. The tablet they used to watch the news looks like something you could pick up at Best Buy this afternoon. [CircuitDigest] saw Iron Man and that inspired him to see if he could control his PC through gestures as they do on that film and so many others (including Minority Report). Although he calls it “virtual reality,” we think of VR as being visually immersed and this is really just the glove, but it is still cool.
The project uses an Arduino on the glove and Processing on the PC. The PC has a webcam which tracks the hand motion and the glove has two Hall effect sensors to simulate mouse clicks. Bluetooth links the glove and the PC. You can see a video of the thing in action, below.
Continue reading “A Minority Report Arduino-Based Hand Controller”
One of the companion technologies in the developing field of augmented reality is gesture tracking. It’s one thing to put someone in a virtual or augmented world, but without a natural way to interact inside of it the user experience is likely to be limited. Of course, gestures can be used to control things in the real world as well, and to that end [Sarah]’s latest project uses this interesting human interface device to control a drone.
The project uses a Leap Motion sensor to detect and gather the gesture data, and feeds all of that information into LabVIEW. A Parrot AR Drone was chosen for this project because of a robust API that works well with this particular software suite. It seems as though a lot of the grunt work of recognizing gestures and sending commands to the drone are taken care of behind-the-scenes in software, so if you’re looking to do this on your own there’s likely to be quite a bit more work involved. That being said, it’s no small feat to get this to work in the first place and the video below is worth a view.
To some, gestures might seem like a novelty technology with no real applications, but they do have real-world uses for people with disabilities or others with unusual workflow that require a hands-free approach. So far we’ve seen hand gesture technologies that drive cars, help people get around in the physical world, and even play tetris.
Continue reading “Drone Takes Off With a Flick of the Wrist”
When [krich] switched keyboards he lost his volume control. So he decided to hack one together out of an Arduino, an old floppy disc case, and a Hover Labs Hover board (not the Back to the Future kind). You can see the result in the videos below.
Continue reading “Making a Gesture”
Who doesn’t like the user interface in the movie Minority Report where [Tom Cruise] manipulates a giant computer screen by just waving his hands in front of it? [AdhamN] wanted to unlock his door with hand gestures. While it isn’t as seamless as [Tom’s] Hollywood interface, it manages to do the job. You just have to hold on to your smartphone while you gesture.
The project uses an Arduino and a servo motor to move a bolt back and forth. The gesture part requires a 1sheeld board. This is a board that interfaces to a phone and allows you to use its capabilities (in this case, the accelerometer) from your Arduino program.
The rest should be obvious. The 1sheeld reads the accelerometer data and when it sees the right gesture, it operates the servo. It would be interesting to do this with a smart watch, which would perhaps look a little less obvious.
We covered the 1sheeld board awhile back. Of course, you could also use NFC or some other sensor technology to trigger the mechanism. You can find a video that describes the 1sheeld below.
Continue reading “Hand Waving Unlocks Door”