Let’s face it, most of the time we’re hacking for no other reason than sheer enjoyment. So we love to see hacks come about that can really make a difference in people’s lives. This time around it’s a video game designed to exercise your eyes. [James Blaha] has an eye condition called Strabismus which is commonly known as crossed-eye. The issue is that the muscles for each eye don’t coordinate with each other in the way they need to in order to produce three-dimensional vision.
Recent research (linked in the reference section of [James’] post) suggests that special exercises may be able to train the eyes to work correctly. He’s been working on developing a video game to promote this type of training. As you can see above, the user (patient?) wears an Oculus Rift headset which makes it possible to show each eye slightly different images, while using a Leap Motion controller for VR interaction. If designed correctly, and paired with the addictive qualities of games, this my be just what the doctor ordered. You know what they say, practice makes perfect!
Continue reading “Video Gaming to Fix Eye Ailments”
There’s been some good .STL manipulation tips in this week.
The first one is called stl_tools, and it’s a Python library to convert images or text to 3D-printable STL files. The examples shown are quite impressive, and it even does a top notch job of taking a 2D company logo into 3D! We can see this being quite handy if you need some quick 3D text, and either don’t use CAD, or really just need a one click solution. Now if only .STLs were easier to edit afterwards…
Continue reading “STL Fun: Converting Images To STL Geometry”
Robots used in laparoscopic surgery are fairly commonplace, but controlling them is far from simple. The usual setup is something akin to a Waldo-style manipulator, allowing a surgeon to cut, cauterise, and stitch from across a room. There is another way to go about this thanks to some new hardware, as [Sriranjan] shows us with his Leap-controlled surgery bot.
[Sriranjan] isn’t using a real laparoscopic surgery robot for his experiments. Instead, he’s using the Le-Sur simulator that puts two virtual robot arms in front of a surgeon in training. Each of these robotic arms have seven degrees of freedom, and by using two Leap controllers (one each in a VM), [Sriranjan] was able to control both of them using his hands.
We’ve seen a lot of creative applications for the Leap sensor, like controlling quadcopters, controlling hexapod robots, and controlling more quadcopters, but this is the first time we’ve seen the Leap do something no other controller can – emulating the delicate touch of a surgeon’s hand
Continue reading “Finally, a practical use for the Leap”
The gang at Bitcraze is at it again, this time developing Leap Motion control for their Crazyflie quadcopter, as well as releasing a Kinect-driven autopilot proof of concept. If you haven’t seen the Crazyflie before, you may not realize how compact it is: 90mm motor to motor and only 19 grams.
As far as we can tell, the Crazyflie still needs a PC to control it, so the Leap and Kinect are natural followups. Hand control with the Leap Motion is what you’d expect: just imagine your open palm controlling it like a marionette, with the height of your hand dictating thrust. The Kinect setup looks the most promising. The guys strapped a red ball to the Crazyflie that provides a trackable object against a white backdrop. The Kinect then monitors the quadcopter while a user steers via mouse clicks. Separate PID controllers correct the roll, pitch and thrust to reposition the Crazyflie from its current coordinates to a new setpoint chosen by a click or a drag. Videos of both Leap and Kinect piloting are below.
Tight on cash but still want to take to the skies? We have two rubber-band-powered devices from earlier this week: the Ornithopter and the hilariously brilliant GoPro Slingshot.
Continue reading “Crazyflie control with Leap and Kinect”
Moving your hand makes this hexapod dance like a stringless marionette. Okay, so there’s obviously one string which is actually a wire but you know what we mean. The device on the floor is a Leap Motion sensor which is monitoring [Queron Williams’] hand gestures. This is done using a Processing library which leverages the Leap Motion API.
Right now the hand signals only affect pitch, roll, and yaw of the hexapod’s body. But [Queron] does plan to add support for monitoring both hands to add more control. We look at the demo after the break and think this is getting pretty close to the manipulations shown by [Tom Cruise] in Minority Report. Add Google Glass for a Heads Up Display and you could have auxiliary controls rendered on the periphery.
While you’re looking at [Queron’s] project post click on his ‘hexapod’ tag to catch a glimpse the build process for the robot.
Continue reading “Leap motion controls hexapod with hand signals”
The Leap Motion is a very cool device, but so far we haven’t seen many applications of interacting with physical devices. [Xavier] wanted to control a cute servo animated desk lamp with his hands, and with the help of a Leap and an Arduino he was able to do just that.
The Leap Motion API has a handy feature that will output all its data over a websocket. It’s a very easy way to transfer hand positions with a minimum amount of overhead, and with just a little bit of Node.js, it’s only two lines of code to connect the Leap to a websocket server.
Video of the build below.
Continue reading “Animating a lamp with the Leap Motion”
Touch screens are nice — we still can’t live without a keyboard but they suffice when on the go. But it is becoming obvious that the end goal with user interface techniques is to completely remove the need to touch a piece of hardware in order to interact with it. One avenue for this goal is the use of voice commands via software like Siri, but another is the use of 3D processing hardware like Kinect or Leap Motion. This project uses the latter to control the image shown on the 3D display.
Continue reading “3D display controlled with the Leap Motion”