We’re pretty sure the Hackaday demographic is a a person who sees a giant tower crane lifting beams and girders above a skyline and says, “that would be fun, at least until I have to go to the bathroom.” Realizing the people who own these cranes probably won’t let any regular joe off the street into the cabin, [Thomas] and [screen Name] (see, this is why we have brackets, kids) built their own miniature version with an Oculus Rift.
Instead of a crane that is hundreds of feet tall, the guys are using a much smaller version, just over a meter tall, that is remotely controlled through a computer via a serial connection. Just below the small plastic cab is a board with two wide-angle webcams. The video from these cameras are sent to the Oculus so the operator can see the boom swinging around, and the winch unwinding to pick up small objects.
The guys have also added a little bit of OpenCV to add color based object detection. This is somewhat useful, but there’s also an approximation of the distance to an object, something that would be very useful if you don’t have a three-inch tall spotter on the ground.
Continue reading “The Crane Game, Oculus Style”
Taking on an autonomous vehicle challenge, [Randy] put together this drone which can locate and pop balloons. It’s been assembled for this year’s Sparkfun Autonomous Vehicle Competition, which will challenge entrants to locate and pop
99 luftbaloons red balloons without human intervention.
The main controller for this robot is the Pixhawk, which runs a modified version of the ArduCopter firmware. These modifications enable the Pixhawk to receive commands from an Odroid U3 computer module. The Odroid uses a webcam to take images, and then processes them using OpenCV. It tries to locate large red objects and fly towards them.
The vision processing and control code on the Odroid was developed using MAVProxy and Drone API. This allows for all the custom code to be developed using Python.
The Sparkfun AVC takes place tomorrow — June 21st in Boulder, Colorado. You can still register to spectate for free. We’re hoping [Randy]’s drone is up to the task, and based on the video after the break, it should be able to complete this challenge.
Continue reading “Autonomous Balloon Popping”
Last weekend wasn’t just about Maker Faire; in Stockholm there was another DIY festival celebrating the protocols that make electronic music possible. It’s MIDI Hack 2014, and [Kristian], [Michael], [Bram], and [Tobias] put together something really cool: a Lego sequencer
The system is set up on a translucent Lego base plate, suspended above a webcam that feeds into some OpenCV and Python goodness. From there, data is sent to Native Instruments Maschine. There’s a step sequencer using normal Lego bricks, a fader controlling beat delay, and a rotary encoder for reverb.
Despite being limited to studs and pegs, the short demo in the video below actually sounds good, with a lot of precision found in the faders and block-based rotary encoder. [Kristian] will be putting up the code and a few more details shortly. Hopefully there will be enough information to use different colored blocks in the step sequencer part of the build for different notes.
Continue reading “Turning Lego Into A Groove Machine”
Even for hobby projects, iteration is very important. It allows us to improve upon and fine-tune our existing designs making them even better. [Max] wrote in to tell us about his latest posture sensor, this time, built around a webcam.
We covered [Max’s] first posture sensor back in February, which utilized an ultrasonic distance sensor to determine if you had correct posture (or not). Having spent time with this sensor and having received lots of feedback, he decided to scrap the idea of using an ultrasonic distance sensor altogether. It simply had too many issues: issues with mounting the sensor on different chairs, constantly hearing the clicking of the sensor, and more. After being inspired by a very similar blog post to his original that mounted the sensor on a computer monitor, [Max] was back to work. This time, rather than using an ultrasonic distance sensor, he decided to use a webcam. Armed with Processing and OpenCV, he greatly improved upon the first version of his posture sensor. All of his code is provided on his website, be sure to check it out and give it a whirl!
Iteration leads to many improvements and it is an integral part of both hacking and engineering. What projects have you redesigned or rebuild? Let us know!
[Courtney] has been hard at work on OSkAR, an OpenCV based speaking robot. OSkAR is [Courney’s] capstone project (pdf link) at Shepherd University in West Virginia, USA. The goal is for OSkAR to be an assistive robot. OSkAR will navigate a typical home environment, reporting objects it finds through speech synthesis software.
To accomplish this, [Courtney] started with a Beagle Bone Black and a Logitech C920 webcam. The robot’s body was built using LEGO Mindstorms NXT parts. This means that when not operating autonomously, OSkAR can be controlled via Bluetooth from an Android phone. On the software side, [Courtney] began with the stock Angstrom Linux distribution for the BBB. After running into video problems, she switched her desktop environment to Xfce. OpenCV provides the machine vision system. [Courtney] created models for several objects for OSkAR to recognize.
Right now, OSkAR’s life consists of wandering around the room looking for pencils and door frames. When a pencil or door is found, OSkAR announces the object, and whether it is to his left or his right. It may sound like a rather boring life for a robot, but the semester isn’t over yet. [Courtney] is still hard at work creating more object models, which will expand OSkAR’s interests into new areas.
Continue reading “Never Lose Your Pencil With OSkAR on Patrol”
Need a quick and easy way to sort through a few hundred random resistors? You could do them one at a time by reading the color codes yourself… or you could get a machine to do it for you!
When [Robert] was faced with a pile of unsorted resistors he quickly decided he did not have the patience to sort them manually. So, he started by writing an Android app using OpenCV to detect and identify resistor color codes. The problem is, most phones have trouble focusing at short distances — and since resistors are so small, holding the phone farther back results in color rings only being a few pixels wide — not the greatest for image recognition!
So, he started again on his computer, using a cheap LED-lit webcam instead. He wrote the app in java so he could re-use parts of the code from the Android app. It seems to work pretty well — check it out in the following video! This would be perfect to pair up with your illuminated storage bin hack.
Continue reading “ReSCan — Automated Resistor Identification!”
[Matthiew] needed to create a system that would allow a computer to read braille. An electromechanical system would be annoying to develop and would require many hardware iterations as the system [Matthew] is developing evolves. Instead, he came up with a much better solution using a webcam and OpenCV that still gets 100% accuracy.
Instead of using a camera to look for raised or lowered pins in this mechanical braille display, [Matthiew] is using OpenCV to detect the shadows. This requires calibrating the camera to the correct angle, or in OpenCV terms, pose.
After looking at the OpenCV tutorials, [Matthiew] found a demo that undistorts an image of a chess board. Using this same technique, he used fiducials from the ARTag project to correctly calibrate an image of his mechanical braille pins.
As for why [Matthiew] went through all the trouble to get a computer to read braille – something that doesn’t make a whole lot of sense if you think about it – he’s building a braille eBook reader, something that just screams awesome mechanical design. We’d be interested in seeing some more info on that project as well.