[Eric] just sent in this awesome Kinect hack that he and a few friends worked on. Playing Super Smash Bros with a Kinect.
The system makes use of two Kinects, and three PCs. The first Kinect records each individual players moves, while the second Kinect watches both players “fight” each other. The first PC runs an Nintendo 64 emulator to play the game.
The second PC runs a camera with OpenCV to add another cool but perhaps unnecessary feature, you see, even the character selection is a physical process, adding to the idea of playing the entire game with your body. A glass table allows players to set their 3D printed token onto the glass, effectively placing it on the character they would like to use.
And when the match ends, a windshield wiper knocks off the losing player’s token from the table.
The third PC is responsible for running both Kinects, which then has to send the resulting commands back to first PC over a TCP connection for input into the game.
They introduced it to the public at MHacks Fall 2014, a hacking competition sponsored by Dell and Intel. Video Below.
Continue reading “Super Smash Bros Gets a Revamp with the Microsoft Kinect”
By now you’ve seen almost anything Tweet. But have you seen the (French) twittering chicken coop? (Google translate link) [Hugo] had kept two chickens as part of a household-waste reduction campaign, and then afterward started work.
Even if you don’t read French, the chickens’ twitter feed basically tells the story.
The setup can take IR photographs of sleeping chickens and notify [Hugo] when it’s time to collect the eggs. Naturally, an abundance of other sensors are available. The coop can tweet based on ambient temperature, nest temperature, light level, motion sensor status, or the amount of remaining chicken feed. You can easily follow whether the two fowl are in the coop or out in the yard. It’s like Big Brother, only for birds.
The application is, frankly, ridiculous. But if you’re into home (or coop) automation, there’s a lot to be learned and the project is very well documented. [Hugo] used OpenCV for visual egg detection, and custom Python code to slightly randomize the tweets’ text. All of these details are up on his Github account.
And if you just can’t get enough chicken-coop hacks, be sure to check out this mobile chicken coop, this coop in the shape of a golden spiral, or this Bluetooth-enabled, talking chicken coop, among others. You’d think our name was Coop-a-Day.
Even with visions of quadcopters buzzing around metropolitan areas delivering everything from pizzas to toilet paper fresh in the minds of tech blogospherites, There’s been a comparatively small amount of research into how to support squadrons of quadcopters and other unmanned aerial vehicles. The most likely cause of this is the FAA’s reactionary position towards UAVs. Good thing [Giovanni] is performing all his research for autonomous recharging and docking for multirotors in Australia, then.
The biggest obstacle of autonomous charging of a quadcopter is landing a quad exactly where the charging station is; run of the mill GPS units only have a resolution of about half a meter, and using a GPS solution would require putting GPS on the charging station as well. The solution comes from powerful ARM single board computers – in this case, an Odroid u3 – along with a USB webcam, OpenCV and a Pixhawk autopilot.
Right now [Giovanni] is still working out the kinks on his software system, but he has all the parts and the right tools to get this project up in the air, down, and back up again.
The project featured in this post is a semifinalist in The Hackaday Prize.
[Bharath] recently uploaded the source code for an OpenCV based pattern recognition platform that can be used for Augmented Reality, or even robots. It was built with C++ and utilized the OpenCV library to translate marker notations within a single frame.
The program started out by focusing in on one object at a time. This method was chosen to eliminate the creation of additional arrays that contained information of all of the blobs inside the image; which could cause some problems.
Although this implementation did not track marker information through multiple frames, it did provide a nice foundation for integrating pattern recognition into computer systems. The tutorial was straightforward and easy to ready. The entire program and source code can be found on Github which comes with a ZERO license so that anyone can use it. A video of the program comes up after the break:
Continue reading “Open Source Marker Recognition for Augmented Reality”
[Dr. Fortin] teaches physics at a French High School, and to get his students interested in the natural world around them, he built a geomagnetic observatory, able to tell his students if they have a chance at seeing an aurora, or if a large truck just drove by.
We’ve seen this sort of device before, and the basic construction is extremely similar – a laser shines on a mirror attached to magnets. When a change occurs in the local magnetic field, the mirror rotates slightly and the laser beam is deflected. Older versions have used photoresistors, but [the doctor] is shining his laser on a piece of paper and logging everything with a webcam and a bit of OpenCV.
The design is a huge improvement over earlier DIY attempts at measuring the local magnetic field, if only because the baseline between the webcam and mirror are so long. When set up in his house, the magnetometer can detect cars parked in front of his building, but the data he’s collecting (French, but it’s just a bunch of graphs) is comparable to the official Russian magnetic field data.
We’re pretty sure the Hackaday demographic is a a person who sees a giant tower crane lifting beams and girders above a skyline and says, “that would be fun, at least until I have to go to the bathroom.” Realizing the people who own these cranes probably won’t let any regular joe off the street into the cabin, [Thomas] and [screen Name] (see, this is why we have brackets, kids) built their own miniature version with an Oculus Rift.
Instead of a crane that is hundreds of feet tall, the guys are using a much smaller version, just over a meter tall, that is remotely controlled through a computer via a serial connection. Just below the small plastic cab is a board with two wide-angle webcams. The video from these cameras are sent to the Oculus so the operator can see the boom swinging around, and the winch unwinding to pick up small objects.
The guys have also added a little bit of OpenCV to add color based object detection. This is somewhat useful, but there’s also an approximation of the distance to an object, something that would be very useful if you don’t have a three-inch tall spotter on the ground.
Continue reading “The Crane Game, Oculus Style”
Taking on an autonomous vehicle challenge, [Randy] put together this drone which can locate and pop balloons. It’s been assembled for this year’s Sparkfun Autonomous Vehicle Competition, which will challenge entrants to locate and pop
99 luftbaloons red balloons without human intervention.
The main controller for this robot is the Pixhawk, which runs a modified version of the ArduCopter firmware. These modifications enable the Pixhawk to receive commands from an Odroid U3 computer module. The Odroid uses a webcam to take images, and then processes them using OpenCV. It tries to locate large red objects and fly towards them.
The vision processing and control code on the Odroid was developed using MAVProxy and Drone API. This allows for all the custom code to be developed using Python.
The Sparkfun AVC takes place tomorrow — June 21st in Boulder, Colorado. You can still register to spectate for free. We’re hoping [Randy]’s drone is up to the task, and based on the video after the break, it should be able to complete this challenge.
Continue reading “Autonomous Balloon Popping”