It all started with a bad smell coming from the heat register. [CuddleBurrito] recalled a time when something stinky ended up in the ductwork of his folks’ house which ended up costing them big bucks to explore. The hacker mindset shies away from those expenditures and toward literally rolling your own solution to investigating the funk. In the process [CuddleBurrito] takes us on a journey into the bowels of his house.
We’d call it a robot, but [Eric Buijs] calls it a dolly. [Eric] bought a Makeblock starter robot kit last year, but never did anything with it. He recently wanted a camera dolly to help shoot project videos and the Makeblock hardware fit the bill.
[Eric] found that one of Makeblock’s example videos showed off a camera dolly but had no construction details. He cracked open the kit and got to work replicating what he had seen. Two 6V motors combined with a reduction gear, a belt, and some wheels, and the dolly now moves under computer control!
We’re on the ground here at the 6th Annual World Maker Faire in Queens, New York. This year the Faire is even bigger, extending out from the New York Hall of Science towards Flushing Meadows Corona Park.
Just inside the gates, we ran into [Tommy Mintz] who was showing off his improved portable automated digital photo collage (IP-ADC). Tommy has connected a Raspberry Pi and its camera on a long extension cable. The camera resides high up on a monopod, giving it a bird’s eye view of the area. The Pi first captures a background image, then grabs shots of things that change within the scene. The resulting collages range from hilarious to the surreal. The entire system is mounted on a cart and powered by batteries.
[Tommy] hooked us up with the WiFi password, so we were able to download photos directly from the IP-ADC. Less technologically inclined folks would be able to grab physical prints from the on-board Epson photo printer.
We’ll be reporting what we see here at the faire, so drop us a tweet @hackaday if you want us to stop by!
Cell phones have killed many industries. It is getting harder and harder to justify buying an ordinary watch, a calculator, or a day planner because your phone does all those things at least as well as the originals. Cell phones have cameras too, so the days of missing a shot because you don’t have a camera with you are over (although we always wonder where the flood of Bigfoot and UFO pictures are). However, you probably still have a dedicated camera tucked away somewhere because, let’s face it, most cell phone cameras are just not that good.
The Raspberry Pi camera is about on par with a cheap cell phone camera. [Martijn Braam] has a Nikon camera, and he noticed that he could get a Raspberry Pi camera with a C-mount for lenses. He picked up a C to F adapter and proceeded to experiment with Nikon DSLR lenses on the Raspberry Pi camera. (Update: We’ve changed the link to [Martijn’s] original blog post instead of a copy of it.)
There is a lot of helpful technology for people with mobility issues. Even something that can help people do something most of us wouldn’t think twice about, like turn on a lamp or control a computer, can make a world of difference to someone who can’t move around as easily. Luckily, [Matt] has been working on using webcams and depth cameras to allow someone to do just that.
[Matt] found that using webcams instead of depth cameras (like the Kinect) tends to be less obtrusive but are limited in their ability to distinguish individual users and, of course, don’t have the same 3D capability. With either technology, though, the software implementation is similar. The camera can detect head motion and control software accordingly by emulating keystrokes. The depth cameras are a little more user-friendly, though, and allow users to move in whichever way feels comfortable for them.
This isn’t the first time something like a Kinect has been used to track motion, but for [Matt] and his work at Beaumont College it has been an important area of ongoing research. It’s especially helpful since the campus has many things on network switches (like lamps) so this software can be used to help people interact much more easily with the physical world. This project could be very useful to anyone curious about tracking motion, even if they’re not using it for mobility reasons. Continue reading “Head Gesture Tracking Helps Limited Mobility Students”
Getting decent macro photos always seems to be a chore. Some important detail always seems to be just outside of the depth of field, or you have to be zoomed in so close that you get great detail in one spot but miss the big picture. [Nate B] had such a problem while trying to document some PC boards, and he came up with a nifty hack that uses a laser cutter and a smart phone camera to do the job.
Having first tried scanning the boards with a flat-bed scanner but finding the depth of field unsatisfactory, [Nate B] then went on to his Samsung phone’s camera. Set to panorama mode, he manually scanned across the boards and let the camera stitch the images together. The results were better, but the wobblies got the better of him and the images showed it. He then decided to use a laser cutter — with the laser disabled, of course — as an impromptu X-Y stage to raster his camera above the boards. In a slightly cringe-worthy move, he gingerly clamped the phone to the cutter gantry, started the panorama, and let the cutter move over the board. This results in a rock-solid pictures of his boards with a lot of detail – perfect for his documentation. As a bonus, the honeycomb laser cutter bed makes for an interesting background texture.
Obviously anything could be used to raster a camera and achieve similar results, but full points here for maximizing available resources and not over-complicating a simple job. Yet another reason you can use to justify that laser-cutter purchase.
[Daniel] and [Tobias] dabble in videography and while they would love a camera slider controlled by their favorite iDevice, commercial motorized camera sliders are expensive, and there’s no great open source alternative out there. They decided to build one for themselves that can be controlled either from a PS3 controller or from its own iPad app with the help of an ESP8266 WiFi module.
The camera slider is a two-axis ordeal, with one axis sliding the camera along two solid rails, and the other panning the camera. The circuit board was milled by the guys and includes an ATMega328 controlling two Pololu stepper drivers. An ESP8266 is thrown into the mix, and is easily implemented on the device; it’s just an MAX232 chip listening to the Tx and Rx lines of the WiFi module and translating that to something the ATMega can understand.
By far the most impressive part of this project is the iPad app. This app can be controlled ‘live’ and the movements can be recorded for later playback. Alternatively, the app has a simple scripting function that performs various actions such as movement and rotation over time. The second mode is great for time lapse shots. Because this camera slider uses websockets for the connection, the guys should also be able to write a web client for the slider, just in case they wanted the ultimate webcam.
You can check out [Daniel] and [Tobias]’ demo reel for their camera slider below.