There is a lot of helpful technology for people with mobility issues. Even something that can help people do something most of us wouldn’t think twice about, like turn on a lamp or control a computer, can make a world of difference to someone who can’t move around as easily. Luckily, [Matt] has been working on using webcams and depth cameras to allow someone to do just that.
[Matt] found that using webcams instead of depth cameras (like the Kinect) tends to be less obtrusive but are limited in their ability to distinguish individual users and, of course, don’t have the same 3D capability. With either technology, though, the software implementation is similar. The camera can detect head motion and control software accordingly by emulating keystrokes. The depth cameras are a little more user-friendly, though, and allow users to move in whichever way feels comfortable for them.
Getting decent macro photos always seems to be a chore. Some important detail always seems to be just outside of the depth of field, or you have to be zoomed in so close that you get great detail in one spot but miss the big picture. [Nate B] had such a problem while trying to document some PC boards, and he came up with a nifty hack that uses a laser cutter and a smart phone camera to do the job.
Having first tried scanning the boards with a flat-bed scanner but finding the depth of field unsatisfactory, [Nate B] then went on to his Samsung phone’s camera. Set to panorama mode, he manually scanned across the boards and let the camera stitch the images together. The results were better, but the wobblies got the better of him and the images showed it. He then decided to use a laser cutter — with the laser disabled, of course — as an impromptu X-Y stage to raster his camera above the boards. In a slightly cringe-worthy move, he gingerly clamped the phone to the cutter gantry, started the panorama, and let the cutter move over the board. This results in a rock-solid pictures of his boards with a lot of detail – perfect for his documentation. As a bonus, the honeycomb laser cutter bed makes for an interesting background texture.
Obviously anything could be used to raster a camera and achieve similar results, but full points here for maximizing available resources and not over-complicating a simple job. Yet another reason you can use to justify that laser-cutter purchase.
[Daniel] and [Tobias] dabble in videography and while they would love a camera slider controlled by their favorite iDevice, commercial motorized camera sliders are expensive, and there’s no great open source alternative out there. They decided to build one for themselves that can be controlled either from a PS3 controller or from its own iPad app with the help of an ESP8266 WiFi module.
The camera slider is a two-axis ordeal, with one axis sliding the camera along two solid rails, and the other panning the camera. The circuit board was milled by the guys and includes an ATMega328 controlling two Pololu stepper drivers. An ESP8266 is thrown into the mix, and is easily implemented on the device; it’s just an MAX232 chip listening to the Tx and Rx lines of the WiFi module and translating that to something the ATMega can understand.
By far the most impressive part of this project is the iPad app. This app can be controlled ‘live’ and the movements can be recorded for later playback. Alternatively, the app has a simple scripting function that performs various actions such as movement and rotation over time. The second mode is great for time lapse shots. Because this camera slider uses websockets for the connection, the guys should also be able to write a web client for the slider, just in case they wanted the ultimate webcam.
You can check out [Daniel] and [Tobias]’ demo reel for their camera slider below.
If there’s one thing about laser cutters that makes them a little difficult to use, it’s the fact that it’s hard for a person to interact with them one-on-one without a clunky computer in the middle of everything. Granted, that laser is a little dangerous, but it would be nice if there was a way to use a laser cutter without having to deal with a computer. Luckily, [Anirudh] and team have been working on solving this problem, creating a laser cutter that can interact directly with its user.
The laser cutter is tied to a visual system which watches for a number of cues. As we’ve featured before, this particular laser cutter can “see” pen strokes and will instruct the laser cutter to cut along the pen strokes (once all fingers are away from the cutting area, of course). The update to this system is that now, a user can import a drawing from a smartphone and manipulate it with a set of physical tokens that the camera can watch. One token changes the location of the cut, and the other changes the scale. This extends the functionality of the laser cutter from simply cutting at the location of pen strokes to being able to cut around any user-manipulated image without interacting directly with a computer. Be sure to check out the video after the break for a demonstration of how this works.
We remember making pinhole cameras as kids out of cigar boxes. The Focal Camera website wants to enable you to make sophisticated cameras from a selection of building blocks. We’re talking cameras with film, not digital cameras (although we wondered if you could mount an image sensor… but that’s another hack).
The modules do require access to a laser cutter, and you’ll need to scrounge or otherwise acquire things like mirrors and lenses. The site has advice on how to hack things like first surface mirrors out of cheap items like acrylic mirrors.
The intent is to be able to build up your own cameras from the modules. They do have a pinhole camera, in case you are nostalgic, but you could also build SLRs, large format cameras, or even stereo cameras. Not all the modules are ready yet, but there are several example cameras and pictures taken with them on the site. Like most building blocks, the real treat will be when users begin to combine them in unexpected ways.
The nice thing with having a hacker cred is that family and friends are always on the lookout for stuff they think might be useful to you. [Craig Hollabaugh]’s son-in-law found an inspection camera and thought it would be handy for his hobby work. The MagniSight Explorer was first introduced in 2001. It is good for surface mount board work and inspection, except that its analog 480p video is quite dated by today’s standards. So [Craig] upgraded it for crystal clear 1080P/30 video and 5 megapixel images using a $35 Raspberry Pi 2 and a $26 Raspberry Pi Camera Module. After the upgrade, the unit is now a great tool for SMT rework.
There’s not a lot to the upgrade, but [Craig] gives a nice rundown in the 15 minute video of the MagniSight’s internals. He shows us the original analog camera module and its video card, which is able to do some additional processing like black and white output and reverse video (negative). As he mentions, it would be easy for him to do the same via software on the Raspberry Pi. A video camera lens takes care of magnification and two shafts coupled to it via flat belts (rubber bands?) take care of zoom and focus. A front coated mirror angled 45 degrees in front of the lens turns the optical path 90 degrees to allow the lens/camera to “look down”. After experimenting a bit to find the correct focal spot behind the lens unit for the Raspberry Pi camera, he covered the camera module with insulation tape and then just glued it to the old camera mount. After hooking it up to an HDMI monitor, the results are quite nice and he reckons he can easily work with components down to 0402 in size.
He’s got a couple of more upgrades in mind to make the system even better. He plans to replace the existing compact fluorescent lamps with a string of LED’s which will provide more uniform illumination. Plus, he can control their brightness, and selectively turn them on or off to get the optimum lighting. The other interesting upgrade would be to add stepper motors to the X-Y translation stage and automate their movement. After looking up a board file and its BoM, he may even be able to search for a part designator and move the stage to bring the part into focus.
Light polarization is an interesting phenomenon that is extremely useful in many situations… but human eyes are blind to detecting any polarization. Luckily, [David] has built a polarization-sensitive camera using a Raspberry Pi and a few off-the-shelf components that allows anyone to view polarization. [David] lists the applications as:
A polarimetric imager to detect invisible pollutants, locate landmines, identify cancerous tissues, and maybe even observe cloaked UFOs!
The build uses a standard Raspberry Pi 2 and a 5 megapixel camera which sits behind a software-controlled electro-optic polarization modulator that was scavenged from an auto-darkening welding mask. The mask is essentially a specialized LCD screen, which is easily electronically controlled. [David] whipped up some scripts on the Pi that control the screen, which is how the camera is able to view various polarizations of light. Since the polarization modulator is software-controlled, light from essentially any angle can be analyzed in any way via the computer.
There is a huge amount of information about this project on the project site, as well as on the project’s official blog. There have been other projects that use polarized light for specific applications, but this is the first we’ve seen of a software-controlled polarizing camera intended for general use that could be made by pretty much anyone.