Digitizing film is a tedious process that becomes a lot more fun if you spend more of your time building a digitizer and less time actually working working with old film. [Heikki Hietala] has been at it for years and his Kotokino Mark IV film scanner is a masterpiece of simple machine building.
Since we first saw the film scanner four years ago it’s undergone a number of excellent improvements. Most notably, the point-and-shoot camera has been swapped out for a DSLR. With the use of a macro reversing ring a normal lens is flipped around to blow up the 8-millimeter-wide film to take advantage of all the megapixels available on the camera sensor.
The key to the setup is the film advancer mechanism which takes care of both advancing the film and triggering the camera. As you can see, a servo motor rotating an axle provides the locomotion. The mechanism keys into the perforations in the film to pull it along on the down stroke and closes a switch to trigger the camera on the upstroke. Directly under the lens, the alignment jig uses lens cleaning fabric to avoid scratching the film, while perfectly positioning it over the light source.
Previous versions have placed the camera on the horizontal plane but it seems some vibrations in the system caused alignment problems between captured frames. This latest version places the camera pointed straight down to solve that issue, and brings the entire thing together into one beautiful finished project. Having gathered numerous fans of the build along the way, [Heikki] has made the design files available so that you may build your own version.
In the world of computer security, the good news is that a lot of vendors are finally taking security seriously now, with the result that direct attacks are harder to pull off. The bad news is that in a lot of cases, they’re still leaving the side-door wide open. Side-channel attacks come in all sorts of flavors, but they all have something in common: they leak information about the state of a system through an unexpected vector. From monitoring the sounds that the keyboard makes as you type to watching the minute vibrations of a potato chip bag in response to a nearby conversation, side-channel attacks take advantage of these leaks to exfiltrate information.
Side-channel exploits can be the bread and butter of black hat hackers, but understanding them can be useful to those of us who are more interested in protecting systems, or perhaps to inform our reverse engineering efforts. Samy Kamkar knows quite a bit more than a thing or two about side-channel attacks, so much so that he gave a great talk at the 2019 Hackaday Superconference on just that topic. He’ll be dropping by the Hack Chat to “extend and enhance” that talk, and to answer your questions about side-channel exploits, and discuss the reverse engineering potential they offer. Join us and learn more about this fascinating world, where the complexity of systems leads to unintended consequences that could come back to bite you, or perhaps even help you.
Click that speech bubble to the right, and you’ll be taken directly to the Hack Chat group on Hackaday.io. You don’t have to wait until Wednesday; join whenever you want and you can see what the community is talking about.
Line scan cameras are advanced devices used for process inspection tasks in industrial applications. Used to monitor the quality of silicon wafers and other high-accuracy tasks, they’re often outfitted with top-quality optics that are highly specialised. [Peter] was able to get his hands on a lens for a line-scan camera, and decided to put it to work on some macro photography instead.
Judging by the specs found online, this is a fairly serious piece of kit. It easily competes with top-shelf commercial optics, which is what piqued [Peter]’s interest in the part. Being such a specialised piece of hardware, you can’t just cruise over to eBay for an off-the-shelf adapter. Instead, a long chain of parts were used to affix this lens to a Sony AIII DSLR, converting from threaded fittings to a Nikon mount and then finally to Sony NEX mount.
Further work involved fitting an aperture into the chain to get the lens as close as possible to telecentric. This improves the lens’s performance for certain tasks, and makes focus stacking macro shots more readily achievable – something we’ve seen [Peter] tinker with before.
You never know what you might find when sorting through surplus industrial gear, could could score some high-performance hardware if you know where to look. It’s always great to see a cheap find become a useful instrument in the hacker toolbox!
From cars to refrigerators, it seems as if every new piece of tech is connected to the Internet. For better or for worse, we’re deep into the “Internet of Things”. But what about your camera? No, not the camera in your smartphone; that one’s already connected to the Internet and selling your secrets to the highest bidder. Don’t you think your trusty DSLR could be improved by an infusion of Wide Area Networking?
Regardless of what you’re answer to that question might be, [Thomas Kittredge] decided his life would be improved by making his beloved Canon EOS Rebel T6 an honorary member of the Internet of Things. Truth be told he says that he hasn’t quite figured out an application for this project. But since he was looking to mess around with both the LTE-enabled Particle Boron development board and designing his own PCB for professional production, this seemed a good a way to get his feet wet as any.
The resulting board is a fairly simple “shield” for the Particle Boron that let’s [Thomas] trigger up to two cameras remotely over the Internet or locally with Bluetooth. If LTE isn’t your sort of thing though, don’t worry. Since the Boron follows the Adafruit Feather specification, there’s a whole collection of development boards with various connectivity options that this little add-on can be used with.
In the GitHub repository, [Thomas] has put up the files for the PCB, the STLs for the 3D printed enclosure, and of course the firmware source code to load onto the Particle board. He currently has code to expose the two shutter triggers as functions the the Particle Cloud API, as well as a practical example that fires off the camera when specific words are used in a Slack channel.
Do you have a need to photographically document the doings of warm-blooded animals? If so, a game camera from the nearest hunting supplier is probably your best bet. But if you don’t need the value-added features such as a weather-resistant housing that can be chained to a tree, this DIY motion trigger for a DSLR is a quick and easy build, and probably loads more fun.
The BOM on [Jeremy S Cook]’s build is extremely short – just a PIR sensor and an optoisolator, with a battery, a plug for the camera’s remote jack, and a 3D-printed bracket. The PIR sensor is housed in a shroud to limit its wide field of view; [Jeremy] added a second shroud when an even narrower field is needed. No microcontroller is needed because all it does is trigger the camera when motion is sensed, but one could be added to support more complicated use cases, like an intervalometer or constraining the motion sensing to certain times of the day. The video below shows the build and some quick tests.
Speaking of intervalometers, we’ve seen quite a few of those over the years. From the tiny to the tinier to the electromechanical, people seem to have a thing for taking snapshots at regular intervals.
Sometimes, less is more. Sometimes, more is more. There is a type of person who believes that if enough photos of the same subject are taken, one of them will shine above the rest as a gleaming example of what is possible with a phone camera and a steady hand. Other people know how to frame a picture before hitting the shutter button. In some cases, the best method may be snapping a handful of photos to get one good one, not by chance, but by design.
[The Thought Emporium]’s video, also below the break, is about getting crisp pictures from a DSLR camera and a microscope using focus stacking, sometimes called image stacking. The premise is to take a series of photos that each have a different part of the subject in focus. In a microscope, this range will be microscopic but in a park, that could be several meters. When the images are combined, he uses Adobe products, the areas in focus are saved while the out-of-focus areas are discarded and the result is a single photo with an impossible depth of focus. We can’t help but remember those light-field cameras which didn’t rely on moving lenses to focus but took many photos, each at a different focal range.
On August 21, 2017, the Moon will cast its shadow across the entire breadth of the United States for the first time in almost a century. It is estimated that 12 million people live within the path in which the sun will be blotted out, and many millions more are expected to pour into the area to experience the wonders of totality.
We’d really love it if you would tell us where you’ll be during the eclipse by creating your own event page, but that’s not what this article’s about. With millions gathered in a narrow swath from Oregon to South Carolina, and with the eclipse falling on a Monday so that the prior two weekend days will be filled with campouts at prime viewing locations, I expect that Eclipse 2017 will be one big coast-to-coast party. This is an event that will attract people of all stripes, from those with no interest in astronomy that have only the faintest idea of what’s actually happening celestially, to those so steeped in the science that they’ll be calling out the exact beginning of totality and when to expect Baily’s Beads to appear.
I suspect our readership leans closer to the latter than the former, and some may want to add to the eclipse experience by participating in a little citizen science. Here’s how you can get involved.