Digital photo frames these days require you to manage the photos stored on it or the cloud-based service tied to the frame’s manufacturer. [Henric Andersson] realized that he and his wife take a lot of photos but find little time to go through them — like photo albums of days past — and add them to any photo frame-like appliance or service. Since Google photos can do a lot of the sorting for them, he decided to incorporate that into a digital photo frame.
Using his wife’s old Viewsonic 24” 1080p monitor, he cracked it open and incorporated the screen into a 24×16 distressed wood frame — reinforcing it to account for the bulky, built-in power supply with pieces of HDF and a lot of glue. The brains behind this digital photo frame is a Raspberry Pi 3 he received from a friend. To turn the whole on/off, he built a small circuit but it turned out it wasn’t strictly necessary since everything started just fine without it.
While functionally complete, it needed one more addition. A little thing called ‘color temperature calibration’ — aka white balance.
Finding the TCS34725 RGB color sensor by Adafruit — and readily available code for easy integration — [Andersson] puzzled over how to add it to the frame. To disguise it while retaining its effectiveness, he had to glue it to the rear of the frame after drilling a hole in the top piece and sticking a plastic stick through the hole to let light through to the sensor.
To get the photos to display, [Henric Andersson] says all he did was add a few queries to Google Photos and it will display all your relevant photos that have been synced to the service. For a breakdown of that side of this hack, check out his other post with the details.
In what began as a personal challenge he issued to himself, [Fred] is in the process of building an underwater camera that’s capable of long-term photography in shallow waters. He’d like it to last about five hours on a charge while taking a photo every five minutes. Ideally, it will be as cheap as possible and constructed from readily available parts. Solving the cheap/available equation would theoretically make the camera easily to replicate, which is the third major requirement.
[Fred] has recently made great strides, both in the circuitry and the capsule design. The latest version uses a Raspberry Pi 3 with a V2 camera module and runs on a 12 V, 2.4 Ah rechargeable lead-acid battery. Everything is mounted on a piece of hardboard that slides into a 110mm piece of PVC. At one end, the camera looks out through a 10mm acrylic lens fixed into a heavy-duty PVC fitting, and a DS1307 RTC provides a handy clock for shooting time lapses. With a friend’s help, he pressure-tested the housing and found that it can withstand 4 bar without leaking. He is still doing dry tests and trying hard to resist the urge to throw it in the water.
PipeCam is a work in progress, and [Fred] has many ideas for improvements. He’d like to add an Arduino to govern the battery use and provide its vital signs back to the Pi, and add an LDR to decide whether there’s enough light to warrant turning the Pi on to take pictures.
Everyone has a box or two at home somewhere full of family photographs and slides from decades past. That holiday with Uncle Joe in Florida perhaps, or an unwelcome reminder of 1987’s Christmas jumper. It’s fair to say that some memories deserve to be left to gather dust, but what about the others in a world of digital images?
You could of course buy a film scanner to digitize Uncle Joe on the beach, but aside from the dubious quality of so many of them where’s the fun in that? Instead, how about 3D printing one? That’s what [Alexander Gee] did, in the form of an adapter to fit the lens mount of his Sony camera that contains both a 50mm enlarger lens and a mount for the slide. It’s a simple enough print, but he’s made enough parts parametric for users to be able to adjust it to their own camera’s mount.
Sometimes builds do not have to be complex, push boundaries, or contain more computing power than took us to the Moon. This one is simple and well-executed, and for anyone prepared to experiment could deliver results with a variety of cameras and lenses. Of course, you have to have some film to scan before you can use it, so perhaps you’d like to try a bit of home developing.
[Eric Strebel] uses a small homemade vehicle with his camera mounted on it to get great tracking shots for the intros to his videos. If the movement is slow enough then the effect is quite professional looking. But he wanted it eight times slower. We not only like the simple way he did it, along with how he machined parts for it, but the result makes it look like a hot rod, hence his name for it, the dolly hot rod. He also has an elegant mechanism for disengaging the motor while he repositions the dolly.
The are many ways to slow down a rotation. We’re assuming he was already at the minimum speed for the vehicle’s 8 RPM motor transmission and electronic speed controller. Gears or pulleys would probably be the next options. But [Eric] went even simpler, switching from roller blade wheels to larger diameter scooter wheels.
As simple as that sounds though, it led to that age-old conundrum, how to attach the wheels to the vehicle. The axle is made up of PVC tubes. So he machined square the ends of some PVC plugs and bolted the plugs to the wheel bearings. That left only to push the PVC plugs into the axle’s tubes. There are a number of ways he could have machined the PVC plugs, and the full explanation of the one he chose is best left to his video below. But basically, it involved first machining a Bondo body filler cylinder with a bolt embedded in it and then using the cylinder to hold onto the PVC plug while he machined that.
Some time before experimenting with MRI machines and building his own CT scanner, [Peter Jansen] wanted to visualize magnetic fields. One of his small side projects is building tricoders — pocket sensor suites that image everything — and after playing around with the magnetometer function on his Roddenberry-endorsed tool, he decided he had to have a way to visualize magnetic fields. After some work, he has the tools to do it at thousands of frames per second. It’s a video camera for magnetic fields, pushing the boundaries of both magnetic imaging technology and the definition of the word ‘camera’.
When we last looked at [Peter]’s Hall effect camera, the device worked, but it wasn’t necessarily complete. The original design used I2C I/O multiplexers for addressing each individual ‘pixel’ of the Hall effect array, limiting the ‘framerate’ of the ‘camera’ to somewhere around 30 Hz. While this would work for visualizing static magnetic fields, the more interesting magnetic fields around us are oscillating — think motors and transformers and such. A much faster magnetic camera was needed, and that’s what [Peter] set out to build.
Instead of an I/O expander, [Peter] re-engineered his design to use analog multiplexers and a binary counter to cycle through each pixel, one at a time. Basically, the new circuit uses two analog muxes for the columns and rows of the Hall effect array, a binary counter to cycle through each pixel at Megahertz speed, and a fast ADC to read each value. It is, bizarrely, the 1970s way of doing things; these are simple chips, and the controller (a Chipkit Max32) only needs to read a single analog value and clock the binary counter really fast.
With the new design, [Peter] is able to get extremely fast frame rates of about 2,000 Hz. That’s fast enough for some beautiful visualizations of spinning motors and transformers, seen in the video below. Further improvements may include three-axis magnetometers, which should allow for some spectacular visualizations similar to [Ted Yapo]’s 3D magnetic field scanner.
It’s depressingly easy to make bad videos, but it only takes a little care to turn that around. After ample lighting and decent audio — and not shooting in portrait — perhaps the biggest improvements come from stabilizing the camera while it’s moving. Giving your viewers motion sickness is bad form, after all, and to smooth out those beauty shots, a camera slider can be a big help.
Not all camera sliders are built alike, though, and we must admit to being baffled while first watching [Rulof Maker]’s build of a smooth, synchronized pan and slide camera rig. We just couldn’t figure out how those gears were going to be put to use, but as the video below progresses, it becomes clear that this is an adjustable pantograph rig, and that [Rulof]’s eBay gears are intended to link the two sets of pantograph arms together. The arms are formed from threaded pipe and tee fittings with bearings pressed into them, which is a pretty clever construction technique that seems highly dependent on having the good fortune to find bearings with an interference fit into the threads. But still, [Rulof] makes it work, and with a little epoxy and a fair amount of finagling, he ends up with a complex linkage that yields the desired effects. And bonus points for being able to configure the motion with small adjustments to the camera bracket pivot points.
We saw a similar pantograph slider a few months back. That one was 3D-printed and linked with timing belts, but the principles are the same and the shots from both look great.
Here’s something really wonderful. [Dave Akerman] wrote up the results of his attempt to use a high-altitude balloon to try to re-create a famous image of NASA’s Bruce McCandless floating freely in space with the Earth in the background. [Dave] did this in celebration of the 34th anniversary of the first untethered spacewalk, even going so far as to launch on the same day as the original event in 1984. He had excellent results, with plenty of video and images recorded by his payload.
Adhering to the actual day of the spacewalk wasn’t the only hurdle [Dave] jumped to make this happen. He tracked down an old and rare “Astronaut with MMU” (Mobile Maneuvering Unit) plastic model kit made by Revell USA and proceeded to build it and arrange for it to remain in view of the cameras. Raspberry Pi Zero Ws with cameras, LoRA hardware, action cameras, and a UBlox GPS unit all make an appearance in the balloon’s payload.
Sadly, [Bruce McCandless] passed away in late 2017, but this project is a wonderful reminder of that first untethered spacewalk. Details on the build and the payload, as well as the tracking system, are covered here on [Dave]’s blog. Videos of the launch and the inevitable balloon burst are embedded below, but more is available in the summary write-up.