The tuatara is a reptile native to New Zealand, and thanks to the descendants of stowaway rats on 17th century ships, these little lizards are critically endangered. [Warren] was asked if he could film one of these hatchlings being born and pulled out a Raspberry Pi to make it happen.
[Warren] constructed a small lasercut box to house the incubating egg, but he hit a few snags figuring out how to properly focus the Raspi camera board. The original idea was to use a Nikkor macro lens, without any kind of adapter between it and the camera board. A bit of googling lead [Warren] to this tutorial for modifying the focus on the Raspi camera, giving him a good picture.
The incubator had no windows and thus no light, making an IR LED array the obvious solution to the lighting problem. Time was of the essence, so an off-the-shelf security camera provided the IR illumination. After dumping the video to his computer, [Warren] had a video of a baby tuatara hatching. You can check that out below.
Continue reading “Recording Time Lapse of Endangered Reptiles Hatching”
Automating the growing process of plants and vegetables is an increasing trend among gardening enthusiasts and hobbyists. It’s no surprise, either, with microcontrollers, moisture sensors, Co2 detectors, and even time-lapse cameras with rotating wooden rigs that are in the hands of millions of amateur gardeners around the world.
This project by [Liz] helps to document the sprouting process of her tiny grapefruit bonsai tree that started to flourish at her apartment in Chicago.
Similar rigs can be used for practically any type of indoor plant. They can also be modified to move the plants and vegetables depending on how much light they are getting. Even further, just add some code to splice the photographs together and you’ve got yourself a custom setup that can produce animated GIF files to be uploaded easily to the internet. Pages and pages of happy and healthy growing plants unearthing themselves from the ground up would be pasted all over the web showing the entire sprouting process. An example video of this by [Liz] is embedded below.
Continue reading “Rotating Plants for Time-Lapse Purposes”
First person video – between Google Glass, GoPro, and other sports cameras, it seems like everyone has a camera on their head these days. If you’re a surfer or skydiver, that might make for some awesome footage. For the rest of us though, it means hours of boring video. The obvious way to fix this is time-lapse. Typically time-lapse throws frames away. Taking 1 of every 10 frames results in a 10x speed increase. Unfortunately, speeding up a head mounted camera often leads to a video so bouncy it can’t be watched without an air sickness bag handy. [Johannes Kopf], [Michael Cohen], and [Richard Szeliski] at Microsoft Research have come up with a novel solution to this problem with Hyperlapse.
Hyperlapse photography is not a new term. Typically, hyperlapse films require careful planning, camera rigs, and labor-intensive post-production to achieve a usable video. [Johannes] and team have thrown computer vision and graphics algorithms at the problem. The results are nothing short of amazing.
The full details are available in the team’s report (35MB PDF warning). To obtain usable data, the fisheye lenses often used on these cameras must be calibrated. The team accomplished that with the OCamCalib toolbox. Imported video is broken down frame by frame. Using structure from motion algorithms, hyperlapse creates a 3D models of the various scenes in the video. With the scenes in this virtual world, the camera can be moved and aimed at will. The team’s algorithms then pick a smooth path that follows the original cameras trajectory. Once the camera’s position is known, it’s simply a matter of rendering the final video.
The results aren’t perfect. The mountain climbing scenes show some artifacts caused by the camera frame rate and exposure changing due to the varied lighting conditions. People appear and disappear in the bicycling portion of the video.
One thing the team doesn’t mention is how long the process takes. We’re sure this kind of rendering must require some serious time and processing power. Still, the output video is stunning.
Continue reading “Hyperlapse Makes Your HeadCam Videos Awesome”
[themonkeybars] recently uploaded a time-lapse video of his DIY synthesizer build. First off the video itself is a pretty neat hack. An iPhone time-lapse app was used to capture one frame every 5 seconds. By the time the build was complete, approximately 46,000 frames had been snapped. This boiled down to over 43 minutes of youtube footage. [themonkeybars] didn’t work full time on the project, so the video covers about a year’s worth of work which we think makes it even cooler. The synth is also featured in much of the video’s soundtrack.
The synthesizer itself would be classified as an analog modular synth, a type we’ve seen before. Modular synthesizers are one of the earlier forms of electronic music. The synthesizer is composed of discrete modules such as oscillators, modulators, and filters. The modules may be housed in the same box, but they are not internally connected. All connections are made via front panel patch cables. This is where the term “Patch” came from. Continue reading “Time-lapse synthesizer build will blow your mind”