From cars to refrigerators, it seems as if every new piece of tech is connected to the Internet. For better or for worse, we’re deep into the “Internet of Things”. But what about your camera? No, not the camera in your smartphone; that one’s already connected to the Internet and selling your secrets to the highest bidder. Don’t you think your trusty DSLR could be improved by an infusion of Wide Area Networking?
Regardless of what you’re answer to that question might be, [Thomas Kittredge] decided his life would be improved by making his beloved Canon EOS Rebel T6 an honorary member of the Internet of Things. Truth be told he says that he hasn’t quite figured out an application for this project. But since he was looking to mess around with both the LTE-enabled Particle Boron development board and designing his own PCB for professional production, this seemed a good a way to get his feet wet as any.
The resulting board is a fairly simple “shield” for the Particle Boron that let’s [Thomas] trigger up to two cameras remotely over the Internet or locally with Bluetooth. If LTE isn’t your sort of thing though, don’t worry. Since the Boron follows the Adafruit Feather specification, there’s a whole collection of development boards with various connectivity options that this little add-on can be used with.
In the GitHub repository, [Thomas] has put up the files for the PCB, the STLs for the 3D printed enclosure, and of course the firmware source code to load onto the Particle board. He currently has code to expose the two shutter triggers as functions the the Particle Cloud API, as well as a practical example that fires off the camera when specific words are used in a Slack channel.
Do you have a need to photographically document the doings of warm-blooded animals? If so, a game camera from the nearest hunting supplier is probably your best bet. But if you don’t need the value-added features such as a weather-resistant housing that can be chained to a tree, this DIY motion trigger for a DSLR is a quick and easy build, and probably loads more fun.
The BOM on [Jeremy S Cook]’s build is extremely short – just a PIR sensor and an optoisolator, with a battery, a plug for the camera’s remote jack, and a 3D-printed bracket. The PIR sensor is housed in a shroud to limit its wide field of view; [Jeremy] added a second shroud when an even narrower field is needed. No microcontroller is needed because all it does is trigger the camera when motion is sensed, but one could be added to support more complicated use cases, like an intervalometer or constraining the motion sensing to certain times of the day. The video below shows the build and some quick tests.
Speaking of intervalometers, we’ve seen quite a few of those over the years. From the tiny to the tinier to the electromechanical, people seem to have a thing for taking snapshots at regular intervals.
Sometimes, less is more. Sometimes, more is more. There is a type of person who believes that if enough photos of the same subject are taken, one of them will shine above the rest as a gleaming example of what is possible with a phone camera and a steady hand. Other people know how to frame a picture before hitting the shutter button. In some cases, the best method may be snapping a handful of photos to get one good one, not by chance, but by design.
[The Thought Emporium]’s video, also below the break, is about getting crisp pictures from a DSLR camera and a microscope using focus stacking, sometimes called image stacking. The premise is to take a series of photos that each have a different part of the subject in focus. In a microscope, this range will be microscopic but in a park, that could be several meters. When the images are combined, he uses Adobe products, the areas in focus are saved while the out-of-focus areas are discarded and the result is a single photo with an impossible depth of focus. We can’t help but remember those light-field cameras which didn’t rely on moving lenses to focus but took many photos, each at a different focal range.
On August 21, 2017, the Moon will cast its shadow across the entire breadth of the United States for the first time in almost a century. It is estimated that 12 million people live within the path in which the sun will be blotted out, and many millions more are expected to pour into the area to experience the wonders of totality.
We’d really love it if you would tell us where you’ll be during the eclipse by creating your own event page, but that’s not what this article’s about. With millions gathered in a narrow swath from Oregon to South Carolina, and with the eclipse falling on a Monday so that the prior two weekend days will be filled with campouts at prime viewing locations, I expect that Eclipse 2017 will be one big coast-to-coast party. This is an event that will attract people of all stripes, from those with no interest in astronomy that have only the faintest idea of what’s actually happening celestially, to those so steeped in the science that they’ll be calling out the exact beginning of totality and when to expect Baily’s Beads to appear.
I suspect our readership leans closer to the latter than the former, and some may want to add to the eclipse experience by participating in a little citizen science. Here’s how you can get involved.
Microscopes are a great way to see the mysteries of the universe hidden at the smaller scale. When they were first developed, scientists had to rely on illustration to convey their findings through the lens. Thankfully we can now rely on photography to help us out instead. Many microscopes come with a special port — often called a trinocular port — for mounting a camera. Using this, [Brian] developed a DSLR mount for his microscope using a hacker staple: PVC pipe.
The virtues of PVC pipe are many and varied. It’s readily available in all manner of shapes and sizes, and there’s a wide variety of couplers, adapters, solvents and glues to go with it. Best of all, you can heat it to a point where it becomes soft and pliable, allowing one to get a custom fit where necessary. [Brian] demonstrates this in using a heat gun to warm up a reducer to friction fit the DSLR lens mount. Beyond that, the mount uses a pair of lenses sourced from jeweller’s loupes to bring the image into focus on the camera’s sensor, mounted tidily inside the PVC couplers.
Like any Moore’s Law-inspired race, the megapixel race in digital cameras in the late 1990s and into the 2000s was a harsh battleground for every manufacturer. With the development of the smartphone, it became a war on two fronts, with Samsung eventually cramming twenty megapixels into a handheld. Although no clear winner of consumer-grade cameras was ever announced (and Samsung ended up reducing their flagship phone’s cameras to sixteen megapixels for reasons we’ll discuss) it seems as though this race is over, fizzling out into a void where even marketing and advertising groups don’t readily venture. What happened?
A brief overview of Moore’s Law predicts that transistor density on a given computer chip should double about every two years. A digital camera’s sensor is remarkably similar, using the same silicon to form charge-coupled devices or CMOS sensors (the same CMOS technology used in some RAM and other digital logic technology) to detect photons that hit it. It’s not too far of a leap to realize how Moore’s Law would apply to the number of photo detectors on a digital camera’s image sensor. Like transistor density, however, there’s also a limit to how many photo detectors will fit in a given area before undesirable effects start to appear.
Image sensors have come a long way since video camera tubes. In the ’70s, the charge-coupled device (CCD) replaced the cathode ray tube as the dominant video capturing technology. A CCD works by arranging capacitors into an array and biasing them with a small voltage. When a photon hits one of the capacitors, it is converted into an electrical charge which can then be stored as digital information. While there are still specialty CCD sensors for some niche applications, most image sensors are now of the CMOS variety. CMOS uses photodiodes, rather than capacitors, along with a few other transistors for every pixel. CMOS sensors perform better than CCD sensors because each pixel has an amplifier which results in more accurate capturing of data. They are also faster, scale more readily, use fewer components in general, and use less power than a comparably sized CCD. Despite all of these advantages, however, there are still many limitations to modern sensors when more and more of them get packed onto a single piece of silicon.
While transistor density tends to be limited by quantum effects, image sensor density is limited by what is effectively a “noisy” picture. Noise can be introduced in an image as a result of thermal fluctuations within the material, so if the voltage threshold for a single pixel is so low that it falsely registers a photon when it shouldn’t, the image quality will be greatly reduced. This is more noticeable in CCD sensors (one effect is called “blooming“) but similar defects can happen in CMOS sensors as well. There are a few ways to solve these problems, though.
First, the voltage threshold can be raised so that random thermal fluctuations don’t rise above the threshold to trigger the pixels. In a DSLR, this typically means changing the ISO setting of a camera, where a lower ISO setting means more light is required to trigger a pixel, but that random fluctuations are less likely to happen. From a camera designer’s point-of-view, however, a higher voltage generally implies greater power consumption and some speed considerations, so there are some tradeoffs to make in this area.
Another reason that thermal fluctuations cause noise in image sensors is that the pixels themselves are so close together that they influence their neighbors. The answer here seems obvious: simply increase the area of the sensor, make the pixels of the sensor bigger, or both. This is a good solution if you have unlimited area, but in something like a cell phone this isn’t practical. This gets to the core of the reason that most modern cell phones seem to be practically limited somewhere in the sixteen-to-twenty megapixel range. If the pixels are made too small to increase megapixel count, the noise will start to ruin the images. If the pixels are too big, the picture will have a low resolution.
There are some non-technological ways of increasing megapixel count for an image as well. For example, a panoramic image will have a megapixel count much higher than that of the camera that took the picture simply because each part of the panorama has the full mexapixel count. It’s also possible to reduce noise in a single frame of any picture by using lenses that collect more light (lenses with a lower f-number) which allows the photographer to use a lower ISO setting to reduce the camera’s sensitivity.
Of course, if you have unlimited area you can make image sensors of virtually any size. There are some extremely large, expensive cameras called gigapixel cameras that can take pictures of unimaginable detail. Their size and cost is a limiting factor for consumer devices, though, and as such are generally used for specialty purposes only. The largest image sensor ever built has a surface of almost five square meters and is the size of a car. The camera will be put to use in 2019 in the Large Synoptic Survey Telescope in South America where it will capture images of the night sky with its 8.4 meter primary mirror. If this was part of the megapixel race in consumer goods, it would certainly be the winner.
With all of this being said, it becomes obvious that there are many more considerations in a digital camera than just the megapixel count. With so many facets of a camera such as physical sensor size, lenses, camera settings, post-processing capabilities, filters, etc., the megapixel number was essentially an easy way for marketers to advertise the claimed superiority of their products until the practical limits of image sensors was reached. Beyond a certain limit, more megapixels doesn’t automatically translate into a better picture. As already mentioned, however, the megapixel count can be important, but there are so many ways to make up for a lower megapixel count if you have to. For example, images with high dynamic range are becoming the norm even in cell phones, which also helps eliminate the need for a flash. Whatever you decide, though, if you want to start taking great pictures don’t worry about specs; just go out and take some photographs!
(Title image: VISTA gigapixel mosaic of the central parts of the Milky Way, produced by European Southern Observatory (ESO) and released under Creative Commons Attribution 4.0 International License. This is a scaled version of the original 108,500 x 81,500, 9-gigapixel image.)
The next giant leap for mankind is to the stars. While we are mostly earthbound — for now — that shouldn’t stop us from gazing upwards to marvel at the night sky. In saying that, if you’re an amateur astrophotographer looking to take long-exposure photos of the Milky Way and other stellar scenes, [Anthony Urbano] has devised a portable tracking setup to keep your photos on point.
When taking pictures of the night sky, the earth’s rotation will cause light trails during long exposures. Designed for ultra-portability, [Urbano’s] rig uses an Arduino UNO controlled Sanryusha P43G geared stepper motor coupled to a camera mounting plate on a small tripod. The setup isn’t designed for anything larger than a DSLR, but is still capable of taking some stellar pictures.