Although scientists have known about Ultra-High Energy Cosmic Rays (UHECRs) for years, nobody can pinpoint their origin. When these UHECRs hit the ground, however, they cause a widespread local disturbance called an air shower. This air shower is a wide dispersion of photons, muons, and electrons at sea level. The means of observing this air shower mandates a widespread geographic region for detecting them. One solution would be a very big detector. Physicists [Daniel] and [Michael] discovered an alternative to pricey hardware, though. By leveraging the CMOS sensors in our smartphones, they can borrow some CPU cycles on our phones to create a worldwide detector network.
According to their paper, the CMOS camera in our smartphones is sensitive to the spectrum of radiation induced by muons and photons from these air showers. With an app running on our phones, [Daniel], [Michael], and other scientists can aggregate the data from multiple detections in a similar region to better understand their origins.
If you’re concerned about CRAYFIS taking away from your talk or web-browsing time, fear not; it runs in the background when a power source has been detected, hopefully, when you are asleep. It’s not the first time we see scientists tap into our computing resources, but this is certainly an achievement made possible in only the last few years by the sensor-loaded smartphone that charges on many of our night stands. With over 1.5 billion smartphones active in the world, we’re thrilled to see a team cleverly leveraging a ubiquitous and already-well-distributed resource.
The Oculus Rift and all the other 3D video goggle solutions out there are great if you want to explore virtual worlds with stereoscopic vision, but until now we haven’t seen anyone exploring real life with digital stereoscopic viewers. [pabr] combined the Kinect-like sensor in an ASUS Xtion with a smartphone in a Google Cardboard-like setup for 3D views the human eye can’t naturally experience like a third-person view, a radar-like display, and seeing what the world would look like with your eyes 20 inches apart.
[pabr] is using an ASUS Xtion depth sensor connected to a Galaxy SIII via the USB OTG port. With a little bit of code, the output from the depth sensor can be pushed to the phone’s display. The hardware setup consists of a VR-Spective, a rather expensive bit of plastic, but with the right mechanical considerations, a piece of cardboard or some foam board and hot glue would do quite nicely.
[pabr] put together a video demo of his build, along with a few examples of what this project can do. It’s rather odd, and surprisingly not a superfluous way to see in 3D. You can check out that video below.
Continue reading “Seeing The World Through Depth Sensing Cameras”
[Daniel Whiteson and Michael Mulhearn], researchers at the University of California, have come up with a novel method of detecting ultra-high energy cosmic rays (UHECR) using smartphones. UHECR are defined as having energy greater than 1018eV. They are rare and very difficult to detect with current arrays. In order to examine enough air showers to detect UHECR, more surface area is needed. Current arrays, like the Pierre Auger Observatory and AGASA, cannot get much larger without dramatically increasing cost. A similar THP Quarterfinalist project is the construction of a low-cost cosmic ray observatory, where it was mentioned that more detection area is needed in order to obtain enough data to be useful.
[Daniel Whiteson and Michael Mulhearn] and colleagues noted that smartphone cameras with CMOS sensors can detect ionizing radiation, which means they also will pick up muons and high-energy photons from cosmic rays. The ubiquitous presence of smartphones makes their collective detection of air showers and UHECR an intriguing possibility. To make all this happen, [Whiteson and Mulhearn] created a smartphone app called CRAYFIS, short for Cosmic RAYs Found In Smartphones. The app turns an idle smartphone into a cosmic ray detector. When the screen goes to sleep and the camera is face-down, CRAYFIS starts taking data from the camera. If a cosmic ray hits the CMOS sensor, the image data is stored on the smartphone along with the arrival time and the phone’s geolocation. This information is uploaded to a central server via the phone’s WiFi. The user does not have to interact with the app beyond installing it. It’s worth noting that CRAYFIS will only capture when the phone is plugged in, so no worries about dead batteries.
The goal of CRAYFIS is to have a minimum of one million smartphones running the app, with a density of 1000 smartphones per square kilometer. As an incentive, anyone whose smartphone data is used in a future scientific paper will be listed as an author. There are CRAYFIS app versions for Android and iOS platforms according to the site. CRAYFIS is still in beta, so the apps aren’t publicly available. Head over to the site to join up!
It’s been a little while since we’ve heard about modular smartphones, but Google has just released the Module Developers Kit (MDK) for Project Ara. The development kit gives an overview of the inner workings of the project, and provides templates for building your own modules.
Once you’ve agreed to the license agreement and downloaded the MDK, you’ll find a large specification document. It explains how a phone will comprise of many modules loaded into an endoskeleton, giving mechanical support and electrical connections. An interface block provides each module with power and data over LVDS. Modules are held in place by an electro-permanent magnet which can be toggled by software.
When you’re finished with the specification document, you can dive into the reference designs. These include templates and actual modules for WiFi, thermal imaging, a battery pack, and more. Mechanical CAD is provided as STEP files and drawings, and electrical design files are provided as Altium projects and PDF schematics.
We discussed both Project Ara and Phonebloks on Hackaday in the past, but now we’re starting to see real details. Google’s Project Ara Developer Conference takes place on April 15th and 16th, and you can register to take part remotely for free. Is this the start of an open, modular phone? Let us know what you think.
[Thanks to Adam for the tip]
[Yoshinok] recently posted an Instructable on doing a $10 smartphone-to-microscope conversion. The hack isn’t so much a conversion as just a handy jig, but it’s still interesting. The basic idea is to set up a platform for the slides, and to mount the smartphone directly above. The trick, and the reason this can be called a microscope, is that [Yoshinok] embeds the lens from a cheap laser pointer into the smartphone holder. He is able to get 40x optical magnification with the lens, and even though it sacrifices quality, he uses the built-in digital zoom to get up to 175x magnification.
By itself, you could use this with a light source to magnify 3D objects. [Yoshinok] demonstrates this with a dime. But since the slide holder is made of clear acrylic, he mounted a cheap LED flashlight in the base to serve as through-sample lighting. Using this setup, he was able to observe the process of plasmolysis.
If you have kids, this is certainly a project to do with them, but we can’t help but think it will be useful for non-parents alike. This sort of magnification is good enough for simple lab experiments, and given that most Hack-a-Day readers have these parts lying around, we figure the cost is closer to $0. If you give it a try, let us know your results in the comments!
Continue reading “Use Your Smartphone as a Microscope for Less Than $10”
[Matt Galisa] decided to try his hand at setting up the Belkin WeMo outlet without using a Smartphone app. The hardware is a pass-through for mains voltage which allows you to switch the plug over the network. It has a built-in WiFi module which normally connects to your home network. But the first time that you power it up it announces its own SSID designed for an iOS (and recently Android Beta) app to connect to in order to enter your AP credentials.
He started with this Python script used for WeMo hacking. It was originally meant to issue commands to the outlet once it had passed the initial setup. [Matt] followed along but couldn’t get an answer on the port he expected. It turns out that the device listens on a different port until the initial setup is complete (probably so that you don’t mess up other outlets on the network that are already working correctly). His next challenge was to manually set the WPA credentials. This never really worked and he ended up using a virtual AP without password protection through DD-WRT. From there he was able to set up a Python script to turn on, off, and toggle the state of the outlet.
If you’re looking to dig deeper into the device’s security check out this project.
Next time you’re waiting in the security line in an airport, why don’t you pull out your smartphone and count all the radiation being emitted by those body scanners and x-rays? There’s an app for that, courtesy of Mr. [Rolf-Dieter Klein].
The app works by blocking all the light coming into a phone’s camera sensor with a piece of tape or plastic. Because high energy radiation will cause artifacts on the CMOS camera sensor inside the phone, radiation will be captured as tiny specks of white light. The title picture for this post was taken from a camera phone at the Helmholtz Research Center in Munich being bathed in 10 Sieverts per hour of Gamma radiation from the decay of Cesium-137.
We have to note that blips of ‘bad data’ from a CMOS camera sensor aren’t unusual. These can come from electrical weirdness in the sensor itself or even the heat from the battery. [Rolf]’s app takes a reading of the noise floor and subtracts it from the counter. Radioactive decay resulting in Beta particles such as the Potassium-40 in bananas or the Uranium in granite counter tops don’t really register, although [Rolf] did have some success with Potassium chloride and a long measurement time. Still though, it’s a really cool way to turn a phone into a tricorder.
Continue reading “Turn your camera phone into a Geiger counter”