It may not change the world, but [Tyler]’s DIY smartphone is a great example of what you can do with off-the-shelf parts. He built a complete, working cell phone using a Raspberry Pi, a few parts from Adafruit, and a 3D printed enclosure.
Inside the Tyfone is a Raspberry Pi Model A, an Adafruit FONA cellular module, a PiTFT, and not much else. There’s a 1200 mAh battery in there, and a 3D printed case keeps everything together.
For the OS, [Tyler] isn’t running Android; that’s only for the Raspi 2, and the Raspberry Pi 2 Model A isn’t out yet. Instead, [Tyler] wrote his own not-OS in Python. It can send and receive SMS messages, make calls, take pictures, connect to WiFi networks, and do just about everything else a Nokia from 2003 can do.
[Tyler] put together a video going over all of the features of his Tyfone. You can check that out below.
Continue reading “Hackaday Prize Entry: A DIY Smartphone”
[Christian Holz, Senaka Buthpitiya, and Marius Knaust] are researchers at Yahoo that have created a biometric solution for those unlucky folks that always forget their smartphone PIN codes. Bodyprint is an authentication system that allows a variety of body parts to act as the password. These range from ears to fists.
Bodyprint uses the phone’s touchscreen as an image scanner. In order to do so, the researchers rooted an LG Nexus 5 and modified the touchscreen module. When a user sets up Bodyprint, they hold the desired body part to the touchscreen. A series of images are taken, sorted into various intensity categories. These files are stored in a database that identifies them by body type and associates the user authentication with them. When the user wants to access their phone, they simply hold that body part on the touchscreen, and Bodyprint will do the rest. There is an interesting security option: the two person authentication process. In the example shown in the video below, two users can restrict file access on a phone. Both users must be present to unlock the files on the phone.
How does Bodyprint compare to capacitive fingerprint scanners? These scanners are available on the more expensive phone models, as they require a higher touchscreen resolution and quality sensor. Bodyprint makes do with a much lower resolution of approximately 6dpi while increasing the false rejection rate to help compensate. In a 12 participant study using the ears to authenticate, accuracy was over 99% with a false rejection rate of 1 out of 13.
Continue reading “Your Body is Your PIN with Bodyprint”
For [Tyler]’s entry to the Hackaday Prize, he’s making something that just a few years ago would be unheard of in a homebrew build. He’s making a DIY smartphone. Yes, with cheap single-board Linux computers, GSM modules, and SPI touchscreen displays, it’s possible to build your own smartphone.
Inside [Tyler]’s DIY smartphone is a Raspberry Pi Model A, a 3.5 inch touchscreen PiTFT with 480×320 resolution, and an Adafruit FONA module The connections are simple enough; the TFT is connected over SPI, and the GSM module over serial. The entire device is powered by a 1200mAh LiIon battery, charged with a powerboost board, runs an operating system written in Python capable of making calls, sending texts, and takes pictures with a Pi camera.
This is not what you would normally call a smartphone. The FONA module is 2G only, meaning you’re limited to 2G speeds and 2G networks. AT&T will be shutting down 2G networks in a little bit, although T-Mobile will be keeping them up for anyone who still has an old Nokia Brick.
That said, [Tyler]’s phone is still exactly what you want in a minimal phone: it just makes calls and receives texts, it has a camera, and unlike the Nokia, you can take it apart and repair it easily. Not that you ever had to do that with a Nokia…
Although scientists have known about Ultra-High Energy Cosmic Rays (UHECRs) for years, nobody can pinpoint their origin. When these UHECRs hit the ground, however, they cause a widespread local disturbance called an air shower. This air shower is a wide dispersion of photons, muons, and electrons at sea level. The means of observing this air shower mandates a widespread geographic region for detecting them. One solution would be a very big detector. Physicists [Daniel] and [Michael] discovered an alternative to pricey hardware, though. By leveraging the CMOS sensors in our smartphones, they can borrow some CPU cycles on our phones to create a worldwide detector network.
According to their paper, the CMOS camera in our smartphones is sensitive to the spectrum of radiation induced by muons and photons from these air showers. With an app running on our phones, [Daniel], [Michael], and other scientists can aggregate the data from multiple detections in a similar region to better understand their origins.
If you’re concerned about CRAYFIS taking away from your talk or web-browsing time, fear not; it runs in the background when a power source has been detected, hopefully, when you are asleep. It’s not the first time we see scientists tap into our computing resources, but this is certainly an achievement made possible in only the last few years by the sensor-loaded smartphone that charges on many of our night stands. With over 1.5 billion smartphones active in the world, we’re thrilled to see a team cleverly leveraging a ubiquitous and already-well-distributed resource.
The Oculus Rift and all the other 3D video goggle solutions out there are great if you want to explore virtual worlds with stereoscopic vision, but until now we haven’t seen anyone exploring real life with digital stereoscopic viewers. [pabr] combined the Kinect-like sensor in an ASUS Xtion with a smartphone in a Google Cardboard-like setup for 3D views the human eye can’t naturally experience like a third-person view, a radar-like display, and seeing what the world would look like with your eyes 20 inches apart.
[pabr] is using an ASUS Xtion depth sensor connected to a Galaxy SIII via the USB OTG port. With a little bit of code, the output from the depth sensor can be pushed to the phone’s display. The hardware setup consists of a VR-Spective, a rather expensive bit of plastic, but with the right mechanical considerations, a piece of cardboard or some foam board and hot glue would do quite nicely.
[pabr] put together a video demo of his build, along with a few examples of what this project can do. It’s rather odd, and surprisingly not a superfluous way to see in 3D. You can check out that video below.
Continue reading “Seeing The World Through Depth Sensing Cameras”
[Daniel Whiteson and Michael Mulhearn], researchers at the University of California, have come up with a novel method of detecting ultra-high energy cosmic rays (UHECR) using smartphones. UHECR are defined as having energy greater than 1018eV. They are rare and very difficult to detect with current arrays. In order to examine enough air showers to detect UHECR, more surface area is needed. Current arrays, like the Pierre Auger Observatory and AGASA, cannot get much larger without dramatically increasing cost. A similar THP Quarterfinalist project is the construction of a low-cost cosmic ray observatory, where it was mentioned that more detection area is needed in order to obtain enough data to be useful.
[Daniel Whiteson and Michael Mulhearn] and colleagues noted that smartphone cameras with CMOS sensors can detect ionizing radiation, which means they also will pick up muons and high-energy photons from cosmic rays. The ubiquitous presence of smartphones makes their collective detection of air showers and UHECR an intriguing possibility. To make all this happen, [Whiteson and Mulhearn] created a smartphone app called CRAYFIS, short for Cosmic RAYs Found In Smartphones. The app turns an idle smartphone into a cosmic ray detector. When the screen goes to sleep and the camera is face-down, CRAYFIS starts taking data from the camera. If a cosmic ray hits the CMOS sensor, the image data is stored on the smartphone along with the arrival time and the phone’s geolocation. This information is uploaded to a central server via the phone’s WiFi. The user does not have to interact with the app beyond installing it. It’s worth noting that CRAYFIS will only capture when the phone is plugged in, so no worries about dead batteries.
The goal of CRAYFIS is to have a minimum of one million smartphones running the app, with a density of 1000 smartphones per square kilometer. As an incentive, anyone whose smartphone data is used in a future scientific paper will be listed as an author. There are CRAYFIS app versions for Android and iOS platforms according to the site. CRAYFIS is still in beta, so the apps aren’t publicly available. Head over to the site to join up!
It’s been a little while since we’ve heard about modular smartphones, but Google has just released the Module Developers Kit (MDK) for Project Ara. The development kit gives an overview of the inner workings of the project, and provides templates for building your own modules.
Once you’ve agreed to the license agreement and downloaded the MDK, you’ll find a large specification document. It explains how a phone will comprise of many modules loaded into an endoskeleton, giving mechanical support and electrical connections. An interface block provides each module with power and data over LVDS. Modules are held in place by an electro-permanent magnet which can be toggled by software.
When you’re finished with the specification document, you can dive into the reference designs. These include templates and actual modules for WiFi, thermal imaging, a battery pack, and more. Mechanical CAD is provided as STEP files and drawings, and electrical design files are provided as Altium projects and PDF schematics.
We discussed both Project Ara and Phonebloks on Hackaday in the past, but now we’re starting to see real details. Google’s Project Ara Developer Conference takes place on April 15th and 16th, and you can register to take part remotely for free. Is this the start of an open, modular phone? Let us know what you think.
[Thanks to Adam for the tip]