You shouldn’t be looking at screens when you’re driving, but what about a heads-up display? A screen that could put relevant information in your field of vision would be great, even more so if it used a Raspberry Pi. That’s exactly what [John] did, only he did it with an airplane.
First up, the legality of this build. [John]’s plane is registered as experimental, which, provided you know what you’re doing, is pretty close to ‘anything goes’ as you would want in a manned aircraft. [John] has a sufficient number of hours in his log book, and he’s built a Zenith 701.
For hardware, the hard part of this build is constructing a heads-up display. Fortunately, aftermarket HUDs exist, and [John] is using a Kivic projector, a $200 piece of equipment that’s readily available on Amazon. If you need a HUD for your car, there you go. The software is another thing entirely, with the goal of having the software decoupled from the display and data sources. This is somewhat easy to accomplish with a Raspberry Pi; the display is actually just some minimal text-based blocky graphics built in PyGame. This build is also decoupled from the data sources by building this as a user interface for Stratux, an independent Raspberry Pi-based ADS-B receiver for pilots.
There are several views available with this HUD, with the AHRS + ADS-B providing information on the aircraft’s attitude and altitude, along with a few indicators of the nearest planes. The traffic view expands on the ADS-B data, showing the nearest eight or so aircraft in the air, with a range, bearing, and difference in altitude. There’s a diagnostic window, and since [John]’s plane is a backcountry STOL thingamado that can hover in a strong wind, there’s also a digital version of a norden bombsight. It’s for dropping bags of flour onto a grass strip. You can check out [John]’s entire AirVenture presentation of the build below, with all the code available here.
Continue reading “Python and Pi Provide Heads Up Display For Your Experimental Airplane”
[Ricardo Ferro] didn’t want to buy a Google Glass, so he made his own.
The Raspberry Pi Zero Prism consists of a 3D-printed headset the side-pieces of which hold a variety of electronic components, including a Pi Zero running Raspbian Jessie, a Pi Noir IR camera, a WiFi/Bluetooth module and a whole mess of SMD tactile push buttons. Video output is provided by a Kopin 922K display module. This module is usually used in smart goggles and uses a prism to reflect information into the wearer’s field of view.
One application [Ricardo] envisions for this Open Source Google Glass is using it in conjunction with facial recognition software and the YouTube-favorite IR camera trick of seeing through clothing. No, he’s not using it for that idea, and you should get your mind out of the gutter. [Ricardo] wants to identify masked criminals. Setting aside the technological challenges of making that technology work, we think that walking around with x-ray specs is likely to get those specs broken off your face by someone who wears clothes for modesty purposes. Still, it’s a fascinating project and we love the way the prism and video assembly comes together.
While most of us will never set foot in a fighter jet, some of us can still try to get as close as possible. One of the most eye-catching features of a fighter jet (at least from the pilot’s point-of-view) is the heads-up display, so that’s exactly what [Frank] decided to build into his car to give it that touch of fighter jet style.
Heads-up displays use the small reflectivity of a transparent surface to work. In this case, [Frank] uses an LED strip placed on the dashboard to shine up into the windshield. A small amount of light is reflected back to the driver which is able to communicate vehicle statues without obscuring view of the road. [Frank]’s system is able to display information reported over the CAN bus, including voltage, engine RPM, and speed.
This display seems to account for all the issues we could think up. It automatically cycles through modes depending on driving style (revving the engine at a stoplight switches it to engine RPM mode, for example), the LEDs automatically dim at night to avoid blinding the driver, and it interfaces with the CAN bus which means the ability to display any other information in the future should be relatively straightforward. [Frank] does note some rough edges, though, namely with the power supply and the fact that there’s a large amount of data on the CAN bus that the Teensy microcontroller has a hard time sorting out.
That being said, the build is well polished and definitely adds a fighter jet quality to the car. And if [Frank] ever wants even more aviation cred for his ground transportation, he should be able to make use of a 747 controller for something on the dashboard, too.
As the Jerusalem mini Makerfaire approached, [Avishay] had to come up with something to build. His final project is something he calls ASTROGUN. The ASTROGUN is a sort of augmented reality game that has the player attempting to blast quickly approaching asteroids before being hit.
It’s definitely reminiscent of the arcade classic, Asteroids. The primary difference is that the player has no space ship and does not move through space. Instead, the player has a first person view and can rotate 360 degrees and look up and down. The radar screen in the corner will give you a rough idea of where the asteroids are coming from. Then it’s up to you to actually locate them and blast them into oblivion before they destroy you.
The game is built around a Raspberry Pi computer. This acts as the brains of the operation. The Pi interfaces with an MPU-9150 inertial measurement unit (IMU). You commonly see IMU’s used in drones to help them keep their orientation. In this case, [Avishay] is using it to track the motion and orientation of the blaster. He claims nine degrees of freedom with this setup.
The Pi generates the graphics and sends the output to a small, high-brightness LCD screen. The screen is mounted perpendicular to the player’s view so the screen is facing “up”. There is a small piece of beam splitting glass mounted above the display at approximately a 45 degree angle. This is a special kind of glass that is partially reflective and partially translucent. The result is that the player sees the real-world background coming through the glass, with the digital graphics overlaid on top of that. It’s similar to some heads-up display technologies.
All of the electronics fit either inside or mounted around a toy gun. The display system was attached with a custom-made fiberglass mount. The code appears to be available via Github. Be sure to watch the video of the system in action below. Continue reading “ASTROGUN is like Asteroids on Steroids”
This edition of Fail of the Week is nothing short of remarkable, and your help could really get the failed project back on track. [Snipor Bob] wanted to replace all of the dashboard readouts on his Mustang and got the idea of making the hacked hardware into a Heads-Up Display. What you see above is simply the early hardware proof of concept for tapping into the vehicle’s data system. But there’s also an interesting test rig for getting the windshield glass working as a reflector for the readout.
Continue reading “Fail of the Week: CAN-Bus Attached HUD for Ford Mustang”
[Cmonaco3’s] girlfriend wanted a better way to control her iPod when driving. She didn’t want to take her eyes of the road and asked him if he could help. He ended up building a heads up display which reads out track information and offers a few simple buttons for control.
The display includes controls for track forward, track back, and play/pause. Those buttons, along with the LCD screen, mount on the windshield using a suction cup. This way the driver doesn’t have to completely remove focus from the road to control the iPod which is sitting in the passenger’s seat.
To accomplish this [Cmonaco] used a dock connector breakout board for communication between an Arduino and the iPod. The Arduino pulls song information to be displayed on the graphic LCD screen, and sends commands to the iPod when it detects a button push. See a quick demo of the setup after the break.
Continue reading “Heads up controls for your iPod”
[Matt Kwan] says that coming up with a personal heads-up display wasn’t that hard. Well that’s because he made design choices that make all the difference.
The goal here was to add some augmented reality to his field of vision. He went with a baseball cap because it’s a pretty easy way to strap something to your head. You can’t see it from this angle, but the setup requires you to cut a rather large hole in brim. The image from a smartphone (HTC Desire Z in this case) which is situated with the screen pointing toward [Matt’s] forehead. The screen reflects off of a small mirror, guiding the image down through a Fresnel lens mounted in the hole of the brim. The image is reflected a second time by the plastic in front of his eyes which is coated with a slightly mirrored material. Since the image is reflected twice it appears right-side up, and the use of the Fresnel lens places the image out about 20 cm in front of his view. He tried to get some images of the effect, but we think you’ve got to see it in person before passing judgement.
This does away with the need to track head movement (there’s a few hacks for that out there though). Augmented reality software is used to turn the view from the smartphone camera into overlay data for the display.