Many cars these days come with a basic Heads Up Display, or HUD. Typically, these display speed, though some also throw in a tachometer or navigational graphics too. Of course, if your car doesn’t have one of these stock, hacking in your own is always an option.
[PowerBroker2] developed this HUD in a somewhat circuitous way, but it’s effective nonetheless. An ELM327 Bluetooth OBD-II reader is hooked up to the car, collecting data on speed and RPM. This data is passed to an ESP-32 and Teensy 3.5. From reading the code, it appears the Teensy is responsible for logging data from the CAN bus on an SD card, and running a small OLED display. The ESP32 is then charged with running the LED display that actually forms the HUD. It’s then combined with a 3D-printed housing, some plexiglass, and reflective windshield film to complete the effect.
It’s a build that probably packs in more hardware than is strictly needed to get the job done, but it does indeed get the job done. Other builds we’ve seen use LED strips as a quick and tidy way to get the job done. Video after the break. Continue reading “Arduino Car HUD Does The Job”
For most of human history, the way to get custom shapes and colors onto one’s retinas was to draw it on a cave wall, or a piece of parchment, or on paper. Later on, we invented electronic displays and used them for everything from televisions to computers, even toying with displays that gave the illusion of a 3D shape existing in front of us. Yet what if one could just skip this surface and draw directly onto our retinas?
Admittedly, the thought of aiming lasers directly at the layer of cells at the back of our eyeballs — the delicate organs which allow us to see — likely does not give one the same response as you’d have when thinking of sitting in front of a 4K, 27″ gaming display to look at the same content. Yet effectively we’d have the same photons painting the same image on our retinas. And what if it could be an 8K display, cinema-sized. Or maybe have a HUD overlay instead, like in video games?
In many ways, this concept of virtual retinal displays as they are called is almost too much like science-fiction, and yet it’s been the subject of decades of research, with increasingly more sophisticated technologies making it closer to an every day reality. Will we be ditching our displays and TVs for this technology any time soon?
Continue reading “The Smallest Large Display Is Projected Straight Onto Your Retina”
You shouldn’t be looking at screens when you’re driving, but what about a heads-up display? A screen that could put relevant information in your field of vision would be great, even more so if it used a Raspberry Pi. That’s exactly what [John] did, only he did it with an airplane.
First up, the legality of this build. [John]’s plane is registered as experimental, which, provided you know what you’re doing, is pretty close to ‘anything goes’ as you would want in a manned aircraft. [John] has a sufficient number of hours in his log book, and he’s built a Zenith 701.
For hardware, the hard part of this build is constructing a heads-up display. Fortunately, aftermarket HUDs exist, and [John] is using a Kivic projector, a $200 piece of equipment that’s readily available on Amazon. If you need a HUD for your car, there you go. The software is another thing entirely, with the goal of having the software decoupled from the display and data sources. This is somewhat easy to accomplish with a Raspberry Pi; the display is actually just some minimal text-based blocky graphics built in PyGame. This build is also decoupled from the data sources by building this as a user interface for Stratux, an independent Raspberry Pi-based ADS-B receiver for pilots.
There are several views available with this HUD, with the AHRS + ADS-B providing information on the aircraft’s attitude and altitude, along with a few indicators of the nearest planes. The traffic view expands on the ADS-B data, showing the nearest eight or so aircraft in the air, with a range, bearing, and difference in altitude. There’s a diagnostic window, and since [John]’s plane is a backcountry STOL thingamado that can hover in a strong wind, there’s also a digital version of a norden bombsight. It’s for dropping bags of flour onto a grass strip. You can check out [John]’s entire AirVenture presentation of the build below, with all the code available here.
Continue reading “Python And Pi Provide Heads Up Display For Your Experimental Airplane”
At some point, a child will inevitably dream of being a superhero. Not all children get the chance to see that dream made manifest, but a few take that destiny into their own hands. Redditor [Lord_of_Bone] — seizing at that goal — has built himself an Iron Man mask with an integrated HUD!
Relying on a conceptually similar project he’d previously built, much of the code was rehashed for this ‘Mark II’ version. Pieces of a smartphone holo pyramid act as projection surfaces — using a lens to focus the image to be viewed at such close distances — and a pair of OLED screens displaying the information. It’s a happy bonus that the lack of backlight results in only the text showing in the user’s field of view.
Instead of speaking with J.A.R.V.I.S., [Lord_of_Bone] is using a Raspberry Pi Zero W as the mask’s brain. Working past some I2C troubles between the OLED screens and an Enviro pHat required a whipped-up veroboard and a bit of hardware hacking. Cramming everything into the mask was no easy task — using Blutack and Sugru to bind them in the limited space — but the pHat had to be surface-mounted in the open anyways for atmospheric and light data.
Continue reading “Iron Man Mask With A HUD!”
There’s a harsh truth underlying all robotic research: compared to evolution, we suck at making things move. Nature has a couple billion years of practice making things that can slide, hop, fly, swim and run, so why not leverage those platforms? That’s the idea behind this turtle with a navigation robot strapped to its back.
This reminds us somewhat of an alternative universe sci-fi story by S.M. Stirling called The Sky People. In the story, Venus is teeming with dinosaurs that Terran colonists use as beasts of burden with brain implants that stimulate pleasure centers to control them. While the team led by [Phill Seung-Lee] at the Korean Advanced Institute of Science and Technology isn’t likely to get as much work from the red-eared slider turtle as the colonists in the story got from their bionic dinosaurs, there’s still plenty to learn from a setup like this. Using what amounts to a head-up display for the turtle in the form of a strip of LEDs, along with a food dispenser for positive reinforcement, the bionic terrapin is trained to associate food with the flashing LEDs. The LEDs are then used as cues as the turtle navigates between waypoints in a tank. Sadly, the full article is behind a paywall, but the video below gives you a taste of the gripping action.
Looking for something between amphibian and fictional dinosaurs to play mind games with? Why not make your best friend bionic? Continue reading “Head-Up Display Augments Bionic Turtle’s Reality”
Adventure travel can be pretty grueling, what with the exotic locations and potential for disaster that the typical tourist destinations don’t offer. One might find oneself dangling over a cliff for that near-death-experience selfie or ziplining through a rainforest canopy. All this is significantly complicated by being blind, of course, so a tool like this Raspberry Pi low-vision system would be a welcome addition to the nearly-blind adventurer’s well-worn rucksack.
[Dan] has had vision problems since childhood, but one look at his YouTube channel shows that he doesn’t let that slow him down. When [Dan] met [Ben] in Scotland, [Ben] noticed that he was using his smartphone as a vision aid, looking at the display up close and zooming in to get as much detail as possible from his remaining vision. [Ben] thought he could help, so he whipped up a heads-up display from a Raspberry Pi and a Pi Camera. Mounted to a 3D-printed frame holding a 5″ HDMI display and worn from a GoPro head mount, the camera provides enough detail to help [Dan] navigate, as seen in the video below.
The rig is a bit unwieldy right now, but as proof of concept (and proof of friendship), it’s a solid start. We think a slimmer profile design might help, in which case [Ben] might want to look into this Google Glass-like display for a multimeter for inspiration on version 2.0.
Continue reading “The ‘All-Seeing Pi’ Aids Low-Vision Adventurer”
If you haven’t been paying attention, big wheel trikes are a thing. There are motor driven versions as well as OG pedal pushing types . [Flux Axiom] is of the OG (you only get one link, now its on you) flavor and has written an instructable that shows how to achieve some nice looking on screen data that he syncs up with the video for a professional looking finished product which you can see in the video after the break.
[Flux Axiom] is using an Arduino Mega in his setup along with a cornucopia of sensors and all their data is being logged onto an SD card. All the code used in his setup is available in his GitHub repository. [Flux Axiom] was also nice enough to include the calibration process he used for the sensors which is also located in the GitHub download.
Sadly [Flux Axiom] uses freedom hating software for combining the video and data, Race Render 3 is his current solution and he is pleased with the results. Leave it in the comments if you have an open source solution for combining the video and data that we can offer him as a replacement.
Edit: Correct spelling of handle.
Continue reading “Video With Sensor Data Overlay Via Arduino Mega”