There are a few very different pathways to building a product, and we gotta applaud the developers taking care to take the open-source path. Today’s highlight is [Mentra], who is releasing an open-source smart glasses OS for their own and others’ devices, letting you develop your smart glasses ideas just once, a single codebase applicable for multiple models.
Currently, the compatibility list covers four models, two of them Mentra’s (Live and Mach 1), one from Vuzix (Z100), and one from Even Realities (G1) — some display-only, and some recording-only. The app store already has a few apps that cover the basics, the repository looks lively, and if the openness is anything to go by, our guess is that we’re sure to see more.
[NullPxl]’s Ban-Rays concept is a wearable that detects when one is in the presence of camera-bearing smartglasses, such as Meta’s line of Ray-Bans. A project in progress, it’s currently focused on how to reliably perform detection without resorting to using a camera itself. Right now, it plays a well-known audio cue whenever it gets a hit.
Once software is nailed down, the device aims to be small enough to fit into glasses.
Currently, [NullPxl] is exploring two main methods of detection. The first takes advantage of the fact that image sensors in cameras act as tiny reflectors for IR. That means camera-toting smartglasses have an identifying feature, which can be sensed and measured. You can see a sample such reflection in the header image, up above.
As mentioned, Ban-Rays eschews the idea of using a camera to perform this. [NullPxl] understandably feels that putting a camera on glasses in order to detect glasses with cameras doesn’t hold much water, conceptually.
The alternate approach is to project IR in a variety of wavelengths while sensing reflections with a photodiode. Initial tests show that scanning a pair of Meta smartglasses in this way does indeed look different from regular eyeglasses, but probably not enough to be conclusive on its own at the moment. That brings us to the second method being used: wireless activity.
Characterizing a device by its wireless activity turned out to be trickier than expected. At first, [NullPxl] aimed to simply watch for BLE (Bluetooth Low-Energy) advertisements coming from smartglasses, but these only seem to happen during pairing and power-up, and sometimes when the glasses are removed from the storage case. Clearly a bit more is going to be needed, but since these devices rely heavily on wireless communications there might yet be some way to actively query or otherwise characterize their activity.
This kind of project is something that is getting some interest. Here’s another smartglasses detector that seems to depend entirely on sniffing OUIs (Organizationally Unique Identifiers); an approach [NullPxl] suspects isn’t scalable due to address randomization in BLE. Clearly, a reliable approach is still in the works.
The increasing numbers of smartglasses raises questions about the impact of normalizing tech companies turning people into always-on recording devices. Of course, the average person is already being subtly recorded by a staggering number of hidden cameras. But at least it’s fairly obvious when an individual is recording you with a personal device like their phone. That may not be the case for much longer.
[SamsonMarch] designs electronic products by day and — apparently — does it in his spare time, too. His latest is a pair of really cool shades that give him turn-by-turn directions as he walks around town. Unlike some smart glasses, these get around the difficult problem of building a heads-up display by using a very simple interface based on colored LEDs visible to your peripheral vision in the temples of the frames.
The glasses themselves look great; designed in Fusion 360 and cut out of wood, no one would give them a second glance. [Sam] says you could 3D print them, too, but we think the wood looks best even if the stock is a cheap bamboo cutting board. He also cut the lenses out of acrylic.
The slots in the temples are where the action is, though. An iPhone app takes input and talks to Apple services to get directions. A lot of thought went into making the app work even though the phone keeps trying to put it to sleep. Each PCB hosts an RGB LED for indicating left/right turn and destination. They talk to the app using BLE and include accelerometers which put the boards — powered by coin cells — into sleep mode when no movement is detected.
Overall a fun and good looking project. There are even covers to hide the boards during normal use. The files you need to reproduce it are on GitHub. Usually, when we see smart glasses, they have some sort of screen which is harder to do. Of course, it is impossible to avoid comparisons to Google Glass.