Google Glass kind of came and went, leaving one significant addition to the English language. Even Google itself used the term “glasshole” for people who used the product in a creepy way. We can’t decide if wearing an obviously homemade set of glasses like the ones made by [Jordan Fung] are more creepy, give you more hacker cred, or just make you look like a Borg. Maybe some combination of all of those. While the cost and complexity of developing for Google Glass was certainly a barrier for hacking on that hardware, this project is just begging for you to build your own and run with the concept.
[Jordan’s] build, called Pedosa Glass, really is pretty respectable for a self-built set up. The Arduino Nano is a bit bulky, and the three push buttons take up some room, but it doesn’t kill the ability to mount them in a glasses form-factor. An FLCoS display lets you see the output of the software which [Jordan] is still developing. Right now features include a timer and a flashlight that uses the head-mounted white LED. Not much, we admit, but enough to prove out the hardware and the whole point would be to add software you wanted.
Admittedly, it isn’t exactly like Google Glass. Although both use FLCoS displays, Pedosa Glass uses a display meant for a camera viewfinder, so you don’t really see through it. Still, there might be some practical use for a little display mounted in your field of vision. The system will improve with a better CPU that is easier to connect to the network with sensors like an accelerometer — there’s plenty of room to iterate on this project. Then again, you do have an entire second ear piece to work with if you wanted to expand the system.
Check out the video demo after the break.
Continue reading “The Simplest Smart Glasses Concept”
Every day your eyeballs are assaulted by advertisements on your box of cereal, billboards, t-shirts, magazines, milk cartons, plastered on the side of buses, buildings, bananas, and written in the sky. [Reed], [Jonathan], [Tom], and [Alex] came up with a solution to this: a Brand Killer that censors all the advertisements and brands you see every minute of every day. It’s a real-world adblock that you can build right now.
The team’s system uses a custom head mounted display made from cardboard, goggles, a webcam, and a seven-inch display. The software for the system uses Python and OpenCV to monitor the images from the webcam, compares them against a list of brands and logos, and filters them out with an unobtrusive blur.
Right now the system just has a few brands and logos that include Dr. Pepper, Hershey’s, McDonalds, Facebook, Starbucks, and clear evidence this was built at UPenn, Wawa and Tastykake. In the video below, the detection and tracking of these various brands is very good. The system is also stereoscopic, meaning this is wearable all day, every day, without a loss of depth perception.
Continue reading “Real World AdBlock”
Ever since [will1384] watched “The Lawnmower Man” as a wee lad, he’s been interested in virtual reality. He has been messing around with it for years and even had a VictorMaxx Stuntmaster, one of the first available head mounted displays. Years later, the Oculus Rift came out and [will1384] wanted to try it out but the $350 price tag put it just out of his price range for a discretionary purchase. He then did what most of us HaD readers would do, try building one himself, and with a goal for doing it for around $100.
The main display is a 7″ LCD with a resolution of 1024×600 pixels and has a mini HDMI input. Some DIY head mounted display projects out on the ‘web use ski goggles or some sort of elastic strap to hold the display to the wearer’s head. [will1384] took a more industrial approach, literally. He used the head mounting system from a welding helmet. This not only has an adjustable band but also has a top strap to prevent the entire contraption from sliding down. Three-dimensional parts were printed out to secure the LCD to the welding helmet parts while at the same time creating a duct to block out external light.
Inside the goggles are a pair of 5x Loupe lenses mounted between the user’s eyes and the LCD screen. These were made to be adjustable so that the wearer can dial them in for the most comfortable viewing experience. The remote mounted to the top strap may look a little out-of-place but it is actually being used to capture head movement. In addition to a standard wireless remote, it is also an air mouse with internal gyroscopes.
For [Tony]’s entry for The Hackaday Prize, he’s doing something we’ve all seen before – a head mounted display, connected to a Bluetooth module, displaying information from a smartphone. What we haven’t seen before is a cheap version of this tech, and a version of Google Glass that folds – you know, like every other pair of glasses on the planet – edges this project over from ‘interesting’ to ‘nearly practical’.
For the display, [Tony] is using a 0.96″ OLED connected to an Arduino Nano. This screen is directed into the wearer’s eye with a series of optics that, along with every other part of the frame, was 3D printed on a Solidoodle 2. The frame itself not only folds along the temples, but also along the bridge, making this HMD surprisingly compact when folded up.
Everything displayed on this head mounted display is controlled by either an Android phone or a Bluetooth connection to a desktop. Using relatively simple display means [Tony] is limited to text and extremely simple graphics, but this is more than enough for some very interesting applications; reading SMS messages and checking email is easy, and doesn’t overpower the ‘duino.
The project featured in this post is an entry in The Hackaday Prize. Build something awesome and win a trip to space or hundreds of other prizes.
[Arvind] has dropped his hat in the game of head mounted displays. With Google Glass pushing $1,500, it’s only natural for hackers to make a cheaper alternative. [Avind’s] $80 version might not be pretty, but it gets the job done.
Using a Raspberry Pi loaded with speech recognition software, a webcam, 2.5 inch LCD display and a handful of other parts, [Arvind’s] hat mounted display allows him to view email, Google Maps, videos or just about anything he wants.
An aspheric loupe magnifier lens lets him see the display even though it sits around 5cm from his eye. No outside light is allowed in. Only the guts of the webcam were used to give him the video and microphone. We’ve seen other head mounted displays before, and this one adds to the growing collection. Be sure to check out [Arvinds] site for a tutorial on how to build your own, and catch a video of it in action after the break.
Continue reading “Smart Hat Puts Your Head in the Game”
[John Ohno] has found what is perhaps the best possible use for steampunk goggles: framing a monocular display for a Raspberry Pi-based wearable computer. [John]’s eventual goal for the computer is a zzstructure-based personal organizer and general notifier. We covered [John]’s zzstructure emulator to our great delight in July 2011. Go ahead and check that out, because it’s awesome. We’ll wait here.
[John] has been interested in wearable computing for some time, but is unimpressed with Google Glass. He had read up on turning head-mounted displays into monocular devices and recognized a great opportunity when his friend gave him most of an Adafruit display. With some steampunk goggles he’d bought at an anime convention, he started on the path to becoming a Gargoyle. He encountered a few problems along the way, namely SD card fail, display output issues, and general keep-the-parts-together stuff, but came out smelling like a rose. [John] has ideas for future input additions such as simple infrared eye tracking, the addition of a chording keyboard, and implementing a motorized glove for haptic learning.
Want to make your own wearable display but have an aversion to steampunk? Check out this homebrew solution with (mostly) 3-D printed frames. And it has servos!
[William Steptoe] is a post-doctoral research associate at University College London. This means he gets to play with some really cool hardware. His most recent project is an augmented reality update to the Oculus Rift. This is much more than hacking a pair of cameras on the Rift though. [William] has created an entire AR/VR user interface, complete with dockable web browser screens. He started with a stock Rift, and a room decked out with a professional motion capture system. The Rift was made wireless with the addition of an ASUS Wavi and a laptop battery system. [William] found that the wireless link added no appreciable latency to the Rift. To move into the realm of augmented reality, [William] added a pair of Logitech C310 cameras. The C310 lens’ field of view was a bit narrow for what he needed, so lenses from a Genius WideCam F100 were swapped in. The Logitech cameras were stripped down to the board level, and mounted on 3D printed brackets that clip onto the Rift’s display. Shapelock was added to the mounts to allow the convergence of the cameras to be easily set.
Stereo camera calibration is a difficult and processor intensive process. Add to that multiple tracking systems (both the 6DOF head tracking on the Rift, and the video tracker built-in to the room) and you’ve got quite a difficult computational process. [William] found that he needed to use a Unity shader running on his PC’s graphics card to get the system to operate in real-time. The results are quite stunning. We didn’t have a Rift handy to view the 3D portions of [William’s] video. However, the sense of presence in the room still showed through. Videos like this make us excited for the future of augmented reality applications, with the Rift, the upcoming castAR, and with other systems.
Continue reading “Oculus Rift Goes from Virtual to Augmented Reality”