In the next few years, VR headsets will be everywhere, and everyone will slowly recede into their own little reality that is presented on high-resolution displays right in front of their eyes. One specific group will be left out: eyeglass wearers. VR just doesn’t work with eyeglasses, and a few people in Germany are fixing this problem. They’re creating custom prescription lenses for Google Cardboard, giving anyone with glasses the opportunity to look just a little more hipster.
The folks behind this Indiegogo already run a specialty optics shop in Germany. They have the tools to make custom lenses for spectacles, and they’re the first company so far that has identified a problem with the current crop of VR headsets and has created a solution. The campaign is for a set of lenses that can be attached to Google Cardboard with double stick tape. There are limitations on how strong of a prescription they can make, but it should work for most four eyes.
It should be noted this Indiegogo isn’t the only way to get custom lenses for a VR headset. If you have your prescription, there are a few places to buy glasses online for $30 or so. Do that, remove the lenses from the frame, and affix them to Cardboard.
[Alan Yates] brought a demo of Valve’s new VR tech that’s the basis of the HTC Vive system to Maker Faire this year. It’s exceptionally clever, and compared to existing VR headsets it’s probably one of the best headtracking solutions out there.
With VR headsets, the problem isn’t putting two displays in front of the user’s eyes. The problem is determining where the user is looking quickly and accurately. IMUs and image processing techniques can be used with varying degrees of success, but to do it right, it needs to be really fast and really cheap.
[Alan] and [Valve]’s ‘Lighthouse’ tracking unit does this by placing a dozen or so IR photodiodes on the headset itself. On the tracking base station, IR lasers scan in the X and Y axes. By scanning these IR lasers across the VR headset, the angle of the headset to the base station can be computed in just a few cycles of a microcontroller. For a bunch of one cent photodiodes, absolute angles and the orientation to a base station can be determined very easily, something that has some pretty incredible applications for everything from VR to robotics.
Remember all of the position tracking hacks that came out as a result of the Nintendo Wii using IR beacons and a tracking camera? This seems like an evolutionary leap forward but in the same realm and can’t wait to see people hacking on this tech!
What would happen if Oculus-quality virtual reality was created in the 80s on the Commodore PET? [Michael Hill] knows, because he created a stereoscopic video headset using a PET.
This build is an extension of [Michael]’s exhibit last year at VCF East where he displayed a video feed with PETSCII. Yes, that means displaying video with characters, not pixels.
This year, he’s doubling the number of screens, and sending everything to two iPhones in a Google Cardboard-like VR headset. Apart from the optics, the setup is pretty simple: cameras get image data, it’s sent over to a PET, and a stream of characters are sent back.
It’s impossible to film, and using it is interesting, to say the least. Video below.
Continue reading “VCF East X: Virtual Reality With PETSCII”
Virtual reality has come a long way but some senses are still neglected. Until Smell-O-Vision happens, the next step might be feeling the wind in your hair. Perhaps dad racing a sportbike or kids giggling on a rollercoaster. Not as hard to build as you might think, you probably have the parts already.
Off-the-shelf devices serve up the seeing and hearing part of your imaginary environment, but they stop there. [Jared] wanted to take the immersion farther by being able to feel the speed, which meant building his own high power wind generator and tying it into the VR system. The failed crowdfunding effort of the “Petal” meant that something new would have to be constructed. Obviously, to move air without actually going on a rollercoaster requires a motor controller and some fans. Powerful fans.
A proponent of going big or going home, [Jared] picked up a pair of fans and modified them so heavily that they will launch themselves off of the table if not anchored down. Who overdrives fans so hard they need custom heatsinks for the motors? He does. He admits he went overboard and sensibly way overbudget for most people but he built it for himself and does not care.
Continue reading ““Superfan” Gaming Peripheral Lets You Feel Your Speed”
Many of us have gone on a stationary romp through some virtual or augmented scape with one of the few headsets out in the wild today. While the experience of viewing a convincing figment of reality is an exciting sensation in itself, [Mark Lee] and [Kevin Wang] are figuring out how to tie other senses into the mix.
The duo from Cornell University have built a mechanical exoskeleton that responds to light with haptic feedback. This means the wearer can touch the sphere of light around a source as if it were a solid object. Photo resistors are mounted like antenna to the tip of each finger, which they filed down around the edges to receive a more diffused amount of light. When the wearer of the apparatus moves their hand towards a light source, the sensors trigger servo motors mounted on the back of the hand to actuate and retract a series of 3D printed tendons which arch upward and connect to the individual fingers of the wearer. This way as the resistors receive varying amounts of light, they can react independently to simulate physical contours.
One of the goals of the project was to produce a working proof of concept with no more than 100 dollars worth of materials, which [Mark] and [Kevin] achieve with some cash to spare. Their list of parts can be found on their blog along with some more details on the project.
Continue reading “Touching Light with Haptic Feedback”
At the Atmel booth at Maker Faire, they were showing off a few very cool bits and baubles. We’ve got a post on the WiFi shield in the works, but the most impressive person at the booth was [Quin]. He has a company, he’s already shipping products, and he has a few projects already in the works. What were you doing at 13?
[Quin]’s Qduino Mini is your basic Arduino compatible board with a LiPo charging circuit. There’s also a ‘fuel gauge’ of sorts for the battery. The project will be hitting Kickstarter sometime next month, and we’ll probably put that up in a links post.
Oh, [Quin] was also rocking some awesome kicks at the Faire. Atmel, I’m trying to give you money for these shoes, but you’re not taking it.
[Sophie] had a really cool installation at the faire, and notably something that was first featured on hackaday.io. Basically, it’s a virtually reality Segway, built with an Oculus, Leap Motion, a Wobbleboard, an Android that allows you to cruise on everyone’s favorite barely-cool balancing scooter through a virtual landscape.
This project was a collaboration between [Sophie], [Takafumi Ide], [Adelle Lin], and [Martha Hipley]. The virtual landscape was built in Unity, displayed on the Oculus, controlled with an accelerometer on a phone, and has input with a Leap Motion. There are destructible and interactable things in the environment that can be pushed around with the Leap Motion, and with the helmet-mounted fans, you can feel the wind in your hair as you cruise over the landscape on your hovering Segway-like vehicle. This is really one of the best VR projects we’ve ever seen.
After a hard Saturday at World Maker Faire, some of the best and brightest in the Hacker/Maker community descended on The Holiday Inn for “Bring A Hack”. Created by [Jeri Ellsworth] several years ago at the Bay Area Maker Faire, Bring A Hack (BAH) is an informal gathering. Sometimes a dinner, sometimes a group getting together at a local bar, BAH is has just one rule: You have to bring a hack!
[Sophi Kravitz] has become the unofficial event organizer for BAH in New York. This year she did a bit of live hacking, as she converted her Wobble Wonder headgear from wired to wireless control.
[Chris Gammell] brought his original Bench BudEE from Contextual Electronics. He showed off a few of his board customizations, including making a TSSOP part fit on the wrong footprint.
[Windell and Lenore] from Evil Mad Scientist Laboratories brought a few hacks along. They picked up an old Radio Shack music player chip at the Electronics Flea Market and built it up on a breadboard. Also on display was their new EggBot Pro. The Pro is a beautifully machined version of the eggbot. Everything is built strong to withstand the sort of duty an EggBot would see at a hackerspace or public library. [Windell] was full of surprises, as he also gave everyone chunks of Sal Ammoniac, which is a great way to bring the tin back to a tired soldering iron tip. The hack was that he found his Sal Ammoniac at a local Indian grocery in the Bay Area. Check out [Windell’s] blog entry for more information.
[Cal Howard] brought his DIY VR goggles. [Cal] converted a Kindle Fire into an Oculus Rift style head mounted display by adding a couple of magnifying lenses, some bamboo kebab sticks to hold the lenses in place. Judicious use of cardboard and duct tape completed the project. His current hurdle is getting past the Fire’s lack of an accelerometer. [Cal] planned to spend Sunday at Maker Faire adding one of his own!
As the hour grew late, everyone started to trickle out. Tired but happy from a long day at Maker Faire, the Bring A Hacker partygoers headed back to their hotels to get some sleep before World Maker Faire’s final day.