Oculus, as we know, was acquired by Facebook for $2 billion, and now the VR community has been buzzing about trying to figure out what to do with all this newly accessible technology. And adding to the interest, the 2nd iteration of the development kits were released, causing a resurgence in virtual reality development as computer generated experiences started pouring out from of every corner of the world. But not everyone can afford the $350 USD price tag to purchase one of these devices, bringing out the need for Do-It-Yourself projects like these 3D printed wearable video goggles via Adafruit.
The design of this project is reminiscent of the VR2GO mobile viewer that came out of the MxR Lab (aka the research environment that spun out Palmer Lucky before he created Oculus). However, the hardware here is more robust and utilizes a 5.6″ display and 50mm aspheric lenses instead of a regular smart phone. The HD monitor is held within a 3D printed enclosure along with an Arduino Micro and 9-DOF motion sensor. The outer hood of the case is composed of a combination of PLA and Ninjaflex printing-filament, keeping the fame rigid while the area around the eyes remain flexible and comfortable. The faceplate is secured with a mounting bracket and a pair of aspheric lenses inside split the screen for stereoscopic video. Head straps were added allowing for the device to fit snugly on one’s face.
At the end of the tutorial, the instructions state that once everything is assembled, all that is required afterwards is to plug in a 9V power adapter and an HDMI cable sourcing video from somewhere else. This should get the console up and running; but it would be interesting to see if this design in the future can eliminate the wires and make this into a portable unit. Regardless of which, this project does a fantastic job at showing what it takes to create a homemade virtual reality device. And as you can see from the product list after the break, the price of the project fits under the $350 DK2 amount, helping to save some money while still providing a fun and educational experience.
Continue reading “3D Printed Virtual Reality Goggles”
Virtual Reality by function pushes the boundaries of what we perceive as existence, tricking the mind into believing that the computer generated environment that the user is thrust into actually contains a real place. So in the spirit of seeing what is possible in VR, a developer named [Jacques] hooked up a Raspberry Pi to an Oculus Rift. He used a computer graphics rendering API called OpenGL ES, which is much like any mobile platform found these days, to render a floating, rotating cube.
All his tests were done on a Release build which utilized the official vertex and fragment shaders. There was no attempt to optimize anything; not like there would be much to do anyways. The scene was rendered twice at 16 milliseconds per frame. From there, he attempted 27 ms per frame with texture, followed by 36 ms/frame, and then 45.
The code used can be found on [Jacques]’s Github account. A simple improvement would use a Banana Pi for better processing speed. However, don’t expect any spectacular results with this type of setup. Really, the project only proves that it’s possible to minimize a VR experience into something that could become portable. And in the same vein, the Pi + Oculus integration can produce an uncomfortable lagging effect if things are not lined up properly. But once the energy/computing power issues are addressed, VR devices could transform into a more fashionable product like Google Glass, where a simple flip of a switch would toggle the view between VR and AR into a something more mixed. And then a motion sensing input camera like this Kinect-mapping space experiment could allow people all over the world to jump into the perspectives of other reality-pushing explorers. That’s all far down the line though, but this project lays the foundation for what the future might hold.
To see [Jacques]’s full set up, view the video after the break.
Continue reading “Testing VR Limits with a Raspberry Pi”
On June 26th, 2014, Clearpath Robotics opened up the doors to their brand new 12,000 square foot robot lair by bringing out a PR2 to cut the ceremonial ribbon and welcome everyone inside. And instead of just programming the ‘locate and destroy’ ribbon sequence, the co-founders opted to use an Oculus Rift to control the robot tearing through the material with flailing arms.
This was accomplished having Jake, the robot, utilize a Kinect 2.0 that fed skeleton tracking data via rosserial_windows, a windows-based set of extension for the Robot Operating System which we heard about in January. The software gathers in a stream of data points each with an X,Y,Z component allowing [Jake] to find himself within a 3D space.Then, the data was collected and published directly into the PR2’s brain. Inject a little python code, and the creature was able to route directions in order to move it’s arms.
Thus, by simply stepping in front of the Kinect 2.0, and putting on the Oculus Rift headset, anyone could teleoperate [Jake] to move around and wave its arms at oncoming ribbons. Once completed, [Jake] would leave the scene, journeying back into the newly created robot lair leaving pieces of nylon and polyester everywhere.
An earlier (un-smoothed) version of the full system can be seen after the break:
Continue reading “Cutting Ribbons with Robots and a Oculus Rift”
One of our trusty tipsters named [Arman] wrote in to tell us about this awesome little Horror VR Hackathon that sought to create a non-lethal electric chair, for a seriously creepy and shocking experience.
[Arman] works in a small prototyping shop, so when a few guys from the local VR group called to ask for help building a non-lethal electric chair, he thought they were joking — until they showed up at the shop! Finally understanding what they really wanted to do, he hooked them up with an EL wire power supply (high voltage AC, low amperage) for their first prototype.
Unfortunately the EL power supply driver took too much juice, so they called [Arman] back the next day to hack together some of those joke gum shockers instead — he hooked them up to an Arduino and they work like a charm. Continue reading “Non-Lethal Electric Chair Brings the Death Row Experience Home”
Craving some virtual reality goodness? Unsure of Oculus after Facebook purchased them? Well — why not make your own then!
At last weeks Google I/O conference, those lucky enough to attend received the Google Cardboard VR kit. It’s basically just a piece of cardboard, two lenses, a few magnets, an NFC tag and some velcro — but when you slide your phone into it and download the Cardboard app — you have virtual reality, on your phone.
This inspired [Wolfgang] to make his own variation of this, except instead of a phone, it fits a tablet much nicer. It really is just a cardboard box with the lenses glued in place — but it works! Of course you could 3D print a nice housing — but if you’re super excited to try out some VR apps — cardboard will do the trick as well!
Besides the Cardboard app there’s a few more Android VR applications worth a look — Tuscany Dive (explore Tuscany from the comfort of your chair), ViewR (a voice controlled experience), and Dive Volcano VR Demo (explore a volcano!).
Only trick is finding those pesky lenses…
Third person video games are never really that realistic — you get a much wider range of vision, you can typically see around things your character can’t actually see… the list goes on. But what would it be like to have a third person perspective, in real life?
That’s exactly what some hackers in Poland decided to do! This is their Real World Third Person Perspective VR / AR Experiment. It makes use of an Oculus Rift, two GoPros, a microprocessor and a few servo motors. It’s essentially a glorified camera on a stick that you wear as a backpack, but nonetheless it has a really cool effect.
The project was built in under 2 days to get into the tight deadline for Intel’s Wearable contest, which has an impressive prize list, including a grand prize of $500,000 for business development! They didn’t place, but it’s still a Hack a Day worthy project!
Check it out!
Continue reading “Third Person Perspective is Guaranteed to Mess With Your Senses”
The Kinect has long been able to create realistic 3D models of real, physical spaces. Combining these Kinect-mapped spaces with an Oculus Rift is something brand new entirely.
[Thomas] and his fellow compatriots within the Kintinuous project are modeling an office space with the old XBox 360 Kinect’s RGB+D sensors. then using an Oculus Rift to inhabit that space. They’re not using the internal IMU in the Oculus to position the camera in the virtual space, either: they’re using live depth sensing from the Kinect to feed the Rift screens.
While Kintinuous is very, very good at mapping large-scale spaces, the software itself if locked up behind some copyright concerns the authors and devs don’t have control over. This doesn’t mean the techniques behind Kintinuous are locked up, however: anyone is free to read the papers (here’s one, and another, PDF of course) and re-implement Kintinuous as an open source project. That’s something that would be really cool, and we’d encourage anyone with a bit of experience with point clouds to give it a shot.
Continue reading “Virtual Physical Reality With Kintinuous And An Oculus Rift”