Many of the early applications for the much anticipated Oculus Rift VR rig have been in gaming. But it’s interesting to see some more useful applications besides gaming, before it’s commercial release sometime this year. [JoLau] at the Institute i4Ds of FHNW School of Engineering wanted to go a step beyond rendering virtual worlds. So he built the Intuitive Rift Explorer a.k.a IRE. The IRE is a moving reality system consisting of a gimbaled stereo-vision camera rig transmitting video to the Rift, and matching head movements received from the Oculus Rift. The vision platform is mounted on a Remote-controlled robot which is completely wireless.
One of the big challenges with using VR headsets is lag, causing motion sickness in some cases. He had to tackle the problem of latency – reducing the time from moving the head to getting a matching image on the headset – Oculus Rift team specified it should be less than 20ms. The other important requirement is a high frame rate, in this case 60 frames per second. [JoLau] succeeded in overcoming most of the problems, although in conclusion he does mention a couple of enhancements that he would like to add, given more time.
[JoLau] provides a detailed description of the various sub-systems that make up IRE – the Stereo camera, audio and video transmission, media processing, servo driven gimbal for the stereo camera, and control system code. [JoLau]’s reasoning on some of the interesting hardware choices for several components used in the project makes for interesting reading. Watch a video of the IRE in action below.
Continue reading “Putting Oculus Rift on a Robot”
[Matt], [Andrew], [Noah], and [Tim] have a pretty interesting build for their capstone project at Ohio Northern University. They’re using a Microsoft Kinect, and a Leap Motion to create a natural user interface for controlling humanoid robots.
The robot the team is using for this project is a tracked humanoid robot they’ve affectionately come to call Johnny Five. Johnny takes commands from a computer, Kinect, and Leap motion to move the chassis, arm, and gripper around in a way that’s somewhat natural, and surely a lot easier than controlling a humanoid robot with a keyboard.
The team has also released all their software onto Github under an open source license. You can grab that over on the Gits, or take a look at some of the pics and videos from the Columbus Mini Maker Faire.
Most of the incredible flight simulator enthusiasts with 737 cockpits in their garage are from the US. What happens when they’re from Slovenia? They built an A320 cockpit. The majority of the build comes from an old Cyprus Airways aircraft, with most of the work being wiring up the switches, lights, and figuring out how to display the simulated world out of the cockpit.
Google Cardboard is the $4 answer to the Oculus Rift – a cardboard box and smartphone you strap to your head. [Frooxius] missed being able to interact with objects in these 3D virtual worlds, so he came up with this thing. He adapted a symbol tracking library for AR, and is now able to hold an object in his hands while looking at a virtual object in 3D.
Heat your house with candles! Yes, it’s the latest Indiegogo campaign that can be debunked with 7th grade math. This “igloo for candles” will heat a room up by 2 or 3 degrees, or a little bit less than a person with an average metabolism will.
Last week, we saw a post that gave the Samsung NX300 the ability to lock the pictures taken by the camera with public key cryptography. [g3gg0] wrote in to tell us he did the same thing with a Canon EOS camera.
The guys at Flite Test put up a video that should be handy for RC enthusiasts and BattleBot contenders alike. They’re tricking out transmitters, putting push buttons where toggle switches should go, on/off switches where pots should go, and generally making a transmitter more useful. It’s also a useful repair guide.
[Frank Zhao] made a mineral oil aquarium and put a computer in it. i7, GTX 970, 16GB RAM, and a 480GB SSD. It’s a little bigger than most of the other aquarium computers we’ve seen thanks to the microATX mobo, and of course there are NeoPixels and a bubbly treasure chest.
[Jason] has been playing around with the Oculus Rift lately and came up with a pretty cool software demonstration. It’s probably been done before in some way shape or form, but we love the idea anyway and he’s actually released the program so you can play with it too!
It’s basically a 3D Windows Manager, aptly called 3DWM — though he’s thinking of changing the name to something a bit cooler, maybe the WorkSphere or something.
As he shows in the following video demonstration, the software allows you to set up multiple desktops and windows on your virtual sphere created by the Oculus — essentially creating a virtual multi-monitor setup. There’s a few obvious cons to this setup which makes it a bit unpractical at the moment. Like the inability to see your keyboard (though this shouldn’t really be a problem), the inability to see people around you… and of course the hardware and it’s lack of proper resolution. But besides that, it’s still pretty awesome!
In the future development he hopes to add Kinect 2 and Leap Motion controller integration to help make it even more user intuitive — maybe more Minority Report style.
Continue reading “Using the Oculus Rift as a Multi-Monitor Replacement”
Most of us have heard of Second Life – that antiquated online virtual reality platform of yesteryear where users could explore, create, and even sell content. You might be surprised to learn that not only are they still around, but they’re also employing the Oculus Rift and completely redesigning their virtual world. With support of the DK2 Rift, the possibilities for a Second Life platform where users can share and explore each other’s creations opens up some interesting doors.
Envision a world where you could log on to a “virtual net”, put on your favorite VR headset and let your imagination run wild. You and some friends could make a city, a planet…and entire universe that you and thousands of others could explore. With a little bit of dreaming
and an arduino, VR can bring dreams to life.
Continue reading “Ask Hackaday: What is The Future of Virtual Reality?”
Ever since [will1384] watched “The Lawnmower Man” as a wee lad, he’s been interested in virtual reality. He has been messing around with it for years and even had a VictorMaxx Stuntmaster, one of the first available head mounted displays. Years later, the Oculus Rift came out and [will1384] wanted to try it out but the $350 price tag put it just out of his price range for a discretionary purchase. He then did what most of us HaD readers would do, try building one himself, and with a goal for doing it for around $100.
The main display is a 7″ LCD with a resolution of 1024×600 pixels and has a mini HDMI input. Some DIY head mounted display projects out on the ‘web use ski goggles or some sort of elastic strap to hold the display to the wearer’s head. [will1384] took a more industrial approach, literally. He used the head mounting system from a welding helmet. This not only has an adjustable band but also has a top strap to prevent the entire contraption from sliding down. Three-dimensional parts were printed out to secure the LCD to the welding helmet parts while at the same time creating a duct to block out external light.
Inside the goggles are a pair of 5x Loupe lenses mounted between the user’s eyes and the LCD screen. These were made to be adjustable so that the wearer can dial them in for the most comfortable viewing experience. The remote mounted to the top strap may look a little out-of-place but it is actually being used to capture head movement. In addition to a standard wireless remote, it is also an air mouse with internal gyroscopes.
The Oculus Rift and all the other 3D video goggle solutions out there are great if you want to explore virtual worlds with stereoscopic vision, but until now we haven’t seen anyone exploring real life with digital stereoscopic viewers. [pabr] combined the Kinect-like sensor in an ASUS Xtion with a smartphone in a Google Cardboard-like setup for 3D views the human eye can’t naturally experience like a third-person view, a radar-like display, and seeing what the world would look like with your eyes 20 inches apart.
[pabr] is using an ASUS Xtion depth sensor connected to a Galaxy SIII via the USB OTG port. With a little bit of code, the output from the depth sensor can be pushed to the phone’s display. The hardware setup consists of a VR-Spective, a rather expensive bit of plastic, but with the right mechanical considerations, a piece of cardboard or some foam board and hot glue would do quite nicely.
[pabr] put together a video demo of his build, along with a few examples of what this project can do. It’s rather odd, and surprisingly not a superfluous way to see in 3D. You can check out that video below.
Continue reading “Seeing The World Through Depth Sensing Cameras”