Virtual Reality by function pushes the boundaries of what we perceive as existence, tricking the mind into believing that the computer generated environment that the user is thrust into actually contains a real place. So in the spirit of seeing what is possible in VR, a developer named [Jacques] hooked up a Raspberry Pi to an Oculus Rift. He used a computer graphics rendering API called OpenGL ES, which is much like any mobile platform found these days, to render a floating, rotating cube.
All his tests were done on a Release build which utilized the official vertex and fragment shaders. There was no attempt to optimize anything; not like there would be much to do anyways. The scene was rendered twice at 16 milliseconds per frame. From there, he attempted 27 ms per frame with texture, followed by 36 ms/frame, and then 45.
The code used can be found on [Jacques]’s Github account. A simple improvement would use a Banana Pi for better processing speed. However, don’t expect any spectacular results with this type of setup. Really, the project only proves that it’s possible to minimize a VR experience into something that could become portable. And in the same vein, the Pi + Oculus integration can produce an uncomfortable lagging effect if things are not lined up properly. But once the energy/computing power issues are addressed, VR devices could transform into a more fashionable product like Google Glass, where a simple flip of a switch would toggle the view between VR and AR into a something more mixed. And then a motion sensing input camera like this Kinect-mapping space experiment could allow people all over the world to jump into the perspectives of other reality-pushing explorers. That’s all far down the line though, but this project lays the foundation for what the future might hold.
To see [Jacques]’s full set up, view the video after the break.
Continue reading “Testing VR Limits with a Raspberry Pi”
A ton of people sent in this video of crazy Russians who have taken a microwave, removed the magnetron, taped it to a broom, and turned it on. Don’t try this at home. Or near us.
You know the Google Cardboard kit that’s a real VR headset made of cardboard (and a smart phone)? Google may have gotten their inspiration from Oculus, because every Oculus Rift DK2 ships with a Samsung Galaxy Note 3 inside.
Ever design a PCB and be disappointed by the quality of the silkscreen? [Paul Allen] has been defining the edges of his PCB labels with the copper layer, and the examples are dramatic. Etching copper is what you actually pay for when you fab a board, so it should come as no surprise that the quality is a little higher.
Dunk tanks are fun, but how about competitive dunk tanks? [Chad] built a dunk tank (really more of a ‘dunk shower’) out of a 2×4 tripod, a garbage can, and a few parts from a the toilet aisle of Home Depot’s plumbing department. Then he built a second. Set up both dunk showers across from each other, give two people a few balls, and see who gets soaked last. Looks fun.
Want a MAME cabinet, but don’t want it taking up room in your house? Build a MAME coffee table! Here’s the reddit thread. Maybe we’re old-fashioned, but we’d rather have a giant NES controller coffee table.
Last week we saw a 16-bobbin rope braiding machine, but odd braiding machines like this aren’t limited to fibers. Here’s a wire twisting machine for making RS422 cables. It only produces a single twisted pair, but that’s really all you need to create a cable. Somebody get some paracord and make some Cat5.
On June 26th, 2014, Clearpath Robotics opened up the doors to their brand new 12,000 square foot robot lair by bringing out a PR2 to cut the ceremonial ribbon and welcome everyone inside. And instead of just programming the ‘locate and destroy’ ribbon sequence, the co-founders opted to use an Oculus Rift to control the robot tearing through the material with flailing arms.
This was accomplished having Jake, the robot, utilize a Kinect 2.0 that fed skeleton tracking data via rosserial_windows, a windows-based set of extension for the Robot Operating System which we heard about in January. The software gathers in a stream of data points each with an X,Y,Z component allowing [Jake] to find himself within a 3D space.Then, the data was collected and published directly into the PR2’s brain. Inject a little python code, and the creature was able to route directions in order to move it’s arms.
Thus, by simply stepping in front of the Kinect 2.0, and putting on the Oculus Rift headset, anyone could teleoperate [Jake] to move around and wave its arms at oncoming ribbons. Once completed, [Jake] would leave the scene, journeying back into the newly created robot lair leaving pieces of nylon and polyester everywhere.
An earlier (un-smoothed) version of the full system can be seen after the break:
Continue reading “Cutting Ribbons with Robots and a Oculus Rift”
One of our trusty tipsters named [Arman] wrote in to tell us about this awesome little Horror VR Hackathon that sought to create a non-lethal electric chair, for a seriously creepy and shocking experience.
[Arman] works in a small prototyping shop, so when a few guys from the local VR group called to ask for help building a non-lethal electric chair, he thought they were joking — until they showed up at the shop! Finally understanding what they really wanted to do, he hooked them up with an EL wire power supply (high voltage AC, low amperage) for their first prototype.
Unfortunately the EL power supply driver took too much juice, so they called [Arman] back the next day to hack together some of those joke gum shockers instead — he hooked them up to an Arduino and they work like a charm. Continue reading “Non-Lethal Electric Chair Brings the Death Row Experience Home”
We’re pretty sure the Hackaday demographic is a a person who sees a giant tower crane lifting beams and girders above a skyline and says, “that would be fun, at least until I have to go to the bathroom.” Realizing the people who own these cranes probably won’t let any regular joe off the street into the cabin, [Thomas] and [screen Name] (see, this is why we have brackets, kids) built their own miniature version with an Oculus Rift.
Instead of a crane that is hundreds of feet tall, the guys are using a much smaller version, just over a meter tall, that is remotely controlled through a computer via a serial connection. Just below the small plastic cab is a board with two wide-angle webcams. The video from these cameras are sent to the Oculus so the operator can see the boom swinging around, and the winch unwinding to pick up small objects.
The guys have also added a little bit of OpenCV to add color based object detection. This is somewhat useful, but there’s also an approximation of the distance to an object, something that would be very useful if you don’t have a three-inch tall spotter on the ground.
Continue reading “The Crane Game, Oculus Style”
Third person video games are never really that realistic — you get a much wider range of vision, you can typically see around things your character can’t actually see… the list goes on. But what would it be like to have a third person perspective, in real life?
That’s exactly what some hackers in Poland decided to do! This is their Real World Third Person Perspective VR / AR Experiment. It makes use of an Oculus Rift, two GoPros, a microprocessor and a few servo motors. It’s essentially a glorified camera on a stick that you wear as a backpack, but nonetheless it has a really cool effect.
The project was built in under 2 days to get into the tight deadline for Intel’s Wearable contest, which has an impressive prize list, including a grand prize of $500,000 for business development! They didn’t place, but it’s still a Hack a Day worthy project!
Check it out!
Continue reading “Third Person Perspective is Guaranteed to Mess With Your Senses”
Ever see a really cool build on YouTube with no build details at all? Frustrating, right? That’s us with the NES Keytar covering the Game of Thrones theme. He’s using a Raspi with the sound chip in the NES to do live chiptunes. Freakin’ awesome. There’s also the ST:TNG theme as well.
A few years ago the folks at Oculus had an idea – because of cellphones, small, high resolution displays are really cheap, so why not make VR goggles? At Google IO this week someone figured out everyone already has a cellphone, so just wrap it in some cardboard and call it a set of VR goggles. You can get a kit here, but the only difficult to source components are the lenses.
What happens when you put liquid nitrogen under a vacuum? Well, it should evaporate more, get colder, and freeze. Then it breaks up into solid nitrogen snow. No idea what you would do with this, but there ‘ya go. Oh, [NC], we’re going to need a writeup of that LN2 generator.
About a month ago, the House4Hack hackerspace in South Africa told us of their plans to bring a glider down from 20km above the Earth. They finally launched it, The CAA only allowed them to glide back from 6km (20,000 feet), but even from there the foam glider hit 230kph (124 knots). That’s a little impressive for a foam FPV platform, and we’re betting something with a larger wingspan would probably break a spar or something. Shout out to HABEX.
All the electronic dice projects we’ve seen have one thing in common: they’re not cubes. Thus uberdice. It’s six nine-pixel displays on the faces of a cube, powered by a battery, and controlled by an accelerometer. Yes, it is by far the most complicated die ever made, but it does look cool.