Here is a virtual spray painting project with a new and DIY twist to it. [Adam Amaral]’s project is an experiment in using the Vive Tracker, which was released earlier this year. [Adam] demonstrates how to interface some simple hardware and 3D printed parts to the Tracker’s GPIO pins, using it as a custom peripheral that is fully tracked and interactive in the Vive’s VR environment. He details not only the custom spray can controller, but also how to handle the device on the software side in the Unreal engine. The 3D printed “spray can controller” even rattles when shaken!
There’s one more trick. Since the Vive Tracker is wireless and completely self-contained, the completed rattlecan operates independently from the VR headset. This means it’s possible to ditch the goggles and hook up a projector, then use the 3D printed spray can to paint a nearby wall with virtual paint; you can see that part in action in the video embedded below.
Continue reading “Spray Paint Goes DIY Virtual with a Vive Tracker”
[Koppany Horvarth] set out to create a dirt-cheap optical tracking rig for VR that uses only two cameras and a certain amount of math to do its thing. He knew he could do theoretically, and wouldn’t cost a lot of money, but still required a lot of work and slightly absurd amount of math.
While playing around with a webcam that he’d set up to run an object-tracking Python script and discovered that his setup tended to display a translucent object with a LED inside of it as pure, washed-out white. This gave [Koppany] the idea that he could use such a light as part of his object tracking project. He 3D-printed 50mm hollow spheres out of transparent PLA, illuminated via a LED and powered by a 5V power supply hacked from an old USB cable. After dealing with some lens flares, he sanded down the PLA a little to diffuse the light and it worked like a charm.
To learn more check out his GitHub code repository. You can also take inspiration in some of the other motion tracking posts we’ve published in the past, like motion tracking on the cheap with a PIC and this OpenCV Airsoft turret.
Polygon reports CastAR is no more.
CastAR is the brainchild of renaissance woman [Jeri Ellsworth], who was hired by Valve to work on what would eventually become SteamVR. Valve let [Jeri] go, but allowed her to take her invention with her. [Jeri] founded a new company, Technical Illusions, with [Rick Johnson] and over the past few years the CastAR has appeared everywhere from Maker Faires to venues better focused towards innovative technologies.
In 2013, Technical Illusions got its start with a hugely successful Kickstarter, netting just north of one million dollars. This success drew the attention of investors and eventually led to a funding round of $15 million. With this success, Technical Illusions decided to refund the backers of its Kickstarter.
We’ve taken a look a CastAR in the past, and it’s something you can only experience first-hand. Unlike the Oculus, Google Cardboard, or any of the other VR plays companies are coming out with, CastAR is an augmented reality system that puts computer-generated objects in a real, physical setting. Any comparison between CastAR and a VR system is incomplete; these are entirely different systems with entirely different use cases. Think of it as the ultimate table top game, or the coolest D&D game you could possibly imagine.
Movies love to show technology they can’t really build yet. Even in 2001: A Space Oddessy (released in 1968), for example, the computer screens were actually projected film. The tablet they used to watch the news looks like something you could pick up at Best Buy this afternoon. [CircuitDigest] saw Iron Man and that inspired him to see if he could control his PC through gestures as they do on that film and so many others (including Minority Report). Although he calls it “virtual reality,” we think of VR as being visually immersed and this is really just the glove, but it is still cool.
The project uses an Arduino on the glove and Processing on the PC. The PC has a webcam which tracks the hand motion and the glove has two Hall effect sensors to simulate mouse clicks. Bluetooth links the glove and the PC. You can see a video of the thing in action, below.
Continue reading “A Minority Report Arduino-Based Hand Controller”
It wasn’t too long ago that one could conjecture that most hackers are not avid video game players. We spend most of our free time taking things apart, tinkering with microcontrollers and reading the latest [Jenny List] article on Hackaday.com. When we do think of video games, our neurons generally fire in the direction of emulating a console on a single board computer, such as a Raspberry Pi or a Beaglebone. Or even emulating the actual console processor on an FPGA. Rarely do we venture off into 3D programs meant to make modern video games. If we can’t export an .STL with it, we’re not interested. It’s just not our bag.
Oculus Rift changed this. The VR headset was originally invented for 3D video games, but quickly became a darling to hackers the world over. Virtual Reality technology is far bigger than just video games, and brings opportunity to many fields such as real estate, construction, product visualization, education, social interaction… the list goes on and on.
The Oculus team got together with the folks over at Unity in the early days to make it easy for video game makers to make content for the Rift. Unity is a game engine designed with a shallow learning curve and is available for free for non-commercial use. The Oculus Rift can be integrated into a Unity environment with the check of a setting and importing a small package, available on the Oculus site. This makes it easy for anyone interested in VR technology to get a Rift and start pumping out content.
Hackers have taken things a step further and have written scripts that allow Unity to communicate with an Arduino. VR is fun. But VR plus physical reality is just down right exciting! In this article, we’re going to walk you through setting up your Oculus Rift and Unity game engine to communicate with the outside world via an Arduino.
Continue reading “You’re the Only One not Playing with Unity”
[Joey Campbell] is studying for his PhD at the Bristol Interaction and Graphics Lab, focusing on the interplay between real and virtual objects within the realm of exergaming–“gamercising” where physical motion and effort drives the game. The goal is to make the physical effort seem to correspond with what’s seen on the headset.
[Joey] set up a test rig where an exercise bike’s gears were adjusted based on the terrain encountered, seeking to find out if that realism inspired a greater feeling of immersion. He also provided some test subjects a HUD with their heart rate and other stats, to see if that encouraged gamers to exercise more.
In his current project, [Joey] has equipped a wheelchair with a pair of Arduino-controlled servos that squeeze the brakes to simulate an obstacle. In the VR realm, a player pushes the wheelchair toward a virtual block and the brakes engage, requiring the player push harder to bypass the obstacle.
One imagines the possibilities of games designed for specifically for wheelchairs. The Eyedrivomatic wheelchair that won the 2015 Hackaday Prize sounds perfect for the job!
Continue reading “Simulating VR Obstacles with Wheelchair Brakes”
Our friend [James Bruton] from XRobots has engaged in another bit of mixed-reality magic by showing how one can seamlessly step from the virtual world into the real world, and back again. Begone, green screens and cumbersome lighting!
Now, most of what you’re seeing is really happening in post-production — for now — but the test footage is the precursor for a more integrated system down the road. As it works now, a GoPro is attached to the front of a HTC Vive headset, allowing [Bruton] to record in both realities at the same time. In the VR test area he has set up is a portal to a virtual green room — only a little smaller than a wardrobe — allowing him to superimpose the GoPro footage over everything he looks at through that doorway, as well as everything surrounding him when he steps through. Unfortunately, [Bruton] is not able to see where he’s going if he is to wear the headset, so he’s forced to hold it in one hand and move about the mixed-reality space. Again, this is temporary.
In action — well, it gets a little surreal when he starts tossing digital blocks through the gateway ‘into’ the real world.
Continue reading “VR and Back Again: An XRobots Tale”