A simple way to integrate physical feedback into a virtual experience is to use a fan to blow air at the user. This idea has been done before, and the fans are usually the easy part. [Paige Pruitt] and [Sean Spielberg] put a twist on things in their (now-canceled) Kickstarter campaign called ZephVR, which featured two small fans mounted onto a VR headset. The bulk of their work was in the software, which watches the audio signal for recognizable “wind” sounds, and uses those to turn on one or both fans in response.
The benefit of using software to trigger fans based on audio cues is that the whole system works independently of everything else, with no need for developers and software to build in support for your project, or to use other middleware. Unfortunately the downside is that the results are only as good as the ability of software to pick the right sounds and act on them. Embedded below is a short video showing a test in action.
Paradise means something different for everyone, it could be a sitting by a fire on a rainy night or lying on a sun-kissed beach. But for us, and makers like [liltreat4you], it’s a well stocked scrap pile out behind the house. After buying a racing wheel and pedals for his Xbox, he took a trip out to his little slice of paradise and found nearly all the hardware he needed to build a professional looking race simulator. According to his breakdown, most of the money he spent on this build ended up going into that sweet red paint job and the speed-enhancing stickers.
Everything the light touches is our kingdom.
Not all of us are as lucky as [liltreat4you], and we probably won’t just happen upon a driver’s seat out of a Mazda, or a bunch of perfectly bent metal pipes from an old trampoline out on the back forty. But trolling Craigslist or cruising around for flea markets can still get you parts like these for cheap, so try not to be too discouraged if your backyard isn’t quite as well stocked.
Once he had the metal pipes and seat from the car, the rest of the build came together pretty quickly. After building an oval out of his salvaged pipes, he attached the seat and the arms that would eventually hold the steering wheel and display. A plate was also added at the bottom for the pedals to sit on. By using long bolts, [liltreat4you] was even able to add a degree of adjustment to the wheel position. Being that he got his seat out of a real car, there’s the usual adjustment you’d expect there as well.
Speaking of which, [liltreat4you] casually mentions that you should disconnect the battery of the donor vehicle before taking out the seat, as it’s possible that the removal of the seat or the disconnection of the seat harness can cause the airbags to deploy. We can neither confirm nor deny this, but it’s probably safe advice to follow.
In July of 2016 a game was released that quickly spread to every corner of the planet. Pokemon Go was an Augmented Reality game that used a smart phone’s GPS location and camera to place virtual creatures into the person’s real location. The game was praised for its creativity and was one of the most popular and profitable apps in 2016. It’s been download over 500 million times since.
Most of its users were probably unaware that they were flirting with a new and upcoming technology called Augmented Reality. A few day ago, [floz] submitted to us a blog from a student who is clearly very aware of what this technology is and what it can do. So aware in fact that they made their own Augmented Reality system with Python and OpenCV.
In the first part of a multi-part series – the student (we don’t know their name) walks you through the basic structure of making a virtual object appear on a real world object through a camera. He 0r she gets into some fairly dense math, so you might want to wait until you have a spare hour or two before digging into this one.
VR is in vogue, but getting on board requires a steep upfront cost. Hackaday.io user [Colin Pate] felt that $800 was a bit much for even the cheapest commercial 360-degree 3D camera, so he thought: ‘why not make my own for half that price?’
[Pate] knew he’d need a lot of bandwidth and many GPIO ports for the camera array, so he searched out the Altera Cyclone V SOC FPGA and a Terasic DE10-Nano development board to host it. At present, he has four Uctronics OV5642 cameras on his rig, chosen for their extensive documentation and support. The camera mount itself is a 3D-printed octagon so eight of the OC5642 can capture a full 360-degree photo.
MIT’s Computer Science and Artificial Intelligence Lab, CSAIL, has created a process of teleoperating a Baxter humanoid robot with an Oculus Rift VR headset. This project is partially aimed towards making manufacturing jobs a hell of a lot of fun telecommutable. It could even be a way to supervise robot workers from a distance.
In a nutshell, the user controls the robot remotely in a virtual reality environment. The user does this specifically in a VR environment modeled like a control room with multiple sensor displays, making it feel like they are sitting inside the robot’s head. By using hand controllers, users can match their movements to the robot’s to complete various tasks. If you’ve seen Pacific Rim, you are probably envisioning a Jaegar right about now — minus the psychic linking.
This video demonstrates a really interesting experiment: sticking a Vive Tracker onto an ordinary chair in order to sync it up perfectly with its VR counterpart. The result? A chair that is visible in VR as a virtual object, but has a 1:1 physical world version occupying the same space. This means that unlike any other virtual object, this chair can be seen, touched, felt, moved, and actually sat in while the user is immersed in VR.
The purpose of this experiment seems to have been to virtually explore seating arrangements for real-world environments, and spawned a theatre planning tool by design studio [Agile Lens]. But we wonder if there’s unrealized potential in the idea of connecting physical objects that can be touched and held (or sat on) with their VR counterparts. Video demos of the chair test are embedded below.
Here is a virtual spray painting project with a new and DIY twist to it. [Adam Amaral]’s project is an experiment in using the Vive Tracker, which was released earlier this year. [Adam] demonstrates how to interface some simple hardware and 3D printed parts to the Tracker’s GPIO pins, using it as a custom peripheral that is fully tracked and interactive in the Vive’s VR environment. He details not only the custom spray can controller, but also how to handle the device on the software side in the Unreal engine. The 3D printed “spray can controller” even rattles when shaken!
There’s one more trick. Since the Vive Tracker is wireless and completely self-contained, the completed rattlecan operates independently from the VR headset. This means it’s possible to ditch the goggles and hook up a projector, then use the 3D printed spray can to paint a nearby wall with virtual paint; you can see that part in action in the video embedded below.