Your VR Doesn’t Stink (Yet)

What does it smell like when the wheels heat up on that Formula 1 car you drive at night and on the weekends? You have no idea because the Virtual Reality experience that lets you do so doesn’t come with a nasal component. Yet.

Shown here is an olfactory device that works with Oculus Rift and other head-mounted displays. The proof of concept is hte work of [Kazuki Hashimoto], [Yosuke Maruno], and [Takamichi Nakamoto] and was shown of at last year’s IEEE VR conference. It lets the wearer smell the oranges when approaching a tree in a virtual environment. In other words, it makes your immersive experience smelly.

As it stands this a pretty cool little device which atomizes odor droplets while a tiny fan wafts them to the wearer’s nose. There is a paper which presumably has more detail but it’s behind a pay wall so for now check out the brief demo video below. Traditionally an issue with scent systems is the substance stuck in the lines, which this prototype overcomes with direct application from the reservoir. Yet to be solved is the availability for numerous different scents.

This build came to our attention via an UploadVR article that does a good job of covering some of the scent-based experiments over the years. They see some of the same hurdles we do: odors linger and there is a limited palette that can be produced. We assume the massive revenue of the gaming industry is going to drive research in this field, but we won’t be lining up to smell gunpowder and dead bodies (or rotting zombies) anytime soon.

The more noble effort is in VR applications like taking the elderly and immobile back for another tour of places they’ll never again be able to visit in their lives. Adding the sense of smell, which has the power to unlock so many memories, makes that use case so much more powerful. We think that’s something everyone can be hopeful about!

Continue reading “Your VR Doesn’t Stink (Yet)”

Oculus Pi

[WayneKeenan] wrote a proof-of-concept virtual reality system that used a Raspberry Pi and an Oculus Rift. It was about a thousand lines of Python and with a battery pack it was even portable. The problem was that the Pi was struggling to create the 3D views.

[Wayne] recently revisited the demo and found that just about everything has gotten better: the Pi 3 is faster, and the Python libraries have become better. He spent some time building a library — VR Zero — and then recreated the original demo in 80 more lines of Python. You can see a video, below.

Continue reading “Oculus Pi”

Facebook To Slurp Oculus Rift Users’ Every Move

The web is abuzz with the news that the Facebook-owned Oculus Rift has buried in its terms of service a clause allowing the social media giant access to the “physical movements and dimensions” of its users. This is likely to be used for the purposes of directing advertising to those users and most importantly for the advertisers, measuring the degree of interaction between user and advert. It’s a dream come true for the advertising business, instead of relying on eye-tracking or other engagement studies on limited subsets of users they can take these metrics from their entire user base and hone their offering on an even more targeted basis for peak interaction to maximize their revenue.

Hardly a surprise you might say, given that Facebook is no stranger to criticism on privacy matters. It does however represent a hitherto unseen level of intrusion into a user’s personal space, even to guess the nature of their activities from their movements, and this opens up fresh potential for nefarious uses of the data.

Fortunately for us there is a choice even if our community doesn’t circumvent the data-slurping powers of their headsets; a rash of other virtual reality products are in the offing at the moment from Samsung, HTC, and Sony among others, and of course there is Google’s budget offering. Sadly though it is likely that privacy concerns will not touch the non-tech-savvy end-user, so competition alone will not stop the relentless desire from big business to get this close to you. Instead vigilance is the key, to spot such attempts when they make their way into the small print, and to shine a light on them even when the organisations in question would prefer that they remained incognito.

Oculus Rift development kit 2 image: By Ats Kurvet – Own work, CC BY-SA 4.0, via Wikimedia Commons.

Being Hit in the Virtual World

Virtual reality could be the next big thing in the gaming world. And while VR displays and headsets are getting more and more sophisticated, many are forgetting perhaps the biggest feature VR will need to succeed — haptic feedback. [Pedro Lopes], [Alexandra Ion] and [Prof. Patric Baudisch] from the Hasso Plattner Institute is hoping to change that, with a project called Impacto: Simulating Physical Impact by Combining Tactile with Electrical Muscle Stimulation.

We’ve covered lots of haptic feedback devices over the past few years — some use mini gyros to simulate resistance, others blow air on you, but this is the first time we’ve seen one that combines muscle stimulation to really cause a physical effect.

They’re using an Oculus rift, and a Microsoft Kinect to perform the research. For their demonstration they use a basic boxing game that allows the user to feel the computer’s punches — but don’t worry, it doesn’t hit that hard!

Continue reading “Being Hit in the Virtual World”

3D Printed Eyeglasses, VR Lenses

[Florian] is hyped for Google Cardboard, Oculus Rifts, and other head mounted displays, and with that comes an interest in lenses. [Floian] wanted to know if it was possible to create these lenses with a 3D printer. Why would anyone want to do this when these lenses can be had from dozens of online retailers for a few dollars? The phrase, ‘because I can’ comes to mind.

The starting point for the lens was a CAD model, a 3D printer, and silicone mold material. Clear casting resin fills the mold, cures, and turns into a translucent lens-shaped blob. This is the process of creating all lenses, and by finely sanding, polishing, and buffing this lens with grits ranging from 200 to 7000, this bit of resin slowly takes on an optically clear shine.

Do these lenses work? Yes, and [Florian] managed to build a head mounted display that can hold an iPhone up to his face for viewing 3D images and movies. The next goal is printing prescription glasses, and [Florian] seems very close to achieving that dream.

The last time we saw home lens making was more than a year ago. Is anyone else dabbling in this dark art? Let us know in the comments below and send in a tip if you have a favorite lens hack in mind.

Hacklet 54 – Virtual Reality Projects

Virtual Reality is finally coming of age. Hackers, Makers and Engineers have dreamed of creating immersive interfaces for years. From the first flight simulators to today’s cellphone powered head mounted displays, VR has always been an exciting field. Many of the advances today are being created by hackers who were inspired by systems like Virtuality from the early 1990’s. Now 25 years on, we’re seeing amazing advances – not only in commercial systems, but in open source VR projects. This week’s Hacklet is all about the best VR projects on!

vr1We start with [j0nno] and D.I.Y Virtual Reality. [J0nno] has become interested in VR, and decided to build his own head mounted display. His goal is to create a setup with full head tracking and an open source software stack. He’s hoping to do this within a budget of just $200 AUD. [J0nno] started with the Ritech3d-V2 VR Goggles, which are a plastic implementation of Google’s project cardboard. For display he’s using a 5.6 inch 1280 x 800 TFT LCD. Tracking is optical, using IR LEDs and a PS3 Eye camera. [J0nno’s] background is in software, so he’s doing great setting up OpenVR and Perception. The hardware side is a bit new to him. This isn’t stopping [J0nno] though! In true hacker spirit, he’s learning all about resistors and driving LEDs as he works on D.I.Y Virtual Reality.

vr2Next up is [Josh Lindsay] with Digitabulum: The last motion-capture glove. Digitabulum is a motion capture glove designed to be able to emulate most other motion capture systems. It is also designed to be relatively low-cost. At $400 per hand, it is less expensive than most other offerings, though we’d still love to see something even cheaper. [Josh] is going with inertial sensors, and a lot of them. Specifically he’s using no less than 17 LSM9DS1 Inertial Measurement Unit (IMU) sensors from ST Microelectronics. IMU sensors like this combine multiple rate gyros, accelerometers, and magnetometers into a single unit. Essentially every segment of every finger has its own sensor suite. As you might imagine, that is quite a bit of data to crunch. An Altera Max II CPLD and an ST Arm processor help boil down the data to something which a VR engine can process. [Josh] has been working on this project for over a year now, and he’s making great progress. The prototype glove looks terrific!

vr3[Thomas] brings augmented reality to the table with Oculus Rift featured Crane control. What started as a hobby experiment became [Thomas’] major project at university. He’s connected an Oculus Rift to a toy crane. A stereo camera on the crane sends a video image to the operator. The camera is mounted on a pan/tilt mechanism driven by the Rift’s head tracking unit. Simple joystick controls allow [Thomas] to move the boom and lower the line. On-screen displays show the current status of the crane. The use of the Rift makes this an immersive demonstration. One could easily see how moving this system into the real world would make crane operations safer for crane operators.

vr4Finally we have [Arcadia Labs] with DIY Augmented Reality Device. This project, which is the [Arcadia Labs] entry in the 2015 Hackaday Prize, uses two 320 x 240 screens to create an augmented reality head mounted display. While the resolution can’t match that of the Oculus Rift or HTC Vive, [Arcadia Labs] is ok with that. They’re going for a lower cost open source alternative for augmented reality. Tracking is achieved with an IMU, while a PS3 Eye camera provides the video. A Raspberry Pi controls the show. [Arcadia Labs] was able to get 50 frames per second on the displays just using the Pi’s SPI interface, however the USB PS3 Eye camera limits things to around 10 FPS. This project is under heavy development right now, so follow along with us to see where [Arcadia Labs] ends up!

If you want VR goodness, check out our new virtual reality projects list! Did I miss your project? Don’t be shy, just drop me a message on If you’re on the left coast of the USA, check out SOCAL Virtual Reality Conference and Expo. Hackaday is a sponsor. The event happens on July 12 at the University of California Irvine.

That’s it for this week’s Hacklet, As always, see you next week. Same hack time, same hack channel, bringing you the best of!

Putting Oculus Rift on a Robot

Many of the early applications for the much anticipated Oculus Rift VR rig have been in gaming. But it’s interesting to see some more useful applications besides gaming, before it’s commercial release sometime this year. [JoLau] at the Institute i4Ds of FHNW School of Engineering wanted to go a step beyond rendering virtual worlds. So he built the Intuitive Rift Explorer a.k.a IRE. The IRE is a moving reality system consisting of a gimbaled stereo-vision camera rig transmitting video to the Rift, and matching head movements received from the Oculus Rift. The vision platform is mounted on a Remote-controlled robot which is completely wireless.

One of the big challenges with using VR headsets is lag, causing motion sickness in some cases. He had to tackle the problem of latency – reducing the time from moving the head to getting a matching image on the headset – Oculus Rift team specified it should be less than 20ms. The other important requirement is a high frame rate, in this case 60 frames per second. [JoLau] succeeded in overcoming most of the problems, although in conclusion he does mention a couple of enhancements that he would like to add, given more time.

[JoLau] provides a detailed description of the various sub-systems that make up IRE – the Stereo camera,  audio and video transmission, media processing, servo driven gimbal for the stereo camera,  and control system code. [JoLau]’s reasoning on some of the interesting hardware choices for several components used in the project makes for interesting reading. Watch a video of the IRE in action below.

Continue reading “Putting Oculus Rift on a Robot”