Many of the early applications for the much anticipated Oculus Rift VR rig have been in gaming. But it’s interesting to see some more useful applications besides gaming, before it’s commercial release sometime this year. [JoLau] at the Institute i4Ds of FHNW School of Engineering wanted to go a step beyond rendering virtual worlds. So he built the Intuitive Rift Explorer a.k.a IRE. The IRE is a moving reality system consisting of a gimbaled stereo-vision camera rig transmitting video to the Rift, and matching head movements received from the Oculus Rift. The vision platform is mounted on a Remote-controlled robot which is completely wireless.
One of the big challenges with using VR headsets is lag, causing motion sickness in some cases. He had to tackle the problem of latency – reducing the time from moving the head to getting a matching image on the headset – Oculus Rift team specified it should be less than 20ms. The other important requirement is a high frame rate, in this case 60 frames per second. [JoLau] succeeded in overcoming most of the problems, although in conclusion he does mention a couple of enhancements that he would like to add, given more time.
[JoLau] provides a detailed description of the various sub-systems that make up IRE – the Stereo camera, audio and video transmission, media processing, servo driven gimbal for the stereo camera, and control system code. [JoLau]’s reasoning on some of the interesting hardware choices for several components used in the project makes for interesting reading. Watch a video of the IRE in action below.
Continue reading “Putting Oculus Rift on a Robot”
With Samsung’s new Gear VR announced, developers and VR enthusiasts are awaiting the release of the smartphone connected VR headset. A few people couldn’t wait to get their hands on the platform, so they created, OpenGear, a Gear VR compatible headset.
The OpenGear starts off with a Samsung Galaxy Note 4, which is the target platform for the Gear VR headset. A cardboard enclosure, similar to the Google Cardboard headset, holds the lenses and straps the phone to your face.
The only missing part is the motion tracking electronics. Fortunately, ST’s STM32F3 Discovery development board has everything needed: a microcontroller with USB device support, a L3GD20 3 axis gyro, and a LSM303DLHC accelerometer/magnetometer. These components together provide a USB inertial measurement unit for tracking your head.
With the Discovery board strapped to the cardboard headset, an open-source firmware is flashed. This emulates the messages sent by a legitimate Oculus Rift motion tracker. The Galaxy Note 4 sees the device as a VR headset, and lets you run VR apps.
If you’re interested, the OpenGear team is offering a development kit. This is a great way for developers to get a head start on their apps before the Gear VR is actually released. The main downside is how you’ll look with this thing affixed to your face. There’s a head-to-head against the real Gear VR after the break.
[via Road To VR]
Continue reading “Build Your Own Gear VR”
If you’ve been keeping your skills fresh with any console video games in the last 15 years (or you’ve acquired a smartphone), you’ll know that “rumble,” or “haptic” feedback plays a key role in augmenting our onscreen (or touch-screen typing) experience. Nevertheless, this sort of rumble feedback is surprisingly boolean, and hasn’t developed into a richer level of precision since it started to be introduced in gaming over 30 years ago. In response, [Martin] and his fellow design teammates at the University of Salzburg, Austria have introduced the TorqueScreen, a mobile haptic attachment that puts a twist on conventional force feedback.
At its core the TorqueScreen is a gyroscope attached to a servo with the respective rotational axes in a perpendicular alignment. When the servo rotates the live gyroscope, the user can sense the tablet’s resistance to rotate about the servos axis.
The team’s conference demonstration model features a brushless motor plucked from an old hard drive controlled by an Arduino and driven manually by a Wii nunchuck. Currently only one rotational axis resists changes in rotation, but a gimbal may be the next step in this project.
We’ve certainly seen budget handheld haptic devices before, but this project allows for an entire spread of responses proportional to the speed at which the gyroscope is rotated about the servo axis.
Unless you’re reading this post on your already-torquescreen-enabled mobile device, it’s a bit difficult to get a feel for what kind of interactions you can produce with this setup. The video (after the break), though, can give you a pretty good idea of what kind of interactivity you’d expect with device clipped onto your tablet.
[via the Tangible, Embedded, and Embodied Interaction Conference]
Continue reading “TorqueScreen: Haptic Feedback On-the-Go”
If you’ve been keeping up with augmented and virtual reality news, you’ll remember that spacial haptic feedback devices aren’t groundbreaking new technology. You’ll also remember, however, that a professional system is notoriously expensive–on the order of several thousand dollars. Grad students [Jonas], [Michael], and [Jordi] and their professor [Eva-Lotta] form the design team aiming to bridge that hefty price gap by providing you with a design that you can build at home.
A quick terminology dive: a spacial haptic device is a physical manipulator that enables exploration of a virtual space through force feedback. A user grips the “manipulandum” (the handle) and moves it within the work area defined by the physical design of the device. Spacial Haptic Devices have been around for years and serve as excellent tools for telling their users (surgeons) what something (tumor) “feels like.”
In our case, this haptic device is a two-link, two-joint system grounded on a base station and providing force feedback with servo motors and tensioned wire ropes. The manipulator itself supports 3-degree-of-freedom movement of the end-effector (translations, but no rotations) which is tracked with encoders placed on all joints. To enable feedback, joints are engaged with cable-drive transmissions.
The design team isn’t new to iterative prototyping. Hailing from CS235, a Stanford course aimed to impart protoyping techniques to otherwise non-tinkerers, the designers have drawn numerous techniques from this course to deliver a fully functional and reproducible setup. In fact, it’s clear that the designers have a strong understanding of their system’s physics, and they capitalize on a few tricks that don’t immediately jump out to us as intuitive. For instance, rather than rigidly fixing their cable to the motor shaft, they simply wrap the cable around the shaft a mere 5 turns such that the force of friction greatly exceeds the threshold amount that would otherwise cause slipping. They also choose plywood–not necessarily because of its price–but more so because of its function as a stiff, layered composite that makes it ideal “lever arm material” for rigidly transferring forces.
For a full breakdown of their design, take a look at their conference paper (PDF) where they evaluate their design techniques and outline the forward kinematics. They’ve also provided a staggeringly comprehensive bill of materials (Google Spreadsheet). Finally, as justifiably open source hardware, they’ve packaged their control software and CAD models into a github repository so that you too can jump into the world of quality force feedback simulation without shelling out the twenty thousand dollars for a professional system.
[via the 2015 Tangible Embedded and Embodied Conference]
Continue reading “Open Source Haptics Kit Aims to Democratize Force Feedback”
At long last I had the opportunity to try out the CastAR, a glasses-based Augmented Reality system developed by Technical Illusions. The hardware has been in the works now for a couple of years, but every time we have come across a demo we were thwarted by the long lines that accompany them. This time I was really lucky. [Jeri] gave us a private demo in a suite at the Palazzo during CES 2015. Reflecting on the experience, CastAR is exactly the type of Virtual Reality hardware I’ve been longing for.
Continue reading “CastAR Hands-On and Off-Record Look at Next Version”
Virtual reality has come a long way but some senses are still neglected. Until Smell-O-Vision happens, the next step might be feeling the wind in your hair. Perhaps dad racing a sportbike or kids giggling on a rollercoaster. Not as hard to build as you might think, you probably have the parts already.
Off-the-shelf devices serve up the seeing and hearing part of your imaginary environment, but they stop there. [Jared] wanted to take the immersion farther by being able to feel the speed, which meant building his own high power wind generator and tying it into the VR system. The failed crowdfunding effort of the “Petal” meant that something new would have to be constructed. Obviously, to move air without actually going on a rollercoaster requires a motor controller and some fans. Powerful fans.
A proponent of going big or going home, [Jared] picked up a pair of fans and modified them so heavily that they will launch themselves off of the table if not anchored down. Who overdrives fans so hard they need custom heatsinks for the motors? He does. He admits he went overboard and sensibly way overbudget for most people but he built it for himself and does not care.
Continue reading ““Superfan” Gaming Peripheral Lets You Feel Your Speed”
Many of us have gone on a stationary romp through some virtual or augmented scape with one of the few headsets out in the wild today. While the experience of viewing a convincing figment of reality is an exciting sensation in itself, [Mark Lee] and [Kevin Wang] are figuring out how to tie other senses into the mix.
The duo from Cornell University have built a mechanical exoskeleton that responds to light with haptic feedback. This means the wearer can touch the sphere of light around a source as if it were a solid object. Photo resistors are mounted like antenna to the tip of each finger, which they filed down around the edges to receive a more diffused amount of light. When the wearer of the apparatus moves their hand towards a light source, the sensors trigger servo motors mounted on the back of the hand to actuate and retract a series of 3D printed tendons which arch upward and connect to the individual fingers of the wearer. This way as the resistors receive varying amounts of light, they can react independently to simulate physical contours.
One of the goals of the project was to produce a working proof of concept with no more than 100 dollars worth of materials, which [Mark] and [Kevin] achieve with some cash to spare. Their list of parts can be found on their blog along with some more details on the project.
Continue reading “Touching Light with Haptic Feedback”