VR headsets have been seeing new life for a few years now, and when it comes to head-mounted displays, the field of view (FOV) is one of the specs everyone’s keen to discover. Valve Software have published a highly technical yet accessibly-presented document that explains why Field of View (FOV) is a complex thing when it pertains to head-mounted displays. FOV is relatively simple when it comes to things such as cameras, but it gets much more complicated and hard to define or measure easily when it comes to using lenses to put images right up next to eyeballs.
The document goes into some useful detail about head-mounted displays in general, the design trade-offs, and naturally talks about the brand-new Valve Index VR headset in particular. The Index uses proprietary lenses combined with a slight outward cant to each eye’s display, and they explain precisely what benefits are gained from each design point. Eye relief (distance from eye to lens), lens shape and mounting (limiting how close the eye can physically get), and adjustability (because faces and eyes come in different configurations) all have a role to play. It’s a situation where every millimeter matters.
If there’s one main point Valve is trying to make with this document, it’s summed up as “it’s really hard to use a single number to effectively describe the field of view of an HMD.” They plan to publish additional information on the topics of modding as well as optics, so keep an eye out on their Valve Index Deep Dive publication list.
What does it smell like when the wheels heat up on that Formula 1 car you drive at night and on the weekends? You have no idea because the Virtual Reality experience that lets you do so doesn’t come with a nasal component. Yet.
Shown here is an olfactory device that works with Oculus Rift and other head-mounted displays. The proof of concept is hte work of [Kazuki Hashimoto], [Yosuke Maruno], and [Takamichi Nakamoto] and was shown of at last year’s IEEE VR conference. It lets the wearer smell the oranges when approaching a tree in a virtual environment. In other words, it makes your immersive experience smelly.
As it stands this a pretty cool little device which atomizes odor droplets while a tiny fan wafts them to the wearer’s nose. There is a paper which presumably has more detail but it’s behind a pay wall so for now check out the brief demo video below. Traditionally an issue with scent systems is the substance stuck in the lines, which this prototype overcomes with direct application from the reservoir. Yet to be solved is the availability for numerous different scents.
This build came to our attention via an UploadVR article that does a good job of covering some of the scent-based experiments over the years. They see some of the same hurdles we do: odors linger and there is a limited palette that can be produced. We assume the massive revenue of the gaming industry is going to drive research in this field, but we won’t be lining up to smell gunpowder and dead bodies (or rotting zombies) anytime soon.
The more noble effort is in VR applications like taking the elderly and immobile back for another tour of places they’ll never again be able to visit in their lives. Adding the sense of smell, which has the power to unlock so many memories, makes that use case so much more powerful. We think that’s something everyone can be hopeful about!
[Harris Shallcross] decided to build a pair of smart glasses and recently completed a first prototype of his project ‘Ochi’ – an STM32 based, BLE-connected, OLED eyeglass display. There are of course several homebrew smart glasses projects out there; many are more polished-looking and nearly all of them also display information from a smartphone over Bluetooth. This one is interesting partly because it highlights many of the design challenges that smart glasses and other near-eye displays face. It also demonstrates the iterative development process: begin by getting something working to learn what does and doesn’t cut it at a basic level, and don’t optimize prematurely; let the process bring problems to the surface.
For his project, [Harris Shallcross] used a small 0.95″ diagonal 96×64 color OLED as the display. The lens is from a knockoff Google Cardboard headset, and is held in a 3D printed piece that slides along a wire rail to adjust focus. The display uses a custom font and is driven by an STM32 microcontroller on a small custom PCB, with an HM11 BLE module to receive data wirelessly. Power is provided by a rechargeable lithium-ion battery with a boost converter. An Android app handles sending small packets of data over Bluetooth for display. The prototype software handles display of time and date, calendar, BBC news feed, or weather information.
Devices like these have a lot to deal with. Weight and distribution of that weight is a concern, the size and comfort of the optics is important, and displaying data on a small OLED is only part of the battle – choosing what information to display and when are vital to the device being actually useful in any way, otherwise it’s just a tech demo.
This project set out to show whether it was possible to use the parts listed to make a glasses mounted smart display that was at least somewhat functional, and the software to support it. Clearly, [Harris Shallcross] succeeded at that, but what really showcases the development process is his list of improvements – what he decided needs to go into a second version, and why. One of those goals is to improve the optics; perhaps there’s something to learn from The $60 Bluetooth Head Mounted Display project, which used a similar OLED and a prism to locate the display off to the side instead of in front.
For [Tony]’s entry for The Hackaday Prize, he’s doing something we’ve all seen before – a head mounted display, connected to a Bluetooth module, displaying information from a smartphone. What we haven’t seen before is a cheap version of this tech, and a version of Google Glass that folds – you know, like every other pair of glasses on the planet – edges this project over from ‘interesting’ to ‘nearly practical’.
For the display, [Tony] is using a 0.96″ OLED connected to an Arduino Nano. This screen is directed into the wearer’s eye with a series of optics that, along with every other part of the frame, was 3D printed on a Solidoodle 2. The frame itself not only folds along the temples, but also along the bridge, making this HMD surprisingly compact when folded up.
Everything displayed on this head mounted display is controlled by either an Android phone or a Bluetooth connection to a desktop. Using relatively simple display means [Tony] is limited to text and extremely simple graphics, but this is more than enough for some very interesting applications; reading SMS messages and checking email is easy, and doesn’t overpower the ‘duino.
The project featured in this post is an entry in The Hackaday Prize. Build something awesome and win a trip to space or hundreds of other prizes.
MTBS3D forum user [Rfurlan] pledged in the oculus rift kickstarter (which concluded last night), but simply couldn’t wait till November/December to get his developer kit. That, and he’s probably only getting one, and who can live with only one? Since [Palmer], the creator of the oculus rift has been very open about parts, [Rfurlan] was able to compile build instructions for your very own Oculus Rift! Keep in mind though, this is only the immersive display, not the tracking component. It is also, possibly not exactly the same as the oculus, but rather the same as a recent prototype.
At one point he was having issues finding the correct lenses and [Palmer] jumped in to make some suggestions to keep things going. That’s the kind of enthusiasm that we love to see from an innovator, even when he’s in the middle of a kickstarted for the very item that [Rfurlan] is creating. This is a testament to the VR community.
Lets take a look at what makes this thing tick, and why it is such a big deal.
If you’ve been following along with immersive gaming, even casually, you’ve probably considered the difficulty in trying to do a comfortable and believable “walk” in a game. The first thing that usually pops into peoples minds are Omni Directional Treadmills, or ODTs. There are many problems with these, one of the biggest simply being cost. They’re very expensive.
[Zalo] at the MTBS3d forums has been working on his own very cost effective solution called the “Simulacrum”. He has built this for under $100 and it allows for a walking motion to be translated into the game. As you can see in the video below it works fairly well, even when one is out of commission for repairs (hence the limp).
Watch out, these sunglasses are actually a head mounted display. [Staffan] says he’s wanted dataglasses since ’95, but whats currently out there makes the user look ridiculous, and we have to agree. While his forum posts are a little lacking in detail, he’s promised us more info soon. And for now lets us know at least the resolution, well sort of: Its either 480×1280 or 480x427x3, you can be the judge. Update: [Staffan] has clarified “The resolution is 480*1280 true pixels. It is accomplished by spanning the screen across two Kopin CyberDisplay VGA modules.”
Regardless, [Staffan] is looking for help perfecting the glasses, with what in particular we’re not sure, but the project looks promising and we hope he keeps up the good work.