For a final design project, [Frank] and his group took on an augmented reality project. The goal was to make objects interactively controllable by pointing a smartphone at them. Their solution was Augmented Reality Universal Controller and Identifier (ARUCI).
The system locates controllable objects by sensing IR beacons that contain identifiers for each object. The IR is received by a Wiimote sensor, which has been integrated into a custom PCB. This board sits in a 3D printed enclosure, and mounts to the back of a smartphone. The electronics are powered by tapping off of the phone’s battery.
Commands are sent to devices using a custom 2.4 GHz protocol which was implemented using the ATmega128RFA1. Each device has another ATmega to receive the signal and control the real world object. In their demo, the group shows the system controlling devices including a TV, a radio, and an RC car.
The system provides an interesting way to interact with objects, and the hardware integration is quite impressive. After the break, watch [Frank] give a demo.
Continue reading “IR Based Augmented Reality”
Over Thanksgiving, [Greg] had a little time on his hands and decided he needed an afternoon project. Having a few bits of plywood, an xacto knife, and some blue paint on hand meant a miniature TARDIS would take shape on his workbench. After finishing the model, [Greg] continued improving it with a blinky LED when the thought of adding an interior to the TARDIS entered his mind. An idea too good to pass up, really.
The TARDIS, of course, is smaller on the outside, so [Greg] needed a way to virtually model the interior of  and ’s home. After playing around with Blender for a few days, [Greg] had a reasonable 3D facsimile of the TARDIS interior. Now the only problem was to display it behind the front door.
[Greg] whipped up a small app for his phone that reads a zebra print pattern behind the door and overlays the 3D modeled TARDIS interior. Yes, it’s only viewable through augmented reality, but tilting the desktop TARDIS from side to side makes the entire console room visible. You can check out [Greg]’s TARDIS interior in the video after the break.
Continue reading “Making a TARDIS bigger on the inside”
[Kevin Mellott’s] take on the VFX1 was to update it so it can be used with modern computers requiring just a USB socket and VGA feed.
The VFX1 is a Virtual Reality Headset that hit the market in the first half of the 90’s. The headset was the first of its kind to hit the home market and was ahead of its time. The VFX1 was developed and marketed by Forte Technologies, who’s assets where purchased in 1997 by Vuzix who now produce modern day Video glasses with optional tracking system.
What [Kevin] has achieved is nothing more than remarkable. The original system required a massive ISA card and a link from this card to the Feature Connector on the display card. [Kevin] did away with the ISA card and FCON replacing it with what he calls the LinkBox. This LinkBox has serial or USB out and accepts stereo/mono VGA input or RGB.
The system can now be used with modern day computers including laptops. Those into VR should really check this out.
Continue reading “VFX1 Virtual Reality Headset LinkBox”
There are so many good ideas crammed into this project its hard for us to believe this isn’t already widely used for critical welding applications. Traditional welding masks simply filter out light to protect the welder’s eyes. This mask doesn’t have a window in it at all. Instead, the mask includes two cameras on the outside and two LCD screens on the inside. It filters light by processing the video which lends itself to that grab-bag of features we mentioned earlier.
Possibly the best of the system is its ability to selectively filter the brightness of the weld. What this means is that areas outside of the welding arc appear at a normal brightness level, whereas before they would have been greatly dimmed. A demonstration of augmented reality is also shown, where a computer monitors the welding surface, giving the welder a target to follow and measuring the distance between the weld and the filament. The video mentions that an FPGA would be well suited for the image processing, making us think this could be produced at a reasonable cost. After all, they already use X-ray machines for some welds, we’d bet a set of these helmets could be supplied to a crew at a similar cost.
Lonely? Bored? Really into J-pop? If you’re any of these things, here’s the build for you. It’s an augmented reality system that allows you to go on a date with one of Japan’s most popular virtual singers.
The character chosen to show off this augmented reality girlfriend tech is [Hatsune Miku], a voice synthesizer personified as a doll-eyed anime avatar. [Miku] is an immensely popular character in Japan, with thousands of people going to her concerts, so choosing her for this augmented reality girlfriend project was an obvious choice.
The build details for this hack are a little sparse, confounded by the horrible Google Translate results of the blog linked in the YouTube description. From what we can gather from the video and this twitter account, the build is based on an ASUS Xtion Kinect clone and a nice pair of video goggles.
We’re expecting the comments for this post to fill up with, ‘Japan is really weird’ comments, but we can see a few very, very cool applications of this tech. For instance, think how cool it would be to be guided around a science museum by [Einstein], or around Philadelphia by [Ben Franklin].
[Bastian] sent in a coffee table he built. This isn’t a place to set your drinks and copies of Make, though: it’s a multitouch table with a 3D display. Since no description can do this table justice, take a look at the video.
The build was inspired by the subject of this Hackaday post where [programming4fun] was able to build a ‘holographic display’ using a regular 2D projector and a Kinect. Both builds work on the principle of redrawing the 3D space in relation to the user’s head – as [Bastian] moves his head around the coffee table, the Kinect tracks his location and moves the 3 dimensional grid of boxes in the opposite direction. It’s extremely clever, and looks to be a promising user interface.
In addition to a Kinect, the coffee table uses a Microsoft Surface-like display; four infrared lasers are placed at the corner and detected with a camera next to the projector in the base.
After the break you can see the demo video and a gallery of the images [Bastion] put up on the NUI group forum.
Continue reading “Multitouch table uses a Kinect for a 3D display”
By now we’re assuming you are all familiar with Google’s “Project Glass”, an ambitious augmented reality project for which they revealed a promotional video last week. [Will Powell] saw the promo vid and was so inspired that he attempted to rig up a demo of Project Glass for himself at home.
While it might seem like a daunting project to take on, [Will] does a lot of work with Kinect-based augmented reality, so his Vuzix/HD webcam/Dragon Naturally Speaking mashup wasn’t a huge step beyond what he does at work. As you can see in the video below, the interface he implemented looks very much like the one Google showed off in their demo, responding to his voice commands in a similar fashion.
He says that the video was recorded in “real time”, though there are plenty of people who debate that claim. We’re guessing that he recorded the video stream fed into the Vuzix glasses rather than recording what was being shown in the glasses, which would make the most sense.
We’d hate to think that the video was faked, mostly because we would love to see Google encounter some healthy competition, but you can decide for yourself.
Continue reading “DIY “Project Glass” clone looks almost too good to be true”