If you have ever had to complete a task such as building a LEGO model over a remote connection, you will know that the challenges are like an absurd grade school group project. The person giving directions often has trouble describing what they are thinking, and the person doing the work has trouble interpreting what the instructor wants. “Turn the blue block over. No, only half way. Go back. Now turn it. No, the other way. NO! Not clockwise, downward. That’s Upward! Geez. Are you even listening‽” Good times.
While you may not be in this situation every day, the Keio University of Japan has an intuitive way to give instructors a way to physically interact with an instructee through a Moore/Swayze experience. The instructor has a camera in typical pirate parrot placement over the shoulder. Two arms are controlled by the instructor who can see through stereoscopic cameras to have a first-person view from across the globe. This natural way to interact with the user’s environment allows muscle memory to pass from the instructor to the wearer.
For some of the other styles of telepresence, see this deep-sea bot and a cylindrical screen that looks like someone is beaming up directly from the holodeck.
Continue reading “Robots Invade Your Personal Space”
VR is in vogue, but getting on board requires a steep upfront cost. Hackaday.io user [Colin Pate] felt that $800 was a bit much for even the cheapest commercial 360-degree 3D camera, so he thought: ‘why not make my own for half that price?’
[Pate] knew he’d need a lot of bandwidth and many GPIO ports for the camera array, so he searched out the Altera Cyclone V SOC FPGA and a Terasic DE10-Nano development board to host it. At present, he has four Uctronics OV5642 cameras on his rig, chosen for their extensive documentation and support. The camera mount itself is a 3D-printed octagon so eight of the OC5642 can capture a full 360-degree photo.
Next: producing an image!
Continue reading “Immersive VR with a 200-Degree Stereoscopic Camera”
Watching robots doing sports is pretty impressive from a technical viewpoint, although we secretly smile when we compare these robots’ humble attempts to our own motoric skills. Now, a new robot named Robomintoner seeks to challenge human players, and it’s already darn good at badminton.
Continue reading “Robomintoner Badminton Bot To Defeat Amateur Humans”
This rig is something of a museum or art installation, but the concept is so simple we thought it could easily inspire your next project. The two mirrors and two video sources make up a stereoscopic display.
The user sits between two displays (computer monitors shown here, but the post also shows images projected on two walls of a room). A pair of mirrors mounted at forty-five degrees form the eye pieces. It’s a V-shaped mirror assembly in which the narrow end pointing toward the bridge of the user’s nose. The mirrors reflect the images from the monitors, giving a different view for each eye.
In this case each monitor is playing back a video loop, but one is just slightly longer than the other. Each monitor has a potentiometer in front of it. The user can turn them to speed or slow the playback in an attempt to bring the video back into sync. We don’t think we’d replicate that portion of the project. But it might be fun to view some stereoscopic clips in this way. There’s even instructions on how two cameras were used to record the scenes.
You can get a closer view of the test apparatus in the clip after the jump.
Continue reading “Stereoscopic display art installation”
This super quick hack will be fun to do with the kids. Remember the days of View-Masters? You’d put a disk of small slides into a little plastic viewer and a stereoscopic image would jump out at you in 3D! Now you can not only view stereoscopic images on your smartphone, but make your own too!
To shoot the images just hold your phone in portrait orientation and take a snapshot of your subject, then move the camera six inches to the right and take a second image. The two pics need to be displayed on the screen at the same time and for this [Plarky] uses a free iPhone app called Pic Stitch. We’re sure you can find an Android equivalent in no time if you do a bit of searching.
To view the stereoscope it helps to make a divider out of cardboard like the one seen above. You’ll need to cross your eyes and focus on a point to bring the two images together. We don’t remember having to do this with the View-Master so we’re hoping someone will take the idea and improve upon it. We’ve already seen a digital View-Master. Now we want to see those dual screens replaced with an iPhone cradle.
While the whole 3d movie/game craze seems to be ramping up, it really isn’t a new thing. We all recall those fancy red-blue glasses that were popular in theaters for a while, but I’m not talking about that. Passive 3d projection (using polarized glasses) has been around for a while too. Many people have figured out cheap ways to build these systems in their homes, but only recently have we seen media created for them in quantity. Now that you can buy 3D games and movies at your local box store, the temptation to have a 3d system in your home is much higher.
Here’s a great read on how to put together a fairly simple projection system that uses two identical projectors with polarizing filters. Basically, all you need are two projectors, two filters, a screen, and the glasses. There are plenty of tips for mounting and setup in the thread to help alleviate any headaches you might encounter.
This system is primarily used with a PC, because it requires two video feeds to function. A cost breakdown might make you wonder why you wouldn’t just jump on amazon and get a 32″ 3d tv for under $400, but sitting in front of that giant screen might make you understand.
Put your face close to the screen and cross your eyes until the two images above become one. You may need to adjust the tilt of your chin to make it happen, but when they come together you’ll see [John Lennon] pop out in 3D. This was made using a 3D rendering script for The Gimp.
The process is not entirely automatic, but it won’t take too long to mask off the outlines for different depth layers. The script makes three different layers from the image. One of them is a color-coded depth map that uses a custom color palatte to choose distance for each item. If you paint the background dark blue it will be processed at the furthest distance from the viewer’s cross-eyed perspective, yellow is the nearest.
[Don] mentions a parallel output and a cross-eyed output in his write up. We understand the cross-eyed version, but are just guessing that the parallel version would be used in a stereoscopic viewer that puts a partition between the two images so that each eye sees a different frame. You know, like a View-Master.