Think of Virtual Reality and it’s mostly fun and games that come to mind. But there’s a lot of useful, real world applications that will soon open up exciting possibilities in areas such as medicine, for example. [Victor] from the Shackspace hacker space in Stuttgart built an Augmented Reality Ultrasound scanning application to demonstrate such possibilities.
But first off, we cannot get over how it’s possible to go dumpster diving and return with a functional ultrasound machine! That’s what member [Alf] turned up with one day. After some initial excitement at its novelty, it was relegated to a corner gathering dust. When [Victor] spotted it, he asked to borrow it for a project. Shackspace were happy to donate it to him and free up some space. Some time later, [Victor] showed off what he did with the ultrasound machine.
As soon as the ultrasound scanner registers with the VR app, possibly using the image taped to the scan sensor, the scanner data is projected virtually under the echo sensor. There isn’t much detail of how he did it, but it was done using Vuforia SDK which helps build applications for mobile devices and digital eye wear in conjunction with the Unity 5 cross-platform game engine. Check out the video to see it in action.
Thanks to [hadez] for sending in this link.
Continue reading “Augmented Reality Ultrasound”
[Alvaro Ferrán Cifuentes] has built the coolest motion capture suit that we’ve seen outside of Hollywood. It’s based on tying a bunch of inertial measurement units (IMUs) to his body, sending the data to a computer, and doing some reasonably serious math. It’s nothing short of amazing, and entirely doable on a DIY budget. Check out the video below the break, and be amazed.
Cellphones all use IMUs to provide such useful functions as tap detection and screen rotation information. This means that they’ve become cheap. The ability to measure nine degrees of freedom on a tiny chip, for chicken scratch, pretty much made this development inevitable, as we suggested back in 2013 after seeing a one-armed proof-of-concept.
But [Alvaro] has gone above and beyond. Everything is open source and documented on his GitHun. An Arduino reads the sensor boards (over multiplexed I2C lines) that are strapped to his limbs, and send the data over Bluetooth to his computer. There, a Python script takes over and passes the data off to Blender which renders a 3D model to match, in real time.
All of this means that you could replicate this incredible project at home right now, on the cheap. We have no idea where this is heading, but it’s going to be cool.
Continue reading “Amazing IMU-based Motion Capture Suit Turns You Into a Cartoon”
It’s a parent’s worst nightmare. Doctors tell you that your baby is sick and there’s nothing they can do. Luckily though, a combination of hacks led to a happy ending for [Teegan Lexcen] and her family.
When [Cassidy and Chad Lexcen]’s twin daughters were born in August, smaller twin [Teegan] was clearly in trouble. Diagnostics at the Minnesota hospital confirmed that she had been born with only one lung and half a heart. [Teegan]’s parents went home and prepared for the inevitable, but after two months, she was still alive. [Cassidy and Chad] started looking for second opinions, and after a few false starts, [Teegan]’s scans ended up at Miami’s Nicklaus Children’s Hospital, where the cardiac team looked them over. They ordered a 3D print of the scans to help visualize possible surgical fixes, but the 3D printer broke.
Not giving up, they threw [Teegan]’s scans into Sketchfab, slapped an iPhone into a Google Cardboard that one of the docs had been playing with in his office, and were able to see a surgical solution to [Teegan]’s problem. Not only was Cardboard able to make up for the wonky 3D printer, it was able to surpass it – the 3D print would only have been the of the heart, while the VR images showed the heart in the context of the rest of the thoracic cavity.[Dr. Redmond Burke] and his team were able to fix [Teegan]’s heart in early December, and she should be able to go home in a few weeks to join her sister [Riley] and make a complete recovery.
We love the effect that creative use of technology can have on our lives. We’ve already seen a husband using the same Sketchfab tool to find a neurologist that remove his wife’s brain tumor. Now this is a great example of doctors doing what it takes to better leverage the data at their disposal to make important decisions.
Virtual reality could be the next big thing in the gaming world. And while VR displays and headsets are getting more and more sophisticated, many are forgetting perhaps the biggest feature VR will need to succeed — haptic feedback. [Pedro Lopes], [Alexandra Ion] and [Prof. Patric Baudisch] from the Hasso Plattner Institute is hoping to change that, with a project called Impacto: Simulating Physical Impact by Combining Tactile with Electrical Muscle Stimulation.
We’ve covered lots of haptic feedback devices over the past few years — some use mini gyros to simulate resistance, others blow air on you, but this is the first time we’ve seen one that combines muscle stimulation to really cause a physical effect.
They’re using an Oculus rift, and a Microsoft Kinect to perform the research. For their demonstration they use a basic boxing game that allows the user to feel the computer’s punches — but don’t worry, it doesn’t hit that hard!
Continue reading “Being Hit in the Virtual World”
[Florian] is hyped for Google Cardboard, Oculus Rifts, and other head mounted displays, and with that comes an interest in lenses. [Floian] wanted to know if it was possible to create these lenses with a 3D printer. Why would anyone want to do this when these lenses can be had from dozens of online retailers for a few dollars? The phrase, ‘because I can’ comes to mind.
The starting point for the lens was a CAD model, a 3D printer, and silicone mold material. Clear casting resin fills the mold, cures, and turns into a translucent lens-shaped blob. This is the process of creating all lenses, and by finely sanding, polishing, and buffing this lens with grits ranging from 200 to 7000, this bit of resin slowly takes on an optically clear shine.
Do these lenses work? Yes, and [Florian] managed to build a head mounted display that can hold an iPhone up to his face for viewing 3D images and movies. The next goal is printing prescription glasses, and [Florian] seems very close to achieving that dream.
The last time we saw home lens making was more than a year ago. Is anyone else dabbling in this dark art? Let us know in the comments below and send in a tip if you have a favorite lens hack in mind.
Straight from the Max Planck Institute for Biological Cybernetics, and displayed at this year’s Driving Simulation Conference & Exhibition is the coolest looking simulation platform we’ve ever seen. It’s a spherical (or icosahedral) roll cage, attached to the corners of a building by cables. With the right kinematics and some very heavy-duty hardware, this simulation platform has three degrees of translation, three degrees of rotation, and thousands of people that want to drive a virtual car or pilot a virtual plane with this gigantic robot.
The Cable Robot Simulator uses electric winches attached to the corners of a giant room to propel a platform with 1.5g of acceleration. The platform can move back and forth, up and down, and to and fro, simulating what a race car driver would feel going around the track, or what a fighter pilot would feel barreling through the canyons of the Mojave. All you need for a true virtual reality system is an Oculus Rift, which the team has already tested with driving and flight simulation programs
An earlier project by the same research group accomplished a similar feat in 2013, but this full-motion robotic simulator was not made of cable-based robotics. The CyberMotion Simulator used a robotic arm with a cockpit of sorts attached to the end of the arm. Inside the cockpit, stereo projectors displayed a wide-angle view, much like what a VR display does. In terms of capability and ability to simulate different environments, the CyberMotion Simulator may be a little more advanced; the Cable Robot Simulator cannot rotate more than about sixty degrees, while the CyberMotion Simulator can turn you upside down.
The Cable Robot Simulator takes up a very large room, and requires some serious engineering – the cables are huge and the winches are very powerful. These facts don’t preclude this technology being used in the future, though, and hopefully this sort of tech will make its way into a few larger arcades.
We often see concepts come in waves. Earlier this week we featured a cable robot used to move pallets around a warehouse.
Continue reading “Cables And Winches Become An Awesome Simulator”
If there’s one thing about laser cutters that makes them a little difficult to use, it’s the fact that it’s hard for a person to interact with them one-on-one without a clunky computer in the middle of everything. Granted, that laser is a little dangerous, but it would be nice if there was a way to use a laser cutter without having to deal with a computer. Luckily, [Anirudh] and team have been working on solving this problem, creating a laser cutter that can interact directly with its user.
The laser cutter is tied to a visual system which watches for a number of cues. As we’ve featured before, this particular laser cutter can “see” pen strokes and will instruct the laser cutter to cut along the pen strokes (once all fingers are away from the cutting area, of course). The update to this system is that now, a user can import a drawing from a smartphone and manipulate it with a set of physical tokens that the camera can watch. One token changes the location of the cut, and the other changes the scale. This extends the functionality of the laser cutter from simply cutting at the location of pen strokes to being able to cut around any user-manipulated image without interacting directly with a computer. Be sure to check out the video after the break for a demonstration of how this works.
Continue reading “Update: What You See Is What You Laser Cut”