Historically, the capabilities of real world humanoid robots have trailed far behind their TV and movie counterparts. But roboticists kept pushing state of the art forward, and Boston Dynamics just shared a progress report: their research platform Atlas can now complete a two-robot parkour routine.
Watching the minute-long routine on YouTube (embedded after the break) shows movements more demanding than their dance to the song “Do You Love Me?“ And according to Boston Dynamics, this new capability is actually even more impressive than it looks. Unlike earlier demonstrations, this routine used fewer preprogrammed motions that made up earlier dance performances. Atlas now makes more use of its onboard sensors to perceive its environment, and more of its onboard computing power to decide how to best move through the world on a case-by-case basis. It also needed to string individual actions together in a continuous sequence, something it had trouble doing earlier.
Such advances are hard to tell from a robot demonstration video, which are frequently edited and curated to show highlighted success and skip all the (many, many) fails along the way. Certainly Boston Dynamics did so themselves before, but this time it is accompanied by almost six minutes worth of behind-the-scenes footage. (Also after the break.) We see the robot stumbling as it learned, and the humans working to put them back on their feet.
Humanoid robot evolution has not always gone smoothly (sometimes entertainingly so) but Atlas is leaps and bounds over its predecessors like Honda Asimo. Such research finds its way to less humanoid looking robots like the Stretch. And who knows, maybe one day real robots will be like their TV and movie counterparts that have, for so long, been played by humans inside costumes.
Continue reading “Boston Dynamics Atlas Dynamic Duo Tackles Obstacle Course”
Boston Dynamics has always built robots with agility few others could match. While great for attention-getting demos, from outside the company it hasn’t been clear how they’ll translate acrobatic skills into revenue. Now we’re getting a peek at a plan in an interview with IEEE Spectrum about their new robot Stretch.
Most Boston Dynamics robots have been research projects, too expensive and not designed for mass production. The closest we got to date was Spot, which was offered for sale and picked up a few high profile jobs like inspecting SpaceX test sites. But Spot was still pretty experimental without an explicit application. In contrast, Stretch has a laser-sharp focus made clear by its official product page: this robot will be looking for warehouse jobs. Specifically, Stretch is designed to handle boxes up to 50 lbs (23 kg). Loading and unloading them, to and from pallets, conveyer belts, trucks, or shipping containers. These jobs are repetitive and tedious back-breaking work with a high injury rate, a perfect opportunity for robots.
But warehouse logistics aren’t as tightly structured as factory automation, demanding more adaptability than typical industrial robots can offer. A niche Boston Dynamics learned it can fill after releasing an earlier demo video showing their research robot Atlas moving some boxes around: they started receiving inquiries into how much that would cost. Atlas is not a product, but wheels were set in motion leading to their Handle robot. Learning from what Handle did well (and not well) in a warehouse environment, the designed evolved to today’s Stretch. The ostrich-like Handle prototype is now relegated to further research into wheeled-legged robots and the occasional fun dance video.
The Stretch preproduction prototypes visible in these videos lacks acrobatic flair of its predecessors, but they still have the perception and planning smarts that made those robots possible. Those skills are just being applied to a narrower problem scope. Once production models are on the job, we look forward to reading some work performance reviews.
Continue reading “Boston Dynamics Stretch Robot Trades Lab Coat For Work Uniform”
The uses of artificial intelligence and machine learning continue to expand, with one of the more recent implementations being video processing. A new method can “fill in” frames to smooth out the appearance of the video, which [LegoEddy] was able to use this in one of his animated LEGO movies with some astonishing results.
His original animation of LEGO figures and sets was created at 15 frames per second. As an animator, he notes that it’s orders of magnitude more difficult to get more frames than this with traditional methods, at least in his studio. This is where the artificial intelligence comes in. The program is able to interpolate between frames and create more frames to fill the spaces between the original. This allowed [LegoEddy] to increase his frame rate from 15 fps to 60 fps without having to actually create the additional frames.
While we’ve seen AI create art before, the improvement on traditionally produced video is a dramatic advancement. Especially since the AI is aware of depth and preserves information about the distance of objects from the camera. The software is also free, runs on any computer with an appropriate graphics card, and is available on GitHub.
Continue reading “Boost Your Animation To 60 FPS Using AI”