Use Machine Learning To Identify Superheroes And Other Miscellany

[Massimiliano Patacchiola] writes this handy guide on using a histogram intersection algorithm to identify different objects. In this case, lego superheroes. All you need to follow along are eyes, Python, a computer, and a bit of machine learning magic.

He gives a good introduction to the idea. You take a histogram of the colors in a properly cropped and filtered photo of the object you want to identify. You then feed that into a neural network and train it to identify the different superheroes by color. When you feed it a new image later, it will compare the new image’s histogram to its model and output confidences as to which set it belongs.

This is a useful thing to know. While a lot of vision algorithms try to make geometric assertions about the things they see, adding color to the mix can certainly help your friendly robot project recognize friend from foe.

 

Train Your Robot To Walk With A Neural Network

[Basti] was playing around with Artificial Neural Networks (ANNs), and decided that a lot of the “hello world” type programs just weren’t zingy enough to instill his love for the networks in others. So he juiced it up a little bit by applying a reasonably simple ANN to teach a four-legged robot to walk (in German, translated here).

While we think it’s awesome that postal systems the world over have been machine sorting mail based on similar algorithms for years now, watching a squirming quartet of servos come to forward-moving consensus is more viscerally inspiring. Job well done! Check out the video embedded below.

Continue reading “Train Your Robot To Walk With A Neural Network”

Otto Bot Is Bob’s Grandson

The Otto DIY robot has just taken first place in the coveted role as “best robot to 3D print for your (inner) child”. It’s cute, it dances, it doesn’t cost too much, it’s completely open source, and it’s not impossible to write code for. It’s probably the most refined Bob design that we’ve seen yet. Watch it move in the video below.

Continue reading “Otto Bot Is Bob’s Grandson”

NABiRoS sideways walking robot

This Robot Goes Through Life Sideways

We humans walk funny. Pivoting one leg forward at the hip creates an offset that puts us off-balance sideways. We have to compensate for this with each step we take. Many robots handle this by instead taking small, calculated steps. Enter NABiRoS, the Non Anthropomorphic Bipedal Robot System (link to the video below). The ‘Non Anthropomorphic’ means that it doesn’t walk like a human, and yet the ‘Bipedal’ means it still walks on two legs. The difference is that it walks sideways.

NABiRoS leg configuration
NABiRoS leg configuration (from video)

Here’s how the folks from RoMeLa (Robotics & Mechanisms Laboratory) at UCLA did it. Imagine you rotated both your legs 90 degrees such that they were facing in opposite directions. Then you rotate your upper body 90 degrees to face one of your legs. You can now move your legs to walk in the direction you’re facing and there’ll be no more tilting sideways each time you take a step. The joints are also simpler as only a single degree of freedom is needed in each of the knee and hip joints. The ankles and feet are done with a compliant, or an elastic, joint much as you see with a lot of prosthetic legs. As you can see in the video below, in addition to walking, they can do some surprisingly active things such as hopping up and down and what we can only call skipping. In fact, the result is sometimes very human.

Continue reading “This Robot Goes Through Life Sideways”

Floating Walking Robot

It’s no secret that we love bizarre robot locomotion, so we are naturally suckers for BALLU (YouTube link, also embedded below) the Bouyancy-Assisted Lightweight Legged Unit. The project started with a simple observation — walking robots are constrained by having to hold themselves up — and removing that constraint make success much easier. Instead of walking, BALLU almost floats and uses what little net weight it does have to push against the ground.

Continue reading “Floating Walking Robot”

Heatmap of vacuum cleaning robot

A Glimpse Into The Mind Of A Robot Vacuum Cleaner

What’s going through the mind of those your autonomous vacuum cleaning robots as they traverse a room? There are different ways to find out such as covering the floor with dirt and seeing what remains afterwards (a less desirable approach) or mounting an LED to the top and taking a long exposure photo. [Saulius] decided to do it by videoing his robot with a fisheye lens from near the ceiling and then making a heatmap of the result. Not being satisfied with just a finished photo, he made a video showing the path taken as the room is being traversed, giving us a glimpse of the algorithm itself.

Looking down on the room and robot
Looking down on the room and robot

The robot he used was the Vorwerk VR200 which he’d borrowed for testing. In preparation he cleared the room and strategically placed a few obstacles, some of which he knew the robot wouldn’t get between. He started the camera and let the robot do its thing. The resulting video file was then loaded into some quickly written Python code that uses the OpenCV library to do background subtraction, normalizing, grayscaling, and then heatmapping. The individual frames were then rendered into an animated gif and the video which you can see below.

Continue reading “A Glimpse Into The Mind Of A Robot Vacuum Cleaner”