After you’ve taken a moment to ponder the turn of phrase used in the title, take a look at this scratch-built robotic vacuum created by [theking3737]. The entire body of the vacuum was 3D printed, and all of the internal electronics are off-the-shelf modular components. We can’t say how well it stacks up against the commercial equivalents from iRobot and the like, but it doesn’t look like it would be too hard to build one yourself to find out.
The body of this rather concerned-looking robot was printed on a DMS DP5 printer, which is a neat trick as it only has a build platform of 200 mm x 200 mm. Once all the pieces were printed, a 3D pen was used to “weld” the sections together. The final result looks a bit rough, but should give a bond that’s just as strong as the printed parts themselves.
The robot has four sets of ultrasonic range finders to detect walls and obstacles, though probably not in the positions you would expect. The right side of the robot has two sets of sensors, while the left side only gets one. We aren’t sure the reasoning behind the asymmetrical layout, but presumably the machine prefers making right turns.
Control is provided by an Arduino Mega and the ever-reliable HC-05 Bluetooth module. A companion Android application was written which allows configuring the robot without having to plug into the Arduino every time you want to tweak a setting.
[madcowswe] starts by pointing out that the entire premise of ODrive (an open-source brushless motor driver board) is to make use of inexpensive brushless motors in industrial-type applications. This usually means using hobby electric aircraft motors, but robotic applications sometimes need more torque than those motors can provide. Adding a gearbox is one option, but there is another: so-called “hoverboard” motors are common and offer a frankly outstanding torque-to-price ratio.
A teardown showed that the necessary mechanical and electrical interfacing look to be worth a try, so prototyping has begun. These motors are really designed for spinning a tire on the ground instead of driving other loads, but [madcowswe] believes that by adding an encoder and the right fixtures, these motors could form the basis of an excellent robot arm. The ODrive project was a contender for the 2016 Hackaday Prize and we can’t wait to see where this ends up.
Almost two years ago, a research team showed that it was possible to get fine motor control from cheap, brushless DC motors. Normally this is not feasible because the motors are built-in such a way that the torque applied is not uniform for every position of the motor, a phenomenon known as “cogging”. This is fine for something that doesn’t need low-speed control like a fan motor, but for robotics it’s a little more important. Since that team published their results, though, we are starting to see others implement their own low-speed brushless motor controllers.
The new method of implementing anti-cogging maps out the holding torque required for any position of the motor’s shaft so this information can be used later on. Of course this requires a fair amount of calibration; [madcowswe] reports that this method requires around 5-10 minutes of calibration. [madcowswe] also did analysis of his motors to show how much harmonic content is contained in these waveforms, which helps to understand how this phenomenon arises and how to help eliminate it.
While [madcowswe] plans to add more features to this motor control algorithm such as reverse-mapping, scaling based on speed, and better memory usage, it’s a good implementation that has visible improvements over the stock motors. The original research is also worth investigating if a cheaper, better motor is something you need.
Here on Hackaday, too often do we turn our heads and gaze at the novelty of 3D printing functional devices. It’s easy to forget that other techniques for assembling functional prototypes exist. Here, [Reuben] nails the aspect of functional prototyping with the laser cutter with a real-world application: a roll-pitch friction differential drive built from just off-the shelf and laser-cut parts!
The centerpiece is held together with friction, where both the order of assembly and the slight wedged edge made from the laser cutter kerf keeps the components from falling apart. Pulleys transfer motion from the would-be motor mounts, where the belts are actually tensioned with a roller bearing mechanism that’s pushed into position. Finally, the friction drive itself is made from roller-blade wheels, where the torque transferred to the plate is driven by just how tightly the top screw is tightened onto the wheels. We’d say that [Reuben] is pushing boundaries with this build–but that’s not true. Rather, he’s using a series of repeatable motifs together to assemble a both beautiful and complex working mechanism.
This design is an old-school wonder from 2012 uncovered from a former Stanford course. The legendary CS235 aimed to teach “unmechanically-minded” roboticists how to build a host of mechanisms in the same spirit as MIT’s How-to-make-almost-Anything class. While CS235 doesn’t exist anymore, don’t fret. [Reuben] kindly posted his best lectures online for the world to enjoy.
Robot design traditionally separates the body geometry from the mechanics of the gait, but they both have a profound effect upon one another. What if you could play with both at once, and crank out useful prototypes cheaply using just about any old 3D printer? That’s where Interactive Robogami comes in. It’s a tool from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) that aims to let people design, simulate, and then build simple robots with a “3D print, then fold” approach. The idea behind the system is partly to take advantage of the rapid prototyping afforded by 3D printers, but mainly it’s to change how the design work is done.
To make a robot, the body geometry and limb design are all done and simulated in the Robogami tool, where different combinations can have a wild effect on locomotion. Once a design is chosen, the end result is a 3D printable flat pack which is then assembled into the final form with a power supply, Arduino, and servo motors.
A white paper is available online and a demonstration video is embedded below. It’s debatable whether these devices on their own qualify as “robots” since they have no sensors, but as a tool to quickly prototype robot body geometries and gaits it’s an excitingly clever idea.
Sudoku is a great way to pass some time, especially on a long flight. However, we don’t think the airlines will let [Sanahm] board with his sudoku-solving robot. The basic machine looks like a 2D plotter made with aluminum extrusion, with the addition of a Raspberry Pi and a camera. The machine can read a sudoku puzzle, solve it, and then fill in the puzzle with a pen. Unlike humans, it should never need to erase its work.
The software uses OpenCV to process the camera data, find the grid, and the cells provided by the puzzle. TensorFlow recognizes the numbers. From there, it is all just math to solve the puzzle. Once solved, the plotter part of the robot takes over and fills in the blanks. After all that, this seems like the easy part.
[Jeremy Cook]’s latest take on the Strandbeest, the ClearWalker, is ready to roll! He’s been at work on this project for a while, and walks us through the electronics and control system as well as final assembly tweaks. The ClearWalker is fully controllable and includes a pan and tilt camera as well as programmable LED segments, and even a tail.
When we last saw [Jeremy] at work on this design, it wasn’t yet functional. He showed us all the important design and assembly details that went into creating a motorized polycarbonate version of [Theo Jansen’s] classic Strandbeest design; there’s far more to the process than simply scaling parts up or down. Happily, [Jeremy] is able to show off the crystal clear beauty in his photo gallery as well as a new video, embedded below.