Rover has 3D printed 4.3:1 reduction planetary gearboxes embedded into each wheel, with off the shelf bearings and brushless motors. A Raspberry Pi sits in the driver’s seat, and the goal is to use a version of NVIDA’s TrailNet framework for GPS-free navigation of paths. As a result, [taylor] hopes to end up with a robotic “trail buddy” that can be made with off-the-shelf components and 3D printed parts.
Moving the motors and gearboxes into the wheels themselves makes for a very small main body to the robot, and it’s more than a bit strange to see the wheel spinning opposite to the wheel’s hub. Check out the video showcasing the latest development of the wheels, embedded below.
We like that the Weedinator Project is thinking big for this year’s Hackaday Prize! This ambitious project by [TegwynTwmffat] is building on a previous effort, which was a tractor mounted weeding machine (shown above). It mercilessly shredded any weeds; the way it did this was by tilling everything that existed between orderly rows of growing leeks. The system worked, but it really wasn’t accurate enough. We suspect it had a nasty habit of mercilessly shredding the occasional leek. The new version takes a different approach.
The new Weedinator will be an autonomous robotic rover using a combination of GPS and colored markers for navigation. With an interesting looking adjustable suspension system to help with fine positioning, the Weedinator will use various attachments to help with plant care. Individual weeds will be identified optically and sent to the big greenhouse in the sky via precise flame from a small butane torch. It’s an ambitious project, but [TegwynTwmffat] is building off experience gained from the previous incarnation and we’re excited to see where it goes.
A team of students in Antwerp, Belgium are responsible for Project Aslan, which is exploring the feasibility of using 3D printed robotic arms for assisting with and translating sign language. The idea came from the fact that sign language translators are few and far between, and it’s a task that robots may be able to help with. In addition to translation, robots may be able to assist with teaching sign language as well.
The project set out to use 3D printing and other technology to explore whether low-cost robotic signing could be of any use. So far the team has an arm that can convert text into finger spelling and counting. It’s an interesting use for a robotic arm; signing is an application for which range of motion is important, but there is no real need to carry or move any payloads whatsoever.
A single articulated hand is a good proof of concept, and these early results show some promise and potential but there is still a long ways to go. Sign language involves more than just hands. It is performed using both hands, arms and shoulders, and incorporates motions and facial expressions. Also, the majority of sign language is not finger spelling (reserved primarily for proper names or specific nouns) but a robot hand that is able to finger spell is an important first step to everything else.
Future directions for the project include adding a second arm, adding expressiveness, and exploring the use of cameras for the teaching of new signs. The ability to teach different signs is important, because any project that aims to act as a translator or facilitator needs the ability to learn and update. There is a lot of diversity in sign languages across the world. For people unfamiliar with signing, it may come as a surprise that — for example — not only is American Sign Language (ASL) related to French sign language, but both are entirely different from British Sign Language (BSL). A video of the project is embedded below.
Not only does the GuitarBot project show off some great design, but the care given to the documentation and directions is wonderful to see. The GuitarBot is an initiative by three University of Delaware professors, [Dustyn Roberts], [Troy Richards], and [Ashley Pigford] to introduce their students to ‘Artgineering’, a beautiful portmanteau of ‘art’ and ‘engineering’.
The GuitarBot It is designed and documented in a way that the three major elements are compartmentalized: the strummer, the brains, and the chord mechanism are all independent modules wrapped up in a single device. Anyone is, of course, free to build the whole thing, but a lot of work has been done to ease the collaboration of smaller, team-based groups that can work on and bring together individual elements.
Some aspects of the GuitarBot are still works in progress, such as the solenoid-activated chord assembly. But everything else is ready to go with Bills of Materials and build directions. An early video of a strumming test proof of concept used on a ukelele is embedded below.
Walkers like the Strandbeest are favorites due in part to their smooth design and fluid motion, but [Leandro] is going a slightly different way with Octo, an octopodal platform for exploring rough terrain. Octo is based on the Klann linkage which was developed in 1994 and intended to act as an alternative to wheels because of its ability to deal with rough terrain. [Leandro] made a small proof of concept out of soldered brass and liked the results. The next version will be larger, made out of aluminum and steel, and capable of carrying a payload.
The Strandbeest and Octo have a lot in common but differ in a few significant ways. Jansen’s linkage (which the Strandbeest uses) uses eight links per leg and requires relatively flat terrain. The Klann linkage used by Octo needs only six links per leg, and has the ability to deal with rougher ground.
[Leandro] didn’t just cut some parts out from a file found online; the brass proof of concept was drawn up based on an animation of a Klann linkage. For the next version, [Leandro] used a simulator to determine an optimal linkage design, aiming for one with a gait that wasn’t too flat, and maximized vertical rise of the leg to aid in clearing obstacles.
We’ve seen the Klann linkage before in a LEGO Spider-bot. We’re delighted to see [Leandro]’s Octo in the ring for the Wheels, wings, and walkers category of The Hackaday Prize.
[David Brown]’s entry for The Hackaday Prize is a design for a tool that normally exists only as an expensive piece of industrial equipment; out of the reach of normal experimenters, in other words. That tool is a 6-axis micro manipulator and is essentially a small robotic actuator that is capable of very small, very precise movements. It uses 3D printed parts and low-cost components.
The manipulator consists of six identical actuators, each consisting of a single piece of SLS 3D printed nylon with a custom PCB to control a motor and read positional feedback. The motor moves the central pivot point of the 3D printed assembly, which in turn deflects the entire piece by a small amount. By anchoring one point and attaching the other, a small amount of highly controllable movement can be achieved. Six actuators in total form a Gough-Stewart Platform for moving the toolhead.
Interestingly, this 6-Axis Micro Manipulator is a sort of side project. [David] is interested in creating his own digital UV exposer, which requires using UV laser diodes with fiber optic pig tails attached. In an industrial setting these are created by empirically determining the optimal position of a fiber optic with regards to the laser diode by manipulating it with a micro manipulator, then holding it steady while it is cemented in place. Seeing a distinct lack of micro manipulators in anything outside of lab or industrial settings, and recognizing that there would be applications outside of his own needs, [David] resolved to build one.
You’ve no doubt by now seen Boston Dynamics latest “we’re living in the future” robotic creation, dubbed Handle. [Mike Szczys] recently covered the more-or-less-official company unveiling of Handle, the hybrid bipedal-wheeled robot that can handle smooth or rugged terrain and can even jump when it has to, all while remaining balanced and apparently handling up to 100 pounds of cargo with its arms. It’s absolutely sci-fi.