ArduRover Boat Uses Tub To Float

There’s nothing quite like the sight of a plastic box merrily sailing its way around a lake to symbolise how easy it is to get started in autonomous robotics. This isn’t a project we’re writing about because of technical excellence, but purely because watching an autonomous plastic box navigate a lake by itself is surprisingly compelling viewing. The reason that [rctestflight] built the vessel was to test out the capabilities of ArduRover. ArduRover is, of course, a flavour of the extremely popular open source ArduPilot, and in this case is running on a Pixhawk.

The hardware itself is deliberately as simple as possible: two small motors with RC car ESCs, a GPS, some power management and a telemetry module are all it takes. The telemetry module allows the course/mission to be updated on the fly, as well as sending diagnostic data back home. Initially, this setup performed poorly; low GPS accuracy combined with a high frequency control loop piloting a device with little inertia lead to a very erratic path. But after applying some filtering to the GPS this improved significantly.

Despite the simplicity of the setup, it wasn’t immune to flaws. Seaweed in the prop was a cause of some stressful viewing, not to mention the lack of power required to sail against the wind. After these problems caused the boat to drift off course past a nearby pontoon, public sightings ranged from an illegal police drone to a dog with lights on its head.

If you want to use your autonomous boat for other purposes than scaring the public, we’ve written about vessels that have been used to map the depth of the sea bed, track aircraft, and even cross the Atlantic.

Continue reading “ArduRover Boat Uses Tub To Float”

[Jessica] Is Soft On Robot Grippers

It is an old movie trope: a robot grips something and accidentally crushes it with its super robot strength. A little feedback goes a long way, of course, but futuristic robots may also want to employ soft grippers. [Jessica] shows how to build soft grippers made of several cast fingers. The fingers are cast from Ecoflex 00-50, and use air pressure.

A 3D-printed mold is used to cast the Ecoflex fingers, which are only workable for 18 minutes after mixing, so it’s necessary to work fast and have everything ready before you start.

Continue reading “[Jessica] Is Soft On Robot Grippers”

Locating Targets With Charm Courtesy Of A Life Size Portal Turret

What better way to count down the last 7 weeks to a big hacker camp like SHA2017 than by embarking on a last-minute, frantic build? That was [Yvo]’s thought when he decided to make a life-sized version of the adorably lethal turrets from the Valve’s Portal video games. Since that build made it to the finish line back then with not all features added, he finished it up for the CCC camp 2019 event, including the ability to close, open, target and shoot Nerf darts.

Originally based on the miniature 2014 turret (covered on Hackaday as well), [Yvo] details this new project in a first and second work log, along with a detailed explanation of how it all goes together and works. While the 2017 version took a mere 50 days to put together, the whole project took about 300 hours of 3D printing. It also comes with four Nerf guns which use flywheels to launch the darts.  The wheels are powered using quadcopter outrunner motors that spin at 25,000 RPM. The theoretical speed of a launched dart is over 100km/h, with 18 darts per gun and a fire rate of 2 darts per second.

The basic movement control for the system is handled by an Arduino Mega, while the talking and vision aspects are taken care of by a Raspberry Pi 3+, which ultimately also makes the decisions about how to move the system. As one can see in the video after the link, the system seems to work pretty well, with a negligible number of fatalities among company employees.

Though decidedly not a project for the inexperienced tinkerer, [Yvo] has made all of the design files available along with the software. We’re still dubious about the claims about the promised cake for completing one of these turrets, however.

Continue reading “Locating Targets With Charm Courtesy Of A Life Size Portal Turret”

DIY Personal Assistant Robot Hears And Sees All

Who wouldn’t want a robot that can fetch them a glass of water? [Saral Tayal] didn’t just think that, he jumped right in and built his own personal assistant robot. This isn’t just some remote-controlled rover though. The robot actually listens to his voice and recognizes his face.

The body of the robot is the common “Rover 5” platform, to which [Saral] added a number of 3D printed parts. A forklift like sled gives the robot the ability to pick things up. Some of the parts are more about form than function – [Saral] loves NASA’s Spirit and Opportunity Mars rovers, so he added some simulated solar cells and other greebles.

The Logitech webcam up front is very functional — images are fed to machine learning models, while audio is processed to listen for commands. This robot can find and pick up 90 unique objects.

The robot’s brains are a Raspberry Pi. It uses TensorFlow for object recognition. Some of the models [Saral] is using are pretty large – so big that the Pi could only manage a couple of frames per second at 100% CPU utilization. A Google Coral coprocessor sped things up quite a bit, while only using about 30% of the Pi’s processor.

It takes several motors to control to robot’s tracks and sled. This is handled by two Roboclaw motor controllers which themselves are commanded by the Pi.

We’ve seen quite a few mobile robot rovers over the years, but [Saral’s] ‘bot is one of the most functional designs out there. Even better is the fact that it is completely open source. You can find the code and 3D models on his GitHub repo.

Check out a video of the personal assistant rover in action after the break.

Continue reading “DIY Personal Assistant Robot Hears And Sees All”

Designing An Advanced Autonomous Robot: Goose

Robotics is hard, maybe not quite as difficult as astrophysics or understanding human relationships, but designing a competition winning bot from scratch was never going to be easy. Ok, so [Paul Bupe, Jr’s] robot, named ‘Goose’, did not quite win the competition, but we’re very interested to learn what golden eggs it might lay in the aftermath.

The mechanics of the bot is based on a fairly standard dual tracked drive system that makes controlling a turn much easier than if it used wheels. Why make life more difficult than it is already? But what we’re really interested in is the design of the control system and the rationale behind those design choices.

The diagram on the left might look complicated, but essentially the system is based on two ‘brains’, the Teensy microcontroller (MCU) and a Raspberry Pi, though most of the grind is performed by the MCU. Running at 96 MHz, the MCU is fast enough to process data from the encoders and IMU in real time, thus enabling the bot to respond quickly and smoothly to sensors. More complicated and ‘heavier’ tasks such as LIDAR and computer vision (CV) are performed on the Pi, which runs ‘Robot operating system’ (ROS), communicating with the MCU by means of a couple of ‘nodes’.

The competition itself dictated that the bot should travel in large circles within the walls of a large box, whilst avoiding particular objects. Obviously, GPS or any other form of dead reckoning was not going to keep the machine on track so it relied heavily on ‘LiDAR point cloud data’ to effectively pinpoint the location of the robot at all times. Now we really get to the crux of the design, where all the available sensors are combined and fed into a ‘particle filter algorithm’:

What we particularly love about this project is how clearly everything is explained, without too many fancy terms or acronyms. [Paul Bupe, Jr] has obviously taken the time to reduce the overall complexity to more manageable concepts that encourage us to explore further. Maybe [Paul] himself might have the time to produce individual tutorials for each system of the robot?

We could well be reading far too much into the name of the robot, ‘Goose’ being Captain Marvel’s bazaar ‘trans-species’ cat that ends up laying a whole load of eggs. But could this robot help reach a de-facto standard for small robots?

We’ve seen other competition robots on Hackaday, and hope to see a whole lot more!

Video after the break: Continue reading “Designing An Advanced Autonomous Robot: Goose”

An Exoskeleton Arm For A Hacker On A Budget

Whether it is motivated by a dream of superhuman strength courtesy of a mech suit or of mobility for those with impaired muscle function, the powered exoskeleton exerts a curious fascination among engineers. The idea of a machine-augmented human body achieving great things is thwarted though by the difficulty of the task, actuators and power sources small enough to be worn comfortably represent a significant challenge that is not easily overcome. It’s a subject that has captivated [Kristjan Berce] since at a young age seeing his grandmother struggling with lifting, and he presents a working powered exoskeleton arm as a proof of his ideas.

It’s a wonderful exercise in low-tech construction with hand tools and a drill press on pieces of aluminium and wood. Motive power comes from an automotive windscreen wiper motor, and electrical power comes from a hefty LiPo attached to the device’s harness. There is a feedback potentiometer incorporated into the elbow joint, and an Arduino oversees the operation under the direction of a pair of glove-mounted buttons. It’s certainly impressive to see it in the video below lifting a bicycle, though we wonder how its weight might affect someone with less muscle function than average.

Projects like this one are very good to see, because there’s a chance that somebody out there may be helped by building one of these. However there is always a note of caution to be struck, as the best solutions come from those who need them and not those who merely think they have the solution. We have written about the Engineer Saviour Trap here in years past.

This isn’t the first prosthetic arm we’ve seen though, we covered a hackerspace in England printing one for a local youngster.

Continue reading “An Exoskeleton Arm For A Hacker On A Budget”

Pick And Place Robot Built With Fischertechnik

We’d be entirely wrong to think that Fichertechnik is just a toy for kids. It’s also perfect for prototyping the control system of robots. [davidatfsg]’s recent entry in the Hackaday Prize, Delta Robot, shows how complex robotics can be implemented without the hardship of having to drill, cut, bolt together or weld components. The added bonus is that the machine can be completely disassembled non-destructively and rebuilt with a new and better design with little or no waste.

The project uses inverse kinematics running on an Arduino Mega to pick coloured objects off a moving conveyor belt and drop them in their respective bins. There’s also also an optical encoder for regulating the speed of the conveyor and a laser light beam for sensing that the object on the conveyor has reached the correct position to be picked.

Not every component is ‘off the shelf’. [davidatfsg] 3D printed a simple nozzle for the actual ‘pick’ and the vacuum required was generated by the clever use of a pair of pneumatic cylinders and solenoid operated air valves.

We’re pretty sure that this will not be the last project on Hackaday that uses Fischertechnik components and it’s the second one that [davidatfsg] has concocted. Videos of the machine working after the break! Continue reading “Pick And Place Robot Built With Fischertechnik”