The Birth Of The Modern Robot

When Isaac Asmiov was writing I, Robot, the field of robotics was still in its infancy. As he notes in The Complete Robot, as the field began to mature, it started showing signs of conforming to the popular ideas held by science fiction writers about what robotics ought to be. Notions of humanoid robots, the functions that robots would have in domestic settings, even the ethical quandaries that AI ethicists face today were all themes of early sci-fi writers.

The idea of a robot – at least of automata – predates the field of robotics. The idea of an independent automata may have existed as early as the ancient Egyptians Chinese, and Greeks, who attempted to build self-operated machines that resembled animals and humans. Myths of clay golems in Jewish legends and clay giants in Norse legends perpetuated the idea of an artificial being that could mimic the actions of living creatures. A 400 BC myth from Crete spoke of a man of bronze who guarded their island from pirates.

Continue reading “The Birth Of The Modern Robot”

You May Trust Driverless Cars, But Do You Trust Driverless Barbers?

Although it is getting more common to have self-driving cars on the road, we have to admit we are still a bit uneasy. After all, we know first hand how hard it is to think of every case and how unreliable things can be. But what about having your hair cut by a robot? At least a car can have airbags and automatically stop at any sign of trouble. But letting a robot hold a sharp instrument up to your head? That’s what Buzz Robotics wants to do and they are starting with a neck trim that you can see in the video below.

Honestly, since the trimmers are probably not that dangerous, we feel a little better. But the fact that the screen says “Calibrated Bad” doesn’t install confidence. While the robobarber might not be able to cut your head off, it could certainly ruin your coiffure.

Continue reading “You May Trust Driverless Cars, But Do You Trust Driverless Barbers?”

ArrBot Is A Fast Way To Get Out Standing In A New Field Of Robotics

[Junglist] correctly points out that agricultural robotics is fast on its way to being the next big thing (TM) and presents his easy to build ArrBot platform so others can get hacking fast. 

The frame is built out of the same brackets and aluminum tubing used to add handrails to stairwells on buildings. Not only is this a fast way to do it, the set-up can be guaranteed to be sturdy since hand rails are often literally standing between life and death. The high ground clearance allows for all sorts of sensors and devices to be mounted while still being able to clear the plants below. 

For motion hub motors driven by an ODrive were re-purposed for the task. He explored turning the wheels as well, but it seems like  differential steer and casters works well for this set-up. ROS on an Nividia Jetson runs the show and deals with the various sensors such as a stereoscopic camera and IMU.

We’re excited to see what hacks people come up with as research in this area grows. (Tee-hee!) For example, [Junglist] wants to see the effect of simply running a UV light over a field rather than spraying with pesticides or fungicides would have.

An Open Assistive Robotic Arm To Help People Feed Themselves

Despite being otherwise capable, not everyone is able to feed themselves. [Julien]’s robot arm project aims to bring this crucial independence back to those people. Assistive devices in this space do exist, but as always they’re prohibitively expensive and the approval process is a nightmare. The development of the arm started by working closely with people who needed it at a local hospital. We note with approval, quite a few cardboard mock-ups to get the size and shape right before more formal work was done in CAD.

The robot arm only has to support a very light payload so its construction can be quite light. A frame of steel rods or plywood is all that’s required. We like how the motion is transferred from stepper motors to the joints of the arm by generously sized timing belts allowing the weight of the arm to remain towards the base. The team behind the project has gotten it to a point, but they’re hoping it will inspire community involvement as they move forward with it.

It’s worth noting, this is not the first assistive eating aid we’ve covered.

RC Strandbeest Is A Head Above The Rest

Prolific maker [Jeremy Cook] recently put the finishing touches (at least, for now) on his impressive ClearCrawler remote controlled Strandbeest, which includes among other things a surprisingly expressive “head” complete with LED matrix eyes. For anyone in the audience who was only mildly terrified of these multi-legged robotic beasties before, you may want to avert your eyes from the video after the break.

The clever locomotive design of [Theo Jansen] known as Strandbeest is a legged walker. What makes it special is that the legs themselves are not independent, but work together for a gliding action more akin to wheeled bots. [Jeremy’s] work with ClearCrawler has taken this to another level of precision and mechanization.

Before installation of the electronics, the ClearCrawler had to be tethered to a bench power supply, and could only move forward and backward. Once the locomotion was working as expected, [Jeremy] was ready to install some brains into the beast.

The robot is controlled by a dual motor driver and an Arduino Nano socketed in an I/O expansion board. Communication between the Nano onboard the walker and the hand-held remote control is provided by of a pair of nRF24L01 modules. The controller itself is a simple affair, comprised of a joystick shield plugged into an Arduino Uno.

The robot’s head is made up of a chunk of clear polycarbonate tube with a 3D printed internal frame to hold the dual 8×8 LED matrices that serve as its animated eyes. This arrangement is mounted on a servo pan and tilt mount, which is controlled by an analog stick on the controller. While the head doesn’t currently serve any practical function, it does give [Jeremy] a chance to emote a bit with his creation; a popular trick when he shows the ClearCrawler off.

A few years ago we covered this robot’s predecessor, the considerably larger ClearWalker. While that machine was surely a beauty to behold, this smaller and more agile iteration of the concept is quite a bit more practical.

Continue reading “RC Strandbeest Is A Head Above The Rest”

Landbeest, A Single Servo Walking Robot

Walking robots have a rich history both on and off the storied pages of Hackaday, but if you will pardon the expression, theirs is not a field that’s standing still. It’s always pleasing to see new approaches to old problems, and the Landbeest built by [Dejan Ristic] is a great example.

It’s a four-legged walker with a gait dictated by a cam-and-follower mechanism that allows it to perform the full range of leg movement with only one motor. Each cam can control more than one leg in synchronisation, and in his most recent prototype, there are two such mechanisms that work on opposite corners of a four-legged machine. The legs are arranged in such a way that the two corner-to-corner pairs pivot at their centres in a similar manner to a pair of scissors; allowing a servo to steer the robot as it walks.

The result certainly isn’t as graceful as [Theo Janssen]’s Strandbeest, from which it evidently takes inspiration for its name, but it’s no less capable for it. After the break you can see a video he’s posted which clearly illustrates its operation and demonstrates its ability to traverse obstacles.

The only thing that’s missing are the files and software should you wish to create your own. He’s unapologetic about this, pointing out that he’d prefer to wait until he is satisfied with it before letting it go. Since he’s put a lot of work in so far and shows no sign of stopping, we’re sure he’ll reach that point soon enough.

Continue reading “Landbeest, A Single Servo Walking Robot”

Robotic Skin Sees When (and How) You’re Touching It

Cameras are getting less and less conspicuous. Now they’re hiding under the skin of robots.

A team of researchers from ETH Zurich in Switzerland have recently created a multi-camera optical tactile sensor that is able to monitor the space around it based on contact force distribution. The sensor uses a stack up involving a camera, LEDs, and three layers of silicone to optically detect any disturbance of the skin.

The scheme is modular and in this example uses four cameras but can be scaled up from there. During manufacture, the camera and LED circuit boards are placed and a layer of firm silicone is poured to about 5 mm in thickness. Next a 2 mm layer doped with spherical particles is poured before the final 1.5 mm layer of black silicone is poured. The cameras track the particles as they move and use the information to infer the deformation of the material and the force applied to it. The sensor is also able to reconstruct the forces causing the deformation and create a contact force distribution. The demo uses fairly inexpensive cameras — Raspberry Pi cameras monitored by an NVIDIA Jetson Nano Developer Kit — that in total provide about 65,000 pixels of resolution.

Apart from just providing more information about the forces applied to a surface, the sensor also has a larger contact surface and is thinner than other camera-based systems since it doesn’t require the use of reflective components. It regularly recalibrates itself based on a convolutional neural network pre-trained with data from three cameras and updated with data from all four cameras. Possible future applications include soft robotics, improving touch-based sensing with the aid of computer vision algorithms.

While self-aware robotic skins may not be on the market quite so soon, this certainly opens the possibility for robots that can detect when too much force is being applied to their structures — the machine equivalent sensation to pain.

Continue reading “Robotic Skin Sees When (and How) You’re Touching It”