Multi Material 3D Printing Makes Soft Robot

When you zoom in on a fractal you find it is made of more fractals. Perhaps that helped inspire the Harvard 3D printers that have various arrays of mixing nozzles. In the video below you can see some of the interesting things you can do with an array of mixing nozzles. The coolest, we think, is a little multi-legged robot that uses vacuum to ambulate across the bench. The paper, however, is behind a paywall.

There are really two ideas here. Mixing nozzles are nothing new. Usually, you use them to mimic a printer with two hot ends. That is, you print one material at a time and purge the old filament out when switching to the new filament. This is often simpler than using two heads because with a two head arrangement, both the heads have to be at the same height, you must know the precise offset between the heads, and you generally lose some print space since the right head can’t cross the left head and vice versa. Add more heads, and you multiply those problems. We’ve also seen mixing nozzles provide different colors.

Continue reading “Multi Material 3D Printing Makes Soft Robot”

Robot Vs. Superbug

Working in a university or research laboratory on interesting, complicated problems in the sciences has a romanticized, glorified position in our culture. While the end results are certainly worth celebrating, often the process of new scientific discovery is underwhelming, if not outright tedious. That’s especially true in biology and chemistry, where scaling up sample sizes isn’t easy without a lot of human labor. A research group from Reading University was able to modify a 3D printer to take some of that labor out of the equation, though.

This 3D printer was used essentially as a base, with the printing head removed and replaced with a Raspberry Pi camera. The printer X/Y axes move the camera around to all of the different sample stored in the print bed, which allows the computer attached to the printer to do most of the work that a normal human would have had to do. This allows them to scale up massively and cheaply, presumably with less tedious inputs from a large number of graduate students.

While the group hopes that this method will have wide applicability for any research group handling large samples, their specific area of interest involves researching “superbugs” or microbes which have developed antibiotic resistance. Their recently-published paper states that any field which involves bacterial motility, colony growth, microtitre plates or microfluidic devices could benefit from this 3D printer modification.

A Fantastic Frontier Of FPGA Flexibility Found In The 2019 Supercon Badge

We have just concluded a successful Hackaday Superconference where a highlight for many was digging into this year’s hardware badge. Shaped in the general form of a Game Boy handheld gaming console, the heart of the badge is a large FPGA opening up new and exciting potential for badge hacking.

Beyond our normal tools of compiling custom code or modifying hardware with a soldering iron, we now have the option to change core hardware behavior with Verilog. And people explored this new frontier to great effect, as seen at the badge hacking ceremony. (Video embedded below.)

FPGAs are not new, technically speaking, why are they exciting now? We can thank their recent growth in capability, their rapidly falling cost, and the relatively new availability of open source toolchains. These developments elevated FPGA into one of the most exciting trends in hardware today, so this year’s badge master [Sprite_TM] built an open FPGA playground for several hundred of his closest Supercon friends. Let’s take a look at what people were able to accomplish in just a few days using this unique and powerful hardware.

Continue reading “A Fantastic Frontier Of FPGA Flexibility Found In The 2019 Supercon Badge”

The Golden Age Of Ever-Changing Computer Architecture

Given the accuracy of Moore’s Law to the development of integrated circuits over the years, one would think that our present day period is no different from the past decades in terms of computer architecture design. However, during the 2017 ACM Turing Award acceptance speech, John L. Hennessy and David A. Patterson described the present as the “golden age of computer architecture”.

Compared to the early days of MS-DOS, when designing user- and kernel-space interactions was still an experiment in the works, it certainly feels like we’re no longer in the infancy of the field. Yet, as the pressure mounts for companies to acquire more computational resources for running expensive machine learning algorithms on massive swaths of data, smart computer architecture design may be just what the industry needs.

Moore’s law predicts the doubling of transistors in an IC, it doesn’t predict the path that IC design will take. When that observation was made in 1965 it was difficult or even impossible to envision where we are today, with tools and processes so closely linked and widely available that the way we conceive processor design is itself multiplying.

Continue reading “The Golden Age Of Ever-Changing Computer Architecture”

Robotic Skin Sees When (and How) You’re Touching It

Cameras are getting less and less conspicuous. Now they’re hiding under the skin of robots.

A team of researchers from ETH Zurich in Switzerland have recently created a multi-camera optical tactile sensor that is able to monitor the space around it based on contact force distribution. The sensor uses a stack up involving a camera, LEDs, and three layers of silicone to optically detect any disturbance of the skin.

The scheme is modular and in this example uses four cameras but can be scaled up from there. During manufacture, the camera and LED circuit boards are placed and a layer of firm silicone is poured to about 5 mm in thickness. Next a 2 mm layer doped with spherical particles is poured before the final 1.5 mm layer of black silicone is poured. The cameras track the particles as they move and use the information to infer the deformation of the material and the force applied to it. The sensor is also able to reconstruct the forces causing the deformation and create a contact force distribution. The demo uses fairly inexpensive cameras — Raspberry Pi cameras monitored by an NVIDIA Jetson Nano Developer Kit — that in total provide about 65,000 pixels of resolution.

Apart from just providing more information about the forces applied to a surface, the sensor also has a larger contact surface and is thinner than other camera-based systems since it doesn’t require the use of reflective components. It regularly recalibrates itself based on a convolutional neural network pre-trained with data from three cameras and updated with data from all four cameras. Possible future applications include soft robotics, improving touch-based sensing with the aid of computer vision algorithms.

While self-aware robotic skins may not be on the market quite so soon, this certainly opens the possibility for robots that can detect when too much force is being applied to their structures — the machine equivalent sensation to pain.

Continue reading “Robotic Skin Sees When (and How) You’re Touching It”

Making A Robotic Dog Better By Adding Springiness Without Springs

Getting a legged robot to stay upright, especially a quadruped or biped, can be a challenging undertaking. To experiment with different approaches, [James Bruton] built robot dog test platform and is playing with “dynamic compliant simulated springs“, or in other words, using the motors to act as though they were springs and dampers..

When robotic legs are kept stiff, they tend to reduce the stability of the platform due to the sudden erratic movements of the robot, especially on uneven surfaces. With a back drivable joint arrangement, [James] is using limited holding current on the motor, and the position of the motor shaft is monitored using an encoder. When a leg experiences a resisting force, with will have some “give” and then the motor will return it to it’s intended position more slowly. Using a IMU on top of the robot, it can detect when it start leaning to a side, and then temporarily soften the other side to balance the robot.

This is quite a common technique in legged robots, but [James] does an excellent job of explaining just how it works. He hopes to use the lessons learned from the test platform to improve or redesign his already impressive OpenDog.

We’ve seen a number of quadruped robots on Hackaday recently. Including Boston Dynamics’ very expensive Spot as well as a low cost robot dog that giving its big brothers a run for their money, and doing some back flips in the process. Check out James’ video after the break. Continue reading “Making A Robotic Dog Better By Adding Springiness Without Springs”

Robotics Controller For The Pi Boasts An Impressive Feature List

[Michael Horne] recently shared his thoughts on the RedBoard+, a motor controller board for the Raspberry Pi aimed at robotic applications. His short version for busy people is: if you’re at all into robotics, get one because it’s fantastic.

At heart the RedBoard+ is a motor controller, but it’s packed with I/O and features that set it above the usual fare. It can drive two DC motors and up to twelve servos, but what is extra useful is the wide input range of 7-24 V and its ability to power and control the underlying Raspberry Pi. A user-programmable button defaults to either doing a reboot or safe shutdown, depending on how long the button is held. Another neat feature is the ability to blink out the IP address of the Pi using the onboard RGB LED, which is always handy in a pinch.

The RedBoard+ has a GitHub repository which provides a variety of test scripts and an easy to use library, as well as a variety of hookup guides and quickstart guides. There’s even a pre-configured SD image for those who prefer to simply dive in.

A brief demo video showing the board in operation is embedded below. If you’re interested in one, Creator [Neil] of RedRobotics has made it available for sale on Tindie.

Continue reading “Robotics Controller For The Pi Boasts An Impressive Feature List”