[Wills] and his purple DIY sorting hat

From Felt To Fate: Building Your Own Sorting Hat

Ever wondered how it feels to have the Sorting Hat decide your fate? [Will Dana] wanted to find out, so he conjured a bit of Hogwarts magic, and crafted a fully animatronic Sorting Hat from scratch. In the video below, he covers every step of bringing this magical purple marvel to life—from rapid joystick movements to the electronics behind it all.

The heart of the project is two 9g servos—one actuates the mouth, and the other controls the eyebrows—powered by an ESP32 microcontroller. Communication between two ESP32 boards ensures smooth operation via the ESP-NOW protocol, making this a wireless wonder. The design process involved using mechanical advantage to solve jittery servo movements, a trick that will resonate with anyone who’s fought with uncooperative motors.

If animatronics or themed projects excite you, Hackaday has covered similar builds, from a DIY BB-8 droid to a robot fox.

Continue reading “From Felt To Fate: Building Your Own Sorting Hat”

3D Printer Eliminates The Printer Bed

Anyone who has operated a 3D printer before, especially those new to using these specialized tools, has likely had problems with the print bed. The bed might not always be the correct temperature leading to problems with adhesion of the print, it could be uncalibrated or dirty or cause any number of other issues that ultimately lead to a failed print. Most of us work these problems out through trial and error and eventually get settled in, but this novel 3D printer instead removes the bed itself and prints on whatever surface happens to be nearby.

The printer is the product of [Daniel Campos Zamora] at the University of Washington and is called MobiPrint. It uses a fairly standard, commercially available 3D printer head but attaches it to the base of a modified robotic vacuum cleaner. The vacuum cleaner is modified with open-source software that allows it to map its environment without the need for the manufacturer’s cloud services, which in turn lets the 3D printer print on whichever surface the robot finds in its travels. The goal isn’t necessarily to eliminate printer bed problems; a robot with this capability could have many more applications in the realm of accessibility or even, in the future, printing while on the move.

There were a few surprising discoveries along the way which were mentioned in an IEEE Spectrum article, as [Campos Zamora] found while testing various household surfaces that carpet is surprisingly good at adhering to these prints and almost can’t be unstuck from the prints made on it. There are a few other 3D printers out there that we’ve seen that are incredibly mobile, but none that allow interacting with their environment in quite this way.

Continue reading “3D Printer Eliminates The Printer Bed”

A Simple Robot For Learning About Robotics

Robots are super interesting, but you probably shouldn’t start learning about them with a full-sized industrial SCARA arm or anything. Better to learn with something smaller and simpler to understand. This simple Arduino-powered robot is called Bug, and it aims to be just that.

The design comes to us from [Joshua Stanley]. It’s based around the ubiquitous Arduino Uno, paired with a motor control and I/O shield for more connectivity. The robot uses treads for locomotion—each side has two wheels wrapped in a belt for grip. The robot has a small DC gearmotor driving each belt so it can be driven forwards, backwards, and steered differentially. To perceive the world, it uses an off-the-shelf ultrasonic transceiver module, and an NRF24L01 module for remote control. All this is wrapped up in a basic 3D-printed housing that positions the ultrasonic modules effectively as “eyes” which is kind of cute, all in all.

Despite its small size and simple construction, Bug gets around perfectly well in testing on an outdoor footpath. It even has enough torque to flip itself up at full throttle. For now, [Joshua] notes it’s a glorified remote control car, but he plans to expand it further with more functionality going forward.

We see lots of educational robots around these parts, like this nifty little robot arm. Video after the break.

Continue reading “A Simple Robot For Learning About Robotics”

Render of life-size robot rat animatronic on blue plane

Robot Rodents: How AI Learned To Squeak And Play

In an astonishing blend of robotics and nature, SMEO—a robot rat designed by researchers in China and Germany — is fooling real rats into treating it like one of their own.

What sets SMEO apart is its rat-like adaptability. Equipped with a flexible spine, realistic forelimbs, and AI-driven behavior patterns, it doesn’t just mimic a rat — it learns and evolves through interaction. Researchers used video data to train SMEO to “think” like a rat, convincing its living counterparts to play, cower, or even engage in social nuzzling. This degree of mimicry could make SMEO a valuable tool for studying animal behavior ethically, minimizing stress on live animals by replacing some real-world interactions.

For builders and robotics enthusiasts, SMEO is a reminder that robotics can push boundaries while fostering a more compassionate future. Many have reservations about keeping intelligent creatures in confined cages or using them in experiments, so imagine applying this tech to non-invasive studies or even wildlife conservation. In a world where robotic dogs, bees, and even schools of fish have come to life, this animatronic rat sounds like an addition worth further exploring. SMEO’s development could, ironically, pave the way for reducing reliance on animal testing.

Continue reading “Robot Rodents: How AI Learned To Squeak And Play”

A Robot Meant For Humans

Although humanity was hoping for a more optimistic robotic future in the post-war era, with media reflecting that sentiment like The Jetsons or Lost in Space, we seem to have shifted our collective consciousness (for good reasons) to a more Black Mirror/Terminator future as real-world companies like Boston Dynamics are actually building these styles of machines instead of helpful Rosies. But this future isn’t guaranteed, and a PhD researcher is hoping to claim back a more hopeful outlook with a robot called Blossom which is specifically built to investigate how humans interact with robots.

For a platform this robot is not too complex, consisting of an accessible frame that can be laser-cut from wood with only a few moving parts controlled by servos. The robot is not too large, either, and can be set on a desk to be used as a telepresence robot. But Blossom’s creator [Michael] wanted this to help understand how humans interact with robots so the latest version is outfitted not only with a large language model with text-to-speech capabilities, but also with a compelling backstory, lore, and a voice derived from Animal Crossing that’s neither human nor recognizable synthetic robot, all in an effort to make the device more approachable.

To that end, [Michael] set the robot up at a Maker Faire to see what sorts of interactions Blossom would have with passers by, and while most were interested in the web-based control system for the robot a few others came by and had conversations with it. It’s certainly an interesting project and reminds us a bit of this other piece of research from MIT that looked at how humans and robots can work productively alongside one another.

Open-Source Robot Transforms

Besides Pokémon, there might have been no greater media franchise for a child of the 90s than the Transformers, mysterious robots fighting an intergalactic war but which can inexplicably change into various Earth-based object, like trucks and airplanes. It led to a number of toys which can also change shapes from fighting robots into various ordinary objects as well. And, perhaps in a way of life imitating art, plenty of real-life robots have features one might think were inspired by this franchise like this transforming quadruped robot.

Called the CYOBot, the robot has four articulating arms with a wheel at the end of each. The arms can be placed in a wide array of positions for different operating characteristics, allowing the robot to move in an incredibly diverse way. It’s based on a previous version called the CYOCrawler, using similar articulating arms but with no wheels. The build centers around an ESP32-S3 microcontroller, giving it plenty of compute power for things like machine learning, as well as wireless capabilities for control or access to more computing power.

Both robots are open source and modular as well, allowing a range of people to use and add on to the platform. Another perk here is that most parts are common or 3d printed, making it a fairly low barrier to entry for a platform with so many different configurations and options for expansion and development. If you prefer robots without wheels, though, we’d always recommend looking at Strandbeests for inspiration.

Supercon 2023: Cuddly Companion Bots

Even in the advanced world of 2024, robots are still better in science fiction than in reality. Star Trek gave us the erudite and refined Data, Rogue One gave us the fierce yet funny K-2SO, and Big Hero 6 gave us the caring charmer named Baymax. All these robots had smarts, capability, and agency. More than that, though—they were faithful(ish) companions to humans, fulfilling what that role entails.

The thing is, we’re not gonna get robots like that unless somebody builds them. [Angela Sheehan] is a artist and an educator, and a maker—and she’s trying to create exactly that. She came down to the 2023 Hackaday Supercon to tell us all about her efforts to create cuddly companion bots for real.

Continue reading “Supercon 2023: Cuddly Companion Bots”