A Closer Look Inside A Robot’s Typewriter-Inspired Mouth

[Ancient] has a video showing off a fascinating piece of work: a lip-syncing robot whose animated electro-mechanical mouth works like an IBM Selectric typewriter. The mouth rapidly flips between different phonetic positions, creating the appearance of moving lips and mouth. This rapid and high-precision movement is the product of a carefully-planned and executed build, showcased from start to finish in a new video.

Behind the face is a ball that, when moving quickly enough, gives the impression of animated mouth and lips. The new video gives a closer look at how it works.

[Ancient] dubs the concept Selectramatronics, because its action is reminiscent of the IBM Selectric typewriter. Instead of each key having a letter on a long arm that would swing up and stamp an ink ribbon, the Selectric used a roughly spherical unit – called a typeball – with letters sticking out of it like a spiky ball. Hitting the ‘A’ key would rapidly turn the typeball so that the ‘A’ faced forward, then satisfyingly smack it into the ink ribbon at great speed. Here’s a look at how that system worked, by way of designing DIY typeballs from scratch. In this robot, the same concept is used to rapidly flip a ball bristling with lip positions.

We first saw this unusual and fascinating design when its creator showed videos of the end result on social media, pronouncing it complete. We’re delighted to see that there’s now an in-depth look at the internals in the form of a new video (the first link in this post, also embedded below just under the page break.)

The new video is wonderfully wordless, preferring to show rather than tell. It goes all the way from introducing the basic concept to showing off the final product, lip-syncing to audio from an embedded Raspberry Pi.

Thanks to [Luis Sousa] for the tip!

Continue reading “A Closer Look Inside A Robot’s Typewriter-Inspired Mouth”

Billy Whiskers: animatronic puppet

Animatronic Puppet Takes Cues From Animation Software

Lip syncing for computer animated characters has long been simplified. You draw a set of lip shapes for vowels and other sounds your character makes and let the computer interpolate how to go from one shape to the next. But with physical, real world puppets, all those movements have to be done manually, frame-by-frame. Or do they?

Billy Whiskers: animatronic puppet
Billy Whiskers: animatronic puppet

Stop motion animator and maker/hacker [James Wilkinson] is working on a project involving a real-world furry cat character called Billy Whiskers and decided that Billy’s lips would be moved one frame at a time using servo motors under computer control while [James] moves the rest of the body manually.

He toyed around with a number of approaches for making the lip mechanism before coming up with one that worked the way he wanted. The lips are shaped using guitar wire soldered to other wires going to servos further back in the head. Altogether there are four servos for the lips and one more for the jaw. There isn’t much sideways movement but it does enough and lets the brain fill in the rest.

On the software side, he borrows heavily from the tools used for lip syncing computer-drawn characters. He created virtual versions of the five servo motors in Adobe Animate and manipulates them to define the different lip shapes. Animate then does the interpolation between the different shapes, producing the servo positions needed for each frame. He uses an AS3 script to send those positions off to an Arduino. An Arduino sketch then uses the Firmata library to receive the positions and move the servos. The result is entirely convincing as you can see in the trailer below. We’ve also included a video which summarizes the iterations he went through to get to the finished Billy Whiskers or just check out his detailed website.

[Jame’s] work shows that there many ways to do stop motion animation, perhaps a part of what makes it so much fun. One of those ways is to 3D print a separate object for each character shape. Another is to make paper cutouts and move them around, which is what [Terry Gilliam] did for the Monty Python movies. And then there’s what many of us did when we first got our hands on a camera, move random objects around on our parent’s kitchen table and shoot them one frame at a time.

Continue reading “Animatronic Puppet Takes Cues From Animation Software”