Mac SE Restomod Has A Floppy Surprise

If he’s anything like us [Duncan Hall] was probably equal parts excited and disgusted when he found a 1987 Macintosh SE case at a garage sale. Excited, because not every day do vintage computers show up at these things. Disgusted, because it had been gutted and coated in house paint; the previous owner apparently wanted to make an aquarium. [Duncan] wanted to make a computer, and after 15 years, he finally did, calling it the PhoeNIX SE.

Note the small hole in the top floppy bay for the laptop webcam.

The NIX part of the name might make you suspect he’s running Linux on it, which yes, he absolutely is. The guts of this restomod were donated from a Dell XPS laptop, whose Core i7 CPU and motherboard power the project. A 9.7″ LCD serves in place of the original monochrome CRT, held in place by 3D printed hardware. While a purist might complain, it’s not like anyone makes replacement CRTs anymore, and once that’s gone? You might as well go full modern. (The analog board, on the other hand, is available. So is the logic board, if you were wondering. Lacking a CRT, some might have chosen e-ink instead, but the LCD looks good here.)

All ports are on the rear, as Steve would have wanted. That original sticker survived under latex paint is a spot of luck.

Having gone full modern, well, there’s no need for the M5011’s dual floppies, so one of them holds a webcam and monitor for a modern experience. A zoom call from that case would be a bit surreal, but we really appreciate the use of the empty floppy bay to keep the clean lines of the Macintosh SE unaltered. The other floppy bay (this is a dual-floppy unit) appears empty; we might have put an SD-card reader or something in there, but we absolutely agree with [Duncan]’s choice to 3D Print a new back panel and keep all I/O on the rear of the case, as God and Steve Jobs intended.

However you feel about restomodding retrocomputers (and we’re aware it’s a controversial practice), I think we can all agree this is a much better fate for the old Mac than becoming an aquarium. Thanks to [Loddington] for the tip.

If you’re on the side of the aisle that prefers to see restorations than restomods, the tips line is waiting for some quality restorations.

Robot Dinosaur YOLOs Colors And Shapes For Kids

YOLO can mean many things, but in the context of [be_riddickulous]’s AI Talking Robot Dinosaur it refers to the “You Only Look Once” YOLOv11 object-detection algorithm by Ultralytics, the method by which this adorable dino recognizes colors and shapes to teach them to children.

If you’re new to using YOLO or object recognition more generally, [be_riddiculous]’s tutorial is not a bad place to get started. She goes through how many images you’ll need and what types to get the shape-and-color recognition needed for this project, as well as how to annotate them and train the model, either locally or in the cloud.

The project itself is an adorable paper-mache dinosaur with a servo-actuated mouth hiding some LEDs and a Raspberry Pi camera module to provide images. In operation, the dinosaur “talks” to children using pre-recorded voice lines, inviting them to play a game and put a specific shape, or shape of a specific color (or both) in its mouth. Then the aforementioned object detection (running on a laptop) goes “YOLO” and identifies the shape so the toy can provide feedback on the child’s choice via a speaker in the belly of the beast.

The link to the game code is currently not valid, but it looks like they used PyGame for the audio output code. A servo motor controls the mouth, but without that code it’s not entirely clear to us what it’s doing. We expect by the time you read this there’s good odds [be_riddickulous] will have fixed that link and you can see for yourself.

The only thing that holds this back from being a great toy to put in every Kindergarten class is the need to have a laptop close by to plug the webcam into. A Raspberry Pi 5 ought to have the horsepower to run YOLOv11, so with a little extra effort the whole thing could be standalone — there might even be room in there for batteries.We’ve had other hacks aimed at little ones, like a kid-friendly computer to relive the glory days of the school computer lab or one of the many iterations of the RFID jukebox idea. If you want to wow the kiddos with AI, perhaps take a look at this talking Santa plush.

Got a cool project, AI, kid-related, or otherwise? Don’t forget to toss us a tip!

2025 Pet Hacks Contest: Fort Bawks Is Guarded By Object Detection

One of the difficult things about raising chickens is that you aren’t the only thing that finds them tasty. Foxes, raccoons, hawks — if it can eat meat, it probably wants a bite of your flock. [donutsorelse] wanted to protect his flock and to be able to know when predators were about without staying up all night next to the hen-house. What to do but outsource the role of Chicken Guardian to a Raspberry pi?

Object detection is done using a YOLOv8 model trained on images of the various predators local to [donutorelse]. The model is running on a Raspberry Pi and getting images from a standard webcam. Since the webcam has no low-light capability, the system also has a motion-activated light that’s arguably goes a long way towards spooking predators away itself. To help with the spooking, a speaker module plays specific sound files for each detected predator — presumably different sounds might work better at scaring off different predators.

If that doesn’t work, the system phones home to activate a siren inside [donutorelse]’s house, using a Blues Wireless Notecarrier F as a cellular USB modem. The siren is just a dumb unit; activation is handled via a TP-Link smart plug that’s hooked into [donutorelse]’s custom smart home setup. Presumably the siren cues [donutorelse] to take action against the predator assault on the chickens.

Weirdly enough, this isn’t the first time we’ve seen an AI-enabled chicken coop, but it is the first one to make into our ongoing challenge, which incidentally wraps up today.

Nerf Blaster Becomes Remote Control Turret

For most of us, turrets that aim and shoot at things are the sole domain of video games. However, they’re remarkably easy to build with modern technology, as [meub] demonstrates. Meet the SwarmTurret.

The build is based around an existing foam blaster, namely the Nerf Swarmfire. This blaster was chosen for being easy to integrate into the build, thanks to its motorized direct-plunger firing mechanism and electronic trigger. It also has the benefit of being far less noisy and quicker to fire than most flywheel blasters.

For this build, the Nerf blaster was slimmed down and fitted to a turret base built with hobby servos and 3D printed components. The blaster is also fitted with a webcam for remote viewing. A Raspberry Pi is running the show, serving up a video feed and allowing aiming commands to be sent via a Websockets-based interface. Thus, you can login via a web browser on your phone or laptop, and fire away at targets to your heart’s content.

We’ve featured some great turrets before, like this Portal-themed unit.

Continue reading “Nerf Blaster Becomes Remote Control Turret”

Tearing Down And Hacking The T2S+ Thermal Camera

[Dmytro] was able to lay his hands on a InfiRay T2S+ camera. It’s a capable thermal imaging unit that comes at a cheaper price than many of its rivals. [Dmytro] decided to pull it apart to see what makes it tick, and he discovered a few interesting things along the way.

Like so much modern hardware, pulling the case apart does require some spudging and levering. Once inside, though, it comes apart in a relatively straightforward manner. Once inside, [Dmytro] notes some similarities between this camera and the Flir Lepton, another affordable thermal camera on the market. He also finds a clone of the Cypress FX2LP chip, which is used for talking USB. There’s also an Gowin FPGA inside, with [Dmytro] suspecting the gateware onboard could be modified. If so, the camera may be a candidate for running open source firmware in future.

What bothered [Dmytro] about this camera, though, was the software. When used with an Android phone, the camera demands the use of a proprietary app with with questionable permissions. It can be used on a regular computer, where it appears as a standard webcam. However, in this mode, the camera fails to self-calibrate, and the images quickly become useless. [Dmytro] worked to hack around this, by figuring out a way to trigger calibrations and run the proper image corrections manually when using the camera without the smartphone app. He also explores techniques to improve the resolution of the thermal measurements made by the camera.

We’ve seen some other neat thermal camera hacks over the years. Video after the break.

Continue reading “Tearing Down And Hacking The T2S+ Thermal Camera”

Contagious Ideas

We ran a story about a wall-mounted plotter bot this week, Mural. It’s a simple, but very well implemented, take on a theme that we’ve seen over and over again in various forms. Two lines, or in this case timing belts, hang the bot on a wall, and two motors drive it around. Maybe a servo pulls the pen in and out, but that’s about it. The rest is motor driving and code.

We were thinking about the first such bot we’ve ever seen, and couldn’t come up with anything earlier than Hektor, a spray-painting version of this idea by [Juerg Lehni]. And since then, it’s reappeared in numerous variations.

Some implementations mount the motors on the wall, some on the bot. There are various geometries and refinements to try to make the system behave more like a simple Cartesian one, but in the end, you always have to deal with a little bit of geometry, or just relish the not-quite-straight lines. (We have yet to see an implementation that maps out the nonlinearities using a webcam, for instance, but that would be cool.) If you’re feeling particularly reductionist, you can even do away with the pen-lifter entirely and simply draw everything as a connected line, Etch-a-Sketch style. Maslow CNC swaps out the pen for a router, and cuts wood.

What I love about this family of wall-plotter bots is that none of them are identical, but they all clearly share the same fundamental idea. You certainly wouldn’t call any one of them a “copy” of another, but they’re all related, like riffing off of the same piece of music, or painting the same haystack in different lighting conditions: robot jazz, or a study in various mechanical implementations of the same core concept. The collection of all wall bots is more than the sum of its parts, and you can learn something from each one. Have you made yours yet?

(Fantastic plotter-bot art by [Sarah Petkus] from her write-up ten years ago!)

Hackaday Podcast Episode 312: Heart Attacks, The Speed Of Light, And Self-balancing

Elliot does the podcast on the road to Supercon Europe, and Al is in the mood for math and nostalgia this week. Listen in and find out what they were reading on Hackaday this week.

The guys talked about the ESP-32 non-backdoor and battery fires. Then it was on to the hacks.

Self-balancing robots and satellite imaging were the appetizers, but soon they moved on to Kinect cameras in the modern day. Think you can’t travel at the speed of light? Turns out that maybe you already are.

Did you know there was a chatbot in 1957? Well, sort of. For the can’t miss stories: watches monitor your heart and what does the number e really mean?

Check out the links below if you want to follow along, and as always, tell us what you think about this episode in the comments!

Download in DRM-free MP3 and stream it on the big speakers.

Continue reading “Hackaday Podcast Episode 312: Heart Attacks, The Speed Of Light, And Self-balancing”