Multifunctional USB controlled PCB on blue background

How A Tiny Relay Became A USB Swiss Army Knife

Meet the little board that could: [alcor6502]’s tiny USB relay controller, now evolved into a multifunction marvel. Originally built as a simple USB relay to probe the boundaries of JLCPCB’s production chops, it has become a compact utility belt for any hacker’s desk drawer. Not only has [alcor6502] actually built the thing, he even provided instructions. If you happened to be at Hackaday in Berlin, you now might even own one, as he handed out twenty of them during his visit. If not, read on and build it yourself.

This thing is not just a relay, and that is what makes it special. Depending on a few solder bridges and minimal components, it shape-shifts into six different tools: a fan controller (both 3- and 4-pin!), servo driver, UART interface, and of course, the classic relay. It even swaps out a crystal oscillator for USB self-sync using STM32F042‘s internal RC – no quartz, less cost, same precision. A dual-purpose BOOT0 button lets you flash firmware and toggle outputs, depending on timing. Clever reuse, just like our mothers taught us.

It’s the kind of design that makes you want to tinker again. Fewer parts. More function. And that little smile when it just works. If this kind of clever compactness excites you too, read [alcor6502]’s build log and instructions here.

Mockup of a printed copy of the Little OS Book

One Book To Boot Them All

Somewhere in the universe, there’s a place that lists every x86 operating system from scratch. Not just some bootloaders, or just a kernel stub, but documentation to build a fully functional, interrupt-handling, multitasking-capable OS. [Erik Helin and Adam Renberg] did just that by documenting every step in The Little Book About OS Development.

This is not your typical dry academic textbook. It’s a hands-on, step-by-step guide aimed at hackers, tinkerers, and developers who want to demystify kernel programming. The book walks you through setting up your environment, bootstrapping your OS, handling interrupts, implementing virtual memory, and even tackling system calls and multitasking. It provides just enough detail to get you started but leaves room for exploration – because, let’s be honest, half the fun is in figuring things out yourself.

Completeness and structure are two things that make this book stand out. Other OS dev guides may give you snippets and leave you to assemble the puzzle yourself. This book documents the entire process, including common pitfalls. If you’ve ever been lost in the weeds of segmentation, paging, or serial I/O, this is the map you need. You can read it online or fetch it as a single 75-page long PDF.

Mockup photo source: Matthieu Dixte

Reconfigurable FPGA For Single Photon Measurements

Detecting single photons can be seen as the backbone of cutting-edge applications like LiDAR, medical imaging, and secure optical communication. Miss one, and critical information could be lost forever. That’s where FPGA-based instrumentation comes in, delivering picosecond-level precision with zero dead time. If you are intrigued, consider sitting in on the 1-hour webinar that [Dr. Jason Ball], engineer at Liquid Instruments, will host on April 15th. You can read the announcement here.

Before you sign up and move on, we’ll peek into a bit of the matter upfront. The power lies in the hardware’s flexibility and speed. It has the ability to timestamp every photon event with a staggering 10 ps resolution. That’s comparable to measuring the time it takes light to travel just a few millimeters. Unlike traditional photon counters that choke on high event rates, this FPGA-based setup is reconfigurable, tracking up to four events in parallel without missing a beat. From Hanbury-Brown-Twiss experiments to decoding pulse-position modulated (PPM) data, it’s an all-in-one toolkit for photon wranglers. [Jason] will go deeper into the subject and do a few live experiments.

Measuring single photons can be achieved with photomultipliers as well. If exploring the possibilities of FPGA’s is more your thing, consider reading this article.

Multi-Divi book with hand thumbing through it

Math, Optimized: Sweden’s Maximal Multi-Divi

Back in the early 1900s, before calculators lived in our pockets, crunching numbers was painstaking work. Adding machines existed, but they weren’t exactly convenient nor cheap. Enter Wilken Wilkenson and his Maximal Multi-Divi, a massive multiplication and division table that turned math into an industrialized process. Originally published in Sweden in the 1910’s, and refined over decades, his book was more than a reference. It was a modular calculating instrument, optimized for speed and efficiency. In this video, [Chris Staecker] tells all about this fascinating relic.

What makes the Multi-Divi special isn’t just its sheer size – handling up to 9995 × 995 multiplications – but its clever design. Wilkenson formatted the book like a machine, with modular sections that could be swapped out for different models. If you needed an expanded range, you could just swap in an extra 200 pages. To sell it internationally, just replace the insert – no translation needed. The book itself contains zero words, only numbers. Even the marketing pushed this as a serious calculating device, rather than just another dusty math bible.

While pinwheel machines and comptometers were available at the time, they required training and upkeep. The Multi-Divi, in contrast, required zero learning curve – just look up the numbers for instant result. And it wasn’t just multiplication: the book also handled division in reverse, plus compound interest, square roots, and even amortizations. Wilkenson effectively created a pre-digital computing tool, a kind of pocket calculator on steroids (if pockets were the size of briefcases).

Of course, no self-respecting hacker would take claims of ‘the greatest invention ever’ at face value. Wilkenson’s marketing, while grandiose, wasn’t entirely wrong – the Multi-Divi outpaced mechanical calculators in speed tests. And if you’re feeling adventurous, [Chris] has scanned the entire book, so you can try it yourself.

Continue reading “Math, Optimized: Sweden’s Maximal Multi-Divi”

Dwingeloo telescope with sun shining through

Dwingeloo To Venus: Report Of A Successful Bounce

Radio waves travel fast, and they can bounce, too. If you are able to operate a 25-meter dish, a transmitter, a solid software-defined radio, and an atomic clock, the answer is: yes, they can go all the way to Venus and back. On March 22, 2025, the Dwingeloo telescope in the Netherlands successfully pulled off an Earth-Venus-Earth (EVE) bounce, making them the second group of amateurs ever to do so. The full breakdown of this feat is available in their write-up here.

Bouncing signals off planets isn’t new. NASA has been at it since the 1960s – but amateur radio astronomers have far fewer toys to play with. Before Dwingeloo’s success, AMSAT-DL achieved the only known amateur EVE bounce back in 2009. This time, the Dwingeloo team transmitted a 278-second tone at 1299.5 MHz, with the round trip to Venus taking about 280 seconds. Stockert’s radio telescope in Germany also picked up the returning echo, stronger than Dwingeloo’s own, due to its more sensitive receiving setup.

Post-processing wasn’t easy either. Doppler shift corrections had to be applied, and the received signal was split into 1 Hz frequency bins. The resulting detections clocked in at 5.4 sigma for Dwingeloo alone, 8.5 sigma for Stockert’s recording, and 9.2 sigma when combining both datasets. A clear signal, loud and proud, straight from Venus’ surface.

The experiment was cut short when Dwingeloo’s transmitter started failing after four successful bounces. More complex signal modulations will have to wait for the next Venus conjunction in October 2026. Until then, you can read our previously published article on achievements of the Dwingeloo telescope.

Microscopic view of chiral magnetic material

Twisting Magnetism To Control Electron Flow

If you ever wished electrons would just behave, this one’s for you. A team from Tohoku, Osaka, and Manchester Universities has cracked open an interesting phenomenon in the chiral helimagnet α-EuP3: they’ve induced one-way electron flow without bringing diodes into play. Their findings are published in the Proceedings of the National Academy of Sciences.

The twist in this is quite literal. By coaxing europium atoms into a chiral magnetic spiral, the researchers found they could generate rectification: current that prefers one direction over another. Think of it as adding a one-way street in your circuit, but based on magnetic chirality rather than semiconductors. When the material flips to an achiral (ferromagnetic) state, the one-way effect vanishes. No asymmetry, no preferential flow. They’ve essentially toggled the electron highway signs with an external magnetic field. This elegant control over band asymmetry might lead to low-power, high-speed data storage based on magnetic chirality.

If you are curious how all this ties back to quantum theory, you can trace the roots of chiral electron flow back to the early days of quantum electrodynamics – when physicists first started untangling how particles and fields really interact.

There’s a whole world of weird physics waiting for us. In the field of chemistry, chirality has been covered by Hackaday, foreshadowing the lesser favorable ways of use. Read up on the article and share with us what you think.

Closeup of the original Manchester Baby CRT screen

Modern Computing’s Roots Or The Manchester Baby

In the heart of Manchester, UK, a groundbreaking event took place in 1948: the first modern computer, known as the Manchester Baby, ran its very first program. The Baby’s ability to execute stored programs, developed with guidance from John von Neumann’s theory, marks it as a pioneer in the digital age. This fascinating chapter in computing history not only reshapes our understanding of technology’s roots but also highlights the incredible minds behind it. The original article, including a video transcript, sits here at [TheChipletter]’s.

So, what made this hack so special? The Manchester Baby, though a relatively simple prototype, was the first fully electronic computer to successfully run a program from memory. Built by a team with little formal experience in computing, the Baby featured a unique cathode-ray tube (CRT) as its memory store – a bold step towards modern computing. It didn’t just run numbers; it laid the foundation for all future machines that would use memory to store both data and instructions. Running a test to find the highest factor of a number, the Baby performed 3.5 million operations over 52 minutes. Impressive, by that time.

Despite criticisms that it was just a toy computer, the Baby’s significance shines through. It was more than just a prototype; it was proof of concept for the von Neumann architecture, showing us that computers could be more than complex calculators. While debates continue about whether it or the ENIAC should be considered the first true stored-program computer, the Baby’s role in the evolution of computing can’t be overlooked.

Continue reading “Modern Computing’s Roots Or The Manchester Baby”