Repurposing Old AMD APUs For AI Work

The BC250 is what AMD calls an APU, or Accelerated Processing Unit. It combines a GPU and CPU into a single unit, and was originally built to serve as the heart of certain Samsung rack mount servers. If you know where to find cheap surplus units of the BC250, you can put them to good use for AI work, as [akandr] demonstrates.

The first thing you’ll have to figure out is how to take an individual BC250 APU and get it up and running. It’s effectively a full system-on-chip, combining a Zen 2 CPU with a Cyan Skillfish RDNA 1.5 GPU. However, it was originally intended to run inside a rackmount server unit rather than a standalone machine. To get it going, you’ll need to hook it up with power and some kind of cooling solution.

From there, it’s a matter of software. [akandr] explains how to get AI workflows running on the BC250 using Ollama and Vulkan, while noting useful hacks to improve performance like disabling the GUI and tweaking the CPU governor. The hardware can be used with a wide range of different models depending on what you’re trying to achieve, it just takes some careful management of the APU’s resources to get the most out of it. Thankfully, that’s all in the guide on GitHub.

We’ve already seen these AMD APUs repurposed before for gaming use. Unfortunately the word is out already  about their capabilities, so prices have risen significantly in response to demand. Still, if you manage to score a BC250 and do something cool with it yourself, be sure to let us know on the tipsline!

Ask Hackaday: What Will An LLM Be Good For In The Plateau Of Productivity?

A friend of mine has been a software developer for most of the last five decades, and has worked with everything from 1960s mainframes to the machines of today. She recently tried AI coding tools to see what all the fuss is about, as a helper to her extensive coding experience rather than as a zero-work vibe coding tool. Her reaction stuck with me; she referenced her grandfather who had been born in rural America in the closing years of the nineteenth century, and recalled him describing the first time he saw an automobile.

Après Nous, Le Krach

The Gartner hype cycle graph. Jeremykemp, CC BY-SA 3.0.

We are living amid a wave of AI slop and unreasonable hype so it’s an easy win to dunk on LLMs, but as the whole thing climbs towards the peak of inflated expectations on the Gartner hype cycle perhaps it’s time to look forward. The current AI hype is inevitably going to crash and burn, but what comes afterwards? The long tail of the plateau of productivity will contain those applications in which LLMs are a success, but what will they be? We have yet to hack together a working crystal ball, but perhaps it’s still time to gaze into the future. Continue reading “Ask Hackaday: What Will An LLM Be Good For In The Plateau Of Productivity?”

The Requirements Of AI

The media is full of breathless reports that AI can now code and human programmers are going to be put out to pasture. We aren’t convinced. In fact, we think the “AI revolution” is just a natural evolution that we’ve seen before. Consider, for example, radios. Early on, if you wanted to have a radio, you had to build it. You may have even had to fabricate some or all of the parts. Even today, winding custom coils for a radio isn’t that unusual.

But radios became more common. You can buy the parts you need. You can even buy entire radios on an IC. You can go to the store and buy a radio that is probably better than anything you’d cobble together yourself. Even with store-bought equipment, tuning a ham radio used to be a technically challenging task. Now, you punch a few numbers in on a keypad.

The Human Element

What this misses, though, is that there’s still a human somewhere in the process. Just not as many. Someone has to design that IC. Someone has to conceive of it to start with. We doubt, say, the ENIAC or EDSAC was hand-wired by its designers. They figured out what they wanted, and an army of technicians probably did the work. Few, if any, of them could have envisoned the machine, but they can build it.

Does that make the designers less? No. If you write your code with a C compiler, should assembly programmers look down on you as inferior? Of course, they probably do, but should they?

If you have ever done any programming for most parts of the government and certain large companies, you probably know that system engineering is extremely important in those environments. An architect or system engineer collects requirements that have very formal meanings. Those requirements are decomposed through several levels. At the end, any competent programmer should be able to write code to meet the requirements. The requirements also provide a good way to test the end product.

Continue reading “The Requirements Of AI”

Bruteforcing Accidental Antenna Designs

Antenna design is often referred to as a black art or witchcraft, even by those experienced in the space. To that end, [Janne] wondered—could years of honed skill be replaced by bruteforcing the problem with the aid of some GPUs? Iterative experiments ensued.

[Janne]’s experience in antenna design was virtually non-existent prior to starting, having a VNA on hand but no other knowledge of the craft. Formerly, this was worked around by simply copying vendor reference designs when putting antennas on PCBs. However, knowing that sometimes a need for something specific arises, they wanted a tool that could help in these regards.

The root of the project came from a research paper using an FDTD tool running on GPUs to inversely design photonic nanostructures. Since light is just another form of radio frequency energy, [Janne] realized this could be tweaked into service as an RF antenna design tool. The core simulation engine of the FDTD tool, along with its gradient solver, were hammered into working as an antenna simulator, with [Janne] using LLMs to also tack on a validation system using openEMS, an open-source electromagnetic field solver. The aim was to ensure the results had some validity to real-world physics, particularly important given [Janne] left most of the coding up to large language models. A reward function development system was then implemented to create antenna designs, rank them on fitness, and then iterate further.

The designs produced by this arcane system are… a little odd, and perhaps not what a human might have created. They also didn’t particularly impress in the performance stakes when [Janne] produced a few on real PCBs. However, they do more-or-less line up with their predicted modelled performance, which was promising. Code is on Github if you want to dive into experimenting yourself. Experienced hands may like to explore the nitty gritty details to see if the LLMs got the basics right.

We’ve featured similar “evolutionary” techniques before, including one project that aimed to develop a radio. If you’ve found ways to creatively generate functional hardware from boatloads of mathematics, be sure to let us know on the tipsline!

Habit Detection For Home Assistant

Computers are very good at doing exactly what they’re told. They’re still not very good at coming up with helpful suggestions of their own. They’re very much more about following instructions than using intuition; we still don’t have a digital version of Jeeves to aid our bumbling Wooster selves. [Sherrin] has developed something a little bit intelligent, though, in the form of a habit detector for use with Home Assistant.

In [Sherrin]’s smart home setup, there are lots of things that they wanted to fully automate, but they never got around to implementing proper automations in Home Assistant. Their wife also wanted to automate things without having to get into writing YAML directly. Thus, they implemented a sidecar which watches the actions taken in Home Assistant.

The resulting tool is named TaraHome. When it detects repetitive actions that happen with a certain regularity, it pops up and suggests automating the task. For example, if it detects lights always being dimmed when media is playing, or doors always being locked at night, it will ask if that task should be set to happen automatically and can whip up YAML to suit. The system is hosted on the local Home Assistant instance. It can be paired with an LLM to handle more complicated automations or specific requests, though this does require inviting cloud services into the equation.

We’ve featured lots of great Home Assistant hacks over the years, like this project that bridges 433 MHz gear to the smart home system. If you’ve found your own ways to make your DIY smart home more intelligent, don’t hesitate to notify the tipsline!

AI. Where do you stand?

[Yang-Hui He] Presents To The Royal Institution About AI And Mathematics

Over on YouTube you can see [Yang-Hui He] present to The Royal Institution about Mathematics: The rise of the machines.

In this one hour presentation [Yang-Hui He] explains how AI is driving progress in pure mathematics. He says that right now AI is poised to change the very nature of how mathematics is done. He is part of a community of hundreds of mathematicians pursuing the use of AI for research purposes.

[Yang-Hui He] traces the genesis of the term “artificial intelligence” to a research proposal from J. McCarthy, M.L. Minsky, N. Rochester, and C.E. Shannon dated August 31, 1955. He says that his mantra has become: connectivism leads to emergence, and goes on to explain what he means by that, then follows with universal approximation theorems.

He goes on to enumerate some of the key moments in AI: Descartes’s bête-machine, 1617; Lovelace’s speculation, 1842; Turing test, 1949; Dartmouth conference, 1956; Rosenblatt’s Perceptron, 1957; Hopfield’s network, 1982; Hinton’s Boltzmann machine, 1984; IBM’s Deep Blue, 1997; and DeepMind’s AlphaGo, 2012.

He continues with some navel-gazing about what is mathematics, and what is artificial intelligence. He considers how we do mathematics as bottom-up, top-down, or meta-mathematics. He mentions about one of his earliest papers on the subject Machine-learning the string landscape (PDF) and his books The Calabi–Yau Landscape: From Geometry, to Physics, to Machine Learning and Machine Learning in Pure Mathematics and Theoretical Physics.

He goes on to explain about Mathlib and the Xena Project. He discusses Machine-Assisted Proof by Terence Tao (PDF) and goes on to talk more about the history of mathematics and particularly experimental mathematics. All in all a very interesting talk, if you can find a spare hour!

In conclusion: Has AI solved any major open conjecture? No. Is AI beginning to help to advance mathematical discovery? Yes. Has AI changed the speaker’s day-to-day research routine? Yes and no.

If you’re interested in more fun math articles be sure to check out Digital Paint Mixing Has Been Greatly Improved With 1930s Math and Painted Over But Not Forgotten: Restoring Lost Paintings With Radiation And Mathematics.

Continue reading “[Yang-Hui He] Presents To The Royal Institution About AI And Mathematics”

Can Skynet Be A Statesman?

There’s been a lot of virtual ink spilled about LLMs and their coding ability. Some people swear by the vibes, while others, like the  FreeBSD devs have sworn them off completely. What we don’t often think about is the bigger picture: What does AI do to our civilization? That’s the thrust of a recent paper from the Boston University School of Law, “How AI Destroys Institutions”. Yes, Betteridge strikes again.

We’ve talked before about LLMs and coding productivity, but [Harzog] and [Sibly] from the school of law take a different approach. They don’t care how well Claude or Gemini can code; they care what having them around is doing to the sinews of civilization. As you can guess from the title, it’s nothing good.

"A computer must never make a management decision."
Somehow the tl;dr was written decades before the paper was.

The paper a bit of a slog, but worth reading in full, even if the language is slightly laywer-y. To summarize in brief, the authors try and identify the key things that make our institutions work, and then show one by one how each of these pillars is subtly corroded by use of LLMs. The argument isn’t that your local government clerk using ChatGPT is going to immediately result in anarchy; rather it will facilitate a slow transformation of the democratic structures we in the West take for granted. There’s also a jeremiad about LLMs ruining higher education buried in there, a problem we’ve talked about before.

If you agree with the paper, you may find yourself wishing we could launch the clankers into orbit… and turn off the downlink. If not, you’ll probably let us know in the comments. Please keep the flaming limited to below gas mark 2.