The Requirements Of AI

The media is full of breathless reports that AI can now code and human programmers are going to be put out to pasture. We aren’t convinced. In fact, we think the “AI revolution” is just a natural evolution that we’ve seen before. Consider, for example, radios. Early on, if you wanted to have a radio, you had to build it. You may have even had to fabricate some or all of the parts. Even today, winding custom coils for a radio isn’t that unusual.

But radios became more common. You can buy the parts you need. You can even buy entire radios on an IC. You can go to the store and buy a radio that is probably better than anything you’d cobble together yourself. Even with store-bought equipment, tuning a ham radio used to be a technically challenging task. Now, you punch a few numbers in on a keypad.

The Human Element

What this misses, though, is that there’s still a human somewhere in the process. Just not as many. Someone has to design that IC. Someone has to conceive of it to start with. We doubt, say, the ENIAC or EDSAC was hand-wired by its designers. They figured out what they wanted, and an army of technicians probably did the work. Few, if any, of them could have envisoned the machine, but they can build it.

Does that make the designers less? No. If you write your code with a C compiler, should assembly programmers look down on you as inferior? Of course, they probably do, but should they?

If you have ever done any programming for most parts of the government and certain large companies, you probably know that system engineering is extremely important in those environments. An architect or system engineer collects requirements that have very formal meanings. Those requirements are decomposed through several levels. At the end, any competent programmer should be able to write code to meet the requirements. The requirements also provide a good way to test the end product.

Continue reading “The Requirements Of AI”

The Complicated Legacy Of Mind Controlled Toys

Imagine a line of affordable toys controlled by the player’s brainwaves. By interpreting biosignals picked up by the dry electroencephalogram (EEG) electrodes in an included headset, the game could infer the wearer’s level of concentration, through which it would be possible to move physical objects or interact with virtual characters. You might naturally assume such devices would be on the cutting-edge of modern technology, perhaps even a spin-off from one of the startups currently investigating brain-computer interfaces (BCIs).

But the toys in question weren’t the talk of 2025’s Consumer Electronics Show, nor 2024, or even 2020. In actual fact, the earliest model is now nearly as old as the original iPhone. Such is the fascinating story of a line of high-tech toys based on the neural sensor technology developed by a company called Neurosky, the first of which was released all the way back in 2009.

Yet despite considerable interest leading up to their release — fueled at least in part by the fact that one of the models featured Star Wars branding and gave players the illusion of Force powers — the devices failed to make any lasting impact, and have today largely fallen into obscurity. The last toy based on Neurosky’s technology was released in 2015, and disappeared from the market only a few years later.

I had all but forgotten about them myself, until I recently came across a complete Mattel Mindflex at a thrift store for $8.99. It seemed a perfect opportunity to not only examine the nearly 20 year old toy, but to take a look at the origins of the product, and find out what ultimately became of Neurosky’s EEG technology. Was the concept simply ahead of its time? In an era when most people still had flip phones, perhaps consumers simply weren’t ready for this type of BCI. Or was the real problem that the technology simply didn’t work as advertised?

Continue reading “The Complicated Legacy Of Mind Controlled Toys”

Is That Ancient Reel Of PLA Any Good?

When it comes to knowledge there are things you know as facts because you have experienced them yourself or had them verified by a reputable source, and there are things that you know because they are common knowledge but unverified. The former are facts, such as that a 100mm cube of water contains a litre of the stuff, while the latter are received opinions, such as the belief among Americans that British people have poor dental care. The first is a verifiable fact, while the second is subjective.

In our line there are similar received opinions, and one of them is that you shouldn’t print with old 3D printing filament because it will ruin the quality of your print. This is one I can now verify for myself, because I was recently given a part roll of blue PLA from a hackerspace, that’s over a decade old. It’s not been stored in a special environment, instead it’s survived a run of dodgy hackerspace premises with all the heat and humidity that’s normal in a slightly damp country. How will it print?

It Ain’t Stringy

In the first instance, looking at the filament, it looks like any other filament. No fading of the colour, no cracking, if I didn’t know its age it could have been opened within the last few weeks. It loads into the printer, a Prusa Mini, fine, it’s not brittle, and I’m ready to print a Benchy.

Continue reading “Is That Ancient Reel Of PLA Any Good?”

Ask Hackaday: How Do You Detect Hidden Cameras?

The BBC recently published an exposé revealing that some Chinese subscription sites charge for access to their network of hundreds of hidden cameras in hotel rooms. Of course, this is presumably without the consent of the hotel management and probably isn’t specifically a problem in China. After all, cameras can now be very tiny, so it is extremely easy to rent a hotel room or a vacation rental and bug it. This is illegal, China has laws against spy cameras, and hotels are required to check for them, the BBC notes. However, there is a problem: At least one camera found didn’t show up on conventional camera detectors. So we wanted to ask you, Hackaday: How do you detect hidden cameras?

How it Works

Commercial detectors typically use one of two techniques. It is easy to scan for RF signals, and if the camera is emitting WiFi or another frequency you expect cameras to use, that works. But it also misses plenty. A camera might be hardwired, for example. Or store data on an SD card for later. If you have a camera that transmits on a strange frequency, you won’t find it. Or you could hide the camera near something else that transmits. So if your scanner shows a lot of RF around a WiFi router, you won’t be able to figure out that it is actually the router and a small camera.

Continue reading “Ask Hackaday: How Do You Detect Hidden Cameras?”

I, Integrated Circuit

In 1958, the American free-market economist Leonard E Read published his famous essay I, Pencil, in which he made his point about the interconnected nature of free market economics by following everything, and we mean Everything, that went into the manufacture of the humble writing instrument.

I thought about the essay last week when I wrote a piece about a new Chinese microcontroller with an integrated driver for small motors, because a commenter asked me why I was featuring a non-American part. As a Brit I remarked that it would look a bit silly were I were to only feature parts made in dear old Blighty — yes, we do still make some semiconductors! — and it made more sense to feature cool parts wherever I found them. But it left me musing about the nature of semiconductors, and whether it’s possible for any of them to truly only come from one country. So here follows a much more functional I, Chip than Read’s original, trying to work out just where your integrated circuit really comes from. It almost certainly takes great liberties with the details of the processes involved, but the countries of manufacture and extraction are accurate. Continue reading “I, Integrated Circuit”

Ask Hackaday: How Do You Digitize Your Documents?

Like many of you, I have a hard time getting rid of stuff. I’ve got boxes and boxes of weirdo bits and bobs, and piles of devices that I’ll eventually get around to stripping down into even more bits and bobs. Despite regular purges — I try to bring a car-load of crap treasure to local hackerspaces and meetups at least a couple times a year — the pile only continues to grow.

But the problem isn’t limited to hardware components. There’s all sorts of things that the logical part of me understands I’ll almost certainly never need, and yet I can’t bring myself to dispose of. One of those things just so happens to be documents. Anything printed is fair game. Could be the notes from my last appointment with the doctor, or fliers for events I attended years ago. Doesn’t matter, the stacks keep building up until I end up cramming it all into a box and start the whole process starts over again.

I’ve largely convinced myself that the perennial accumulation of electronic bric-à-brac is an occupational hazard, and have come to terms with it. But I think there’s a good chance of moving the needle on the document situation, and if that involves a bit of high-tech overengineering, even better. As such, I’ve spent the last couple of weeks investigating digitizing the documents that have information worth retaining so that the originals can be sent along to Valhalla in my fire pit.

The following represents some of my observations thus far, in the hopes that others going down a similar path may find them useful. But what I’m really interested in is hearing from the Hackaday community. Surely I’m not the only one trying to save some storage space by turn piles of papers into ones and zeros.

Continue reading “Ask Hackaday: How Do You Digitize Your Documents?”

Zombie Netscape Won’t Die

The very concept of the web browser began with a humble piece of software called NCSA Mosaic, all the way back in 1993. It was soon eclipsed by Netscape Navigator, and later Internet Explorer, which became the titans of the 1990s browser market. In turn, they too would falter. Navigator’s dying corpse ended up feeding what would become Mozilla Firefox, and Internet Explorer later morphed into the unexceptional browser known as Edge.

Few of us have had any reason to think about Netscape Navigator since its demise in 2008. And yet, the name lingers on. A zombie from a forgotten age, risen again to haunt us today.

Continue reading “Zombie Netscape Won’t Die”