Adversarial IR Hoodie Lets You Own The Night In Anonymity

If you’re in the market for something to obfuscate your nefarious nocturnal activities, rejoice — this adversarial infrared hoodie may be just what you’re looking for.

Not that we condone illegal activities, of course, and neither does artist [Mac Pierce], who created “The Camera-Shy Hoodie.” His purpose seems to be exploring the nature of the surveillance state, or rather to perplex it in the name of anonymity. The idea is simple — equip a standard hoodie with a ring of super-bright IR LEDs, and control them with an RP2040.

We’ve seen blinding hoodies before, but here the LEDs strobe on and off in one of three different patterns, all of which are timed to confound the autoexposure mechanism in just about any surveillance camera by not giving it time to adjust to the rapidly and drastically changing light level. The result is near-total obfuscation of the wearer’s facial features, at least when the camera is in night-vision mode. Check out the results in the video below.

There are some nice touches to [Mac]’s approach, like aluminum PCBs for the LEDs and the use of soldered-on fabric snaps to attach them to the inside of the hoodie, making them easy to remove for laundering. With the LEDs peeking through holes in the fabric, the hoodie looks pretty run-of-the-mill — until, of course, night falls and the USB battery bank in the hoodie’s pocket powers up the light show.

Granted, this won’t exactly help you avoid detection — the big ball of light around your head will be instantly seen by even the most casual observer. But at least it makes it easier to keep your face to yourself. And it won’t help much in daylight — for that, you might want something a little more like this passive adversarial ugly sweater.

Continue reading “Adversarial IR Hoodie Lets You Own The Night In Anonymity”

Render Yourself Invisible To AI With This Adversarial Sweater Of Doom

Ugly sweater season is rapidly approaching, at least here in the Northern Hemisphere. We’ve always been a bit baffled by the tradition of paying top dollar for a loud, obnoxious sweater that gets worn to exactly one social event a year. We don’t judge, of course, but that’s not to say we wouldn’t look a little more favorably on someone’s fashion choice if it were more like this AI-defeating adversarial ugly sweater.

The idea behind this research from the University of Maryland is not, of course, to inform fashion trends, nor is it to create a practical invisibility cloak. It’s really to probe machine learning systems for vulnerabilities by making small changes to the input while watching for changes in the output. In this case, the ML system was a YOLO-based vision system which has little trouble finding humans in an arbitrary image. The adversarial pattern was generated by using a large set of training images, some of which contain the objects of interest — in this case, humans. Each time a human is detected, a random pattern is rendered over the image, and the data is reassessed to see how much the pattern lowers the object’s score. The adversarial pattern eventually improves to the point where it mostly prevents humans from being recognized. Much more detail is available in the research paper (PDF) if you want to dig into the guts of this.

The pattern, which looks a little like a bad impressionist painting of people buying pumpkins at a market and bears some resemblance to one we’ve seen before in similar work, is said to work better from different viewing angles. It also makes a spiffy pullover, especially if you’d rather blend in at that Christmas party.