How Smart Are AI Chips, Really?

The best part about the term “Artificial Intelligence” is that nobody can really tell you what it exactly means. The main reason for this stems from the term “intelligence”, with definitions ranging from the ability to practice logical reasoning to the ability to perform cognitive tasks or dream up symphonies. When it comes to human intelligence, properties such as self-awareness, complex cognitive feats, and the ability to plan and motivate oneself are generally considered to be defining features. But frankly, what is and isn’t “intelligence” is open to debate.

What isn’t open to debate is that AI is a marketing goldmine. The vagueness has allowed for marketing departments around the world to go all AI-happy, declaring that their product is AI-enabled and insisting that their speech assistant responds ‘intelligently’ to one’s queries. One might begin to believe that we’re on the cusp of a fantastic future inhabited by androids and strong AIs attending to our every whim.

In this article we’ll be looking at the reality behind these claims and ponder humanity’s progress towards becoming a Type I civilization. But this is Hackaday, so we’re also going to dig into the guts of some AI chips, including the Kendryte K210 and see how the hardware of today fits into our Glorious Future. Continue reading “How Smart Are AI Chips, Really?”

5G Is For Robots

Ecclesiastes 1:9 reads “What has been will be again, what has done will be done again; there is nothing new under the sun.” Or in other words, 5G is mostly marketing nonsense; like 4G, 3G, and 2G was before it. Let’s not forget LTE, 4G LTE, Advance 4G, and Edge.

Just a normal everyday antenna array in a Seattle parking garage.

Technically, 5G means that providers could, if they wanted to, install some EHF antennas; the same kind we’ve been using forever to do point to point microwave internet in cities. These frequencies are too lazy to pass through a wall, so we’d have to install these antennas in a grid at ground level. The promised result is that we’ll all get slightly lower latency tiered internet connections that won’t live up to the hype at all. From a customer perspective, about the only thing it will do is let us hit the 8Gb ceiling twice as faster on our “unlimited” plans before they throttle us. It might be nice on a laptop, but it would be a historically ridiculous assumption that Verizon is going to let us tether devices to their shiny new network without charging us a million Yen for the privilege.

So, what’s the deal? From a practical standpoint we’ve already maxed out what a phone needs. For example, here’s a dirty secret of the phone world: you can’t tell the difference between 1080p and 720p video on a tiny screen. I know of more than one company where the 1080p on their app really means 640 or 720 displayed on the device and 1080p is recorded on the cloud somewhere for download. Not a single user has noticed or complained. Oh, maybe if you’re looking hard you can feel that one picture is sharper than the other, but past that what are you doing? Likewise, what’s the point of 60fps 8k video on a phone? Or even a laptop for that matter?

Are we really going to max out a mobile webpage? Since our device’s ability to present information exceeds our ability to process it, is there a theoretical maximum to the size of an app? Even if we had Gbit internet to every phone in the world, from a user standpoint it would be a marginal improvement at best. Unless you’re a professional mobile game player (is that a thing yet?) latency is meaningless to you. The buffer buffs the experience until it shines.

So why should we care about billion dollar corporations racing to have the best network for sending low resolution advertising gifs to our disctracto cubes? Because 5G is for robots.

Continue reading “5G Is For Robots”

Hacking Mars: InSight Mole Is On The Move Again

Your job might be tough, but spare a thought for any of the engineers involved in the Mars InSight lander mission when they learned that one of the flagship instruments aboard the lander, indeed the very instrument for which the entire mission was named, appeared to be a dud. That’s a bad day at work by anyone’s standards, and it happened over the summer when it was reported that the Mars Interior Exploration using Seismic Investigations, Geodesy and Heat Transport lander’s Heat Flow and Physical Properties Package (HP³), commonly known as “The Mole”, was not drilling itself into the Martian regolith as planned.

But now, after months of brainstorming and painstaking testing on Earth and on Mars, it looks as if the mole is working again. NASA has announced that, with a little help from the lander’s backhoe bucket, the HP³ penetrator has dug itself 2 cm into the soil. It’s a far cry from the 5-meter planned depth for its heat-transfer experiments, but it’s progress, and the clever hack that got the probe that far might just go on to salvage a huge chunk of the science planned for the $828 million program.

Continue reading “Hacking Mars: InSight Mole Is On The Move Again”

Developing Guidelines For Sustainable Spaceflight

In the early days of spaceflight, when only the governments of the United States and the Soviet Union had the ability to put an object into orbit, even the most fanciful of futurists would have had a hard time believing that commercial entities would one day be launching sixty satellites at a time. What once seemed like an infinite expanse above our heads is now starting to look quite a bit smaller, and it’s only going to get more crowded as time goes on. SpaceX is gearing up to launch nearly 12,000 individual satellites for their Starlink network by the mid-2020s, and that’s just one of the “mega constellations” currently in the works.

Just some of the objects in orbit around the Earth

It might seem like overcrowding of Earth orbit is a concern for the distant future, but one needs only look at recent events to see the first hints of trouble. On September 2nd, the European Space Agency announced that one of its research spacecraft had to perform an evasive maneuver due to a higher than acceptable risk of colliding with one of the first-generation Starlink satellites. Just two weeks later, Bigelow Aerospace were informed by the United States Air Force that there was a 1 in 20 chance that a defunct Russian Cosmos 1300 satellite would strike their Genesis II space station prototype.

A collision between two satellites in orbit is almost certain to be catastrophic, ending with both spacecraft either completely destroyed or severely damaged. But in the worst case, the relative velocity between the vehicles can be so great that the impact generates thousands of individual fragments. The resulting cloud of shrapnel can circle the Earth for years or even decades, threatening to tear apart any spacecraft unlucky enough to pass by.

Fortunately avoiding these collisions shouldn’t be difficult, assuming everyone can get on the same page before it’s too late. The recently formed Space Safety Coalition (SSC) is made up of more than twenty aerospace companies that realize the importance of taking proactive steps to ensure humanity retains the unfettered access to outer space by establishing some common “Rules of the Road” for future spacecraft.

Continue reading “Developing Guidelines For Sustainable Spaceflight”

DNS-over-HTTPS Is The Wrong Partial Solution

Openness has been one of the defining characteristics of the Internet for as long as it has existed, with much of the traffic today still passed without any form of encryption. Most requests for HTML pages and associated content are in plain text, and the responses are returned in the same way, even though HTTPS has been around since 1994.

But sometimes there’s a need for security and/or privacy. While the encryption of internet traffic has become more widespread for online banking, shopping, the privacy-preserving aspect of many internet protocols hasn’t kept pace. In particular, when you look up a website’s IP address by hostname, the DNS request is almost always transmitted in plain text, allowing all the computers and ISPs along the way to determine what website you were browsing, even if you use HTTPS once the connection is made.

The idea of also encrypting DNS requests isn’t exactly new, with the first attempts starting in the early 2000s, in the form of DNSCrypt, DNS over TLS (DoT), and others. Mozilla, Google, and a few other large internet companies are pushing a new method to encrypt DNS requests: DNS over HTTPS (DoH).

DoH not only encrypts the DNS request, but it also serves it to a “normal” web server rather than a DNS server, making the DNS request traffic essentially indistinguishable from normal HTTPS. This is a double-edged sword. While it protects the DNS request itself, just as DNSCrypt or DoT do, it also makes it impossible for the folks in charge of security at large firms to monitor DNS spoofing and it moves the responsibility for a critical networking function from the operating system into an application. It also doesn’t do anything to hide the IP address of the website that you just looked up — you still go to visit it, after all.

And in comparison to DoT, DoH centralizes information about your browsing in a few companies: at the moment Cloudflare, who says they will throw your data away within 24 hours, and Google, who seems intent on retaining and monetizing every detail about everything you’ve ever thought about doing.

DNS and privacy are important topics, so we’re going to dig into the details here. Continue reading “DNS-over-HTTPS Is The Wrong Partial Solution”

Worn Out EMMC Chips Are Crippling Older Teslas

It should probably go without saying that the main reason most people buy an electric vehicle (EV) is because they want to reduce or eliminate their usage of gasoline. Even if you aren’t terribly concerned about your ecological footprint, the fact of the matter is that electricity prices are so low in many places that an electric vehicle is cheaper to operate than one which burns gas at $2.50+ USD a gallon.

Another advantage, at least in theory, is reduced overal maintenance cost. While a modern EV will of course be packed with sensors and complex onboard computer systems, the same could be said for nearly any internal combustion engine (ICE) car that rolled off the lot in the last decade as well. But mechanically, there’s a lot less that can go wrong on an EV. For the owner of an electric car, the days of oil changes, fouled spark plugs, and the looming threat of a blown head gasket are all in the rear-view mirror.

Unfortunately, it seems the rise of high-tech EVs is also ushering in a new era of unexpected failures and maintenance woes. Case in point, some owners of older model Teslas are finding they’re at risk of being stranded on the side of the road by a failure most of us would more likely associate with losing some documents or photos: a disk read error.

Continue reading “Worn Out EMMC Chips Are Crippling Older Teslas”

The Final Days Of The Fire Lookouts

For more than a century, the United States Forest Service has employed men and women to monitor vast swaths of wilderness from isolated lookout towers. Armed with little more than a pair of binoculars and a map, these lookouts served as an early warning system for combating wildfires. Eventually the towers would be equipped with radios, and later still a cellular or satellite connection to the Internet, but beyond that the job of fire lookout has changed little since the 1900s.

Like the lighthouse keepers of old, there’s a certain romance surrounding the fire lookouts. Sitting alone in their tower, the majority of their time is spent looking at a horizon they’ve memorized over years or even decades, carefully watching for the slightest whiff of smoke. The isolation has been a prison for some, and a paradise for others. Author Jack Kerouac spent the summer of 1956 in a lookout tower on Desolation Peak in Washington state, an experience which he wrote about in several works including Desolation Angels.

But slowly, in a change completely imperceptible to the public, the era of the fire lookouts has been drawing to a close. As technology improves, the idea of perching a human on top of a tall tower for months on end seems increasingly archaic. Many are staunchly opposed to the idea of automation replacing human workers, but in the case of the fire lookouts, it’s difficult to argue against it. Computer vision offers an unwavering eye that can detect even the smallest column of smoke amongst acres of woodland, while drones equipped with GPS can pinpoint its location and make on-site assessments without risk to human life.

At one point, the United States Forest Service operated more than 5,000 permanent fire lookout towers, but today that number has dwindled into the hundreds. As this niche job fades even farther into obscurity, let’s take a look at the fire lookout’s most famous tool, and the modern technology poised to replace it.

Continue reading “The Final Days Of The Fire Lookouts”