AI Mistakes Are Different, And That’s A Problem

People have been making mistakes — roughly the same ones — since forever, and we’ve spent about the same amount of time learning to detect and mitigate them. Artificial Intelligence (AI) systems make mistakes too, but [Bruce Schneier] and [Nathan E. Sanders] make the observation that, compared to humans, AI models make entirely different kinds of mistakes. We are perhaps less equipped to handle this unusual problem than we realize.

The basic idea is this: as humans we have tremendous experience making mistakes, and this has also given us a pretty good idea of what to expect our mistakes to look like, and how to deal with them. Humans tend to make mistakes at the edges of our knowledge, our mistakes tend to clump around the same things, we make more of them when bored or tired, and so on. We have as a result developed controls and systems of checks and balances to help reduce the frequency and limit the harm of our mistakes. But these controls don’t carry over to AI systems, because AI mistakes are pretty strange.

The mistakes of AI models (particularly Large Language Models) happen seemingly randomly and aren’t limited to particular topics or areas of knowledge. Models may unpredictably appear to lack common sense. As [Bruce] puts it, “A model might be equally likely to make a mistake on a calculus question as it is to propose that cabbages eat goats.” A slight re-wording of a question might be all it takes for a model to suddenly be confidently and utterly wrong about something it just a moment ago seemed to grasp completely. And speaking of confidence, AI mistakes aren’t accompanied by uncertainty. Of course humans are no strangers to being confidently wrong, but as a whole the sort of mistakes AI systems make aren’t the same kinds of mistakes we’re used to.

There are different ideas on how to deal with this, some of which researchers are (ahem) confidently undertaking. But for best results, we’ll need to invent new ways as well. The essay also appeared in IEEE Spectrum and isn’t terribly long, so take a few minutes to check it out and get some food for thought.

And remember, if preventing mistakes at all costs is the goal, that problem is already solved: GOODY-2 is undeniably the world’s safest AI.

44 thoughts on “AI Mistakes Are Different, And That’s A Problem

  1. In my humble experience, LLMs are great for coding as long as what you require is strictly a very standard and vanilla implementation of something.
    The moment you want anything even slightly complex, or something slightly different from the “standard” thing, it will fail to give you what you want.

    Also its best for getting small functions or snippets written, as opposed to large files full of code. It starts hallucinating if you do that.

    1. so it’s a super inefficient template generator?

      seriously, just have the natural language parsing tokenize it into something that can filter a list of templates and pay actual programmer to generate them. Why any company would allow the liability issues AI code generates baffles me.

    1. I often model LLMs as containing a random number generator which occasionally causes it to output the worst possible thing at any given moment. Additionally, the worst possible thing at any moment will also be output if the hash of the input has a couple of trailing zeroes, meaning it doesn’t take very long for a malicious prompter to find a prompt injection that works.

      So far nobody has found any way to meaningfully improve LLMs in a way that would allow this model to be substituted with a more preferable one. Not with larger models, no with improved training, not with feedback, not with any kind of handcrafted or hardcoded filter, outside system, or check, and not with ensembles of LLMs checking other LLMs either (see Q* aka strawberry).

      Once you start modelling LLMs this way, you realize there are precious few non-adversarial (because prompt injection) and fault-tolerant (because catastrophic failure happens often and without warning) use cases left. Using LLMs as text adventure games is probably the only reliable one.

  2. Honestly, the kinds of mistakes that I’ve seen them make is more in-line with a rush job done by amateurs. Things that are obviously wrong, but would normally get skipped over because rush-jobs by amateurs isn’t what you might think with AI.

    I’ve found that to lower the error rates it becomes important to treat the model like an amateur for whatever the task is. Being detailed with the ask, such as noting any logical loopholes in your ask or difficulties, can go a long way to lowering that error rate.

    Likewise, amateurs need instructions a bit at a time. Give a new hire a 500 page manual and tell em’ to get to work and watch them fail. Give that new hire a single page with basic instructions for a given task, and your error rate drastically lowers.

    This phenomenon has a name no less: “Cognitive Load”, and is usually utilized in eLearning circles to reduce cognitive load on who you’re trying to teach (or give instructions to).

    These are all things I’ve tested myself, extensively, on my own local AI lab that I invested in when this hubbub all started. So by what I’m seeing there’s a pattern here.

    A lot of this was also for work. We released an AI assistant, and I needed to reduce the error rate. And looking at our public opinion on the feature, looks to be working too!

    1. Have you read “Hitchhikers Guide to the Galaxy”? Do you remember when the computer Deep Thought was asked the answer to the question of “Life, the Universe and Everything”? It took 7 million years and came back with “42”. It seems like you’re describing a similar situation, if you ask the AI an ill defined question it will come back with an stupid or useless answer, but if you ask it a simple, well explained and thought out question, you will usually get a useful response. Did I understand you right or am I making an NI mistake?

      1. That’s not how LLMs work. You can pose the most carefully crafted, detailed question, but if an actual human hasn’t actually thought about it, nothing remotely correct will be in the training set and the output will be gibberish.

        1. “if an actual human hasn’t actually thought about it, nothing remotely correct will be in the training set and the output will be gibberish.”

          … and someone has pointed out that the good news is because these things proliferate the training set will contain more and more output from them. I was already worried about the poisoning of the global knowledge pool by fake facts, now I have given up all hope for the future of the human race.

          1. “I was already worried about the poisoning of the global knowledge pool by fake facts”

            Yeah, this was already a serious issue due to sensationalist science journalism: in upper-level physics I can’t tell you the amount of completely wrong things that are stupidly common on the Internet. This is just gonna make it a hundred times worse.

            The biggest issue is that with theoretical stuff, the scientific method can’t fix it, because it’s all conceptual crap. So it doesn’t matter that it’s wrong and doesn’t make sense – because it’s all ungrounded theory no one notices.

      1. Or rather a filter for seeing who needs better training to do their job, an academic can be the worst coder and a autodidact can be the best coder you have ever seen. The autodidact might just need a different approach when learning a new API or a new language.

    2. This is similar to my experience in the coding realm. AI is a useful tool, but you can ask it to do too much and it will valiantly and confidently give an incorrect implementation. I find it to be hugely valuable at reducing “scut” work like the autogeneration of function documents and routine accessors and the like. It saves a lot of typing, provided an actual software developer is paying attention to the output. I would not put something like Copilot in the hands of a junior developer. It takes some degree of experience to recognize and reject gibberish.

      1. “It saves a lot of typing, provided an actual software developer is paying attention to the output.”

        How is this useful, though?? Seriously, how long would it take to create a script/macro which just… does it for you perfectly without burning up half the Amazon?

        We’ve been dealing with boilerplate for decades now: everyone generation so far has managed to create tools to help them that didn’t require more computing power than the actual thing being developed.

  3. Great post from Bruce Schneier. There are most of the time, to me, something worrying and alien in ai texts, erring differently than humans is probably part of the explanation.

    1. All AI generated material I have read so far are too shallow and evasive. It’s like explaining a Cake by enumerating and explaining how evey single ingredient is produced, but completely missing the recipe.

  4. Maybe first and foremost, we should stop using AI for what it is not made to do. Just like you mix a cake batter with a whisk, and not a hammer, LLMs are made to deal with text, without any obligation of fully respecting semantics.

    ChatGPT for an example, is made to look as if it was a human writing, and conveniently has access to a dataset that encompass large portion of the internet, but it doesn’t know right or wrong. Ask some math to it, specially calculus, he shows a lot of things that look as if it makes sense, then check the results against wolfram alpha, and its just gibberish. Ask him about academic publications and he will create fake papers, authors and even DOIs for the fake papers, but it feels legit.

    LLMs are text tools, not facts tools, they will create good text, with 0 guaratee of having any real facts in it.

    AI in general, has nothing of that Intelligence everyone get so excited about. It is just statistics. The catch isn’t on the dataset it uses, but rather, what it is looking into inside the dataset?
    ChatGPT looks into syntax, grammar, and some other things to make texts that look right.

  5. Tried to write a comment, but it went poof.
    So! TLDR Version: Use techniques to control cognitive load and you’ll reduce the error rate.

    Source: Built a major AI product with a major company that’s now doing complex tasks with grace and acceptable error rates to make it a useful tool for the product.

  6. I fail to see the great insight in this observation. It’s the fallacy to see these things as human, because they mime human experiences. But in human terms “there is nobody home” and the hope of researchers that this is “something that’ll just shake out” given enough data… is probably just “toxic positivity” (favorite term; explains so much, teams BSing themselves into better outcomes).

    But one thing is for sure, things will have to get worse before they get better. There is too much money and momentum behind this technology, you WILL PAY for it. I just hope your self driving car won’t put me in a wheelchair, because my heatlhcare provider AI has become very good not paying for it.

  7. Oh, also AIs don’t make mistakes, they have perfect replay, just save the state (pseudo random included) and run it again, same result.

    The people that make, sell and operate AIs make mistakes. They just don’t want to be responsible for them.

      1. We’re discovering an interesting corollary:

        When you’re building a statistical average of a large amount of input, you don’t get the peak ability of that input. You get a mediocre average idiot.

  8. I’ve barely used chatgpt but Ive seen some pretty funny mistakes. I had it generate some code to initialize registers for the pi pico and it actually did pretty well, after a few bug fixes. But one issue was that a register had a value that seemed incorrect. So I asked it to show me the calculation. It responded with a series of nicely formatted latek style calculation steps, in such detail it would be considered patronizing if it was human. And one of the steps was something like 4000/50 = 200. So I respond “4000/50 is not 200”. And it hilariously goes “oh I’m sorry, that is wrong” and then it redid the calculation correctly. Lol.

    I think what makes it weird is that it can makes extremely dumb mistakes, while simultaneously writing with impeccable grammar and formatting. Humans don’t usually work that way.

  9. LLM gives you 3 tipes os answers:
    One obious, one truth and one wrong…and your job is find wich one is wich one.

    I really think LLMs are the future of software development, but as they are not perfect, its a problem find where they fail cause they make better and better good looking code so its harder ans harder find where the bug is hidden.

    Good for us, the solution of code is not the same as hard as find if an answer is a real truth, in software development the truth is what makes the test pass, so you should pay attention to the tests, even more, generate the tests in a isolated way from the code(if this is even possible). Also, once you have the tests you need, you can generate more and more iterations for “free” until all the tests pass.

      1. a clock running backwards at 43200x speed is right any second…

        if you can ask a question so precise that a LLM gives a reliable answer, then a) you can already answer the question yourself and b) finding the right answer takes less effort than finding the right prompt for the LLM

  10. What you think you know about AI was probably 100% correct 6 months ago, and it will probably be 100% wrong (or irrelevant) in 6 months from now. This is one pattern I have observed in the chatter on X over recent years, a lot of people looking foolish for not considering the temporal context of any factoid. Not a new phenomena but one that is, due to the rapidly accelerating pace of change, becoming increasingly impactful on people’s reputations.

    1. OpenAI has had a colossal trainwreck trying to deliver a GPT 5. Everyone in the industry is stagnant, training near-identical models on near-identical datasets. It’d been literal years since the last big breakthrough. Nobody has any ideas at all about how to move the field forward. We have already passed peak LLM.

  11. “Models may unpredictably appear to lack common sense.”
    No, in fact they actually do have zero sense of any kind. They are returning the likely next action according to their training combined with a small amount of randomness for the appearance of variety. They don’t have sense, intelligence, thoughts, or value systems.

  12. “Next word predictor can’t demonstrate intent, planning, or any type of higher-level thinking. In fact maybe the first clumsy attempts at AI aren’t similar to how human minds work at all. More news at 7”

  13. If we sat down for an hour dinner and you started talking for 10 minutes then at the end of that dinner I pulled out a piece of paper that predicted every word you said after that 10 minutes, would that make you think different about yourself? What can you infer about me?

    Do I understand you? Do I have a mental model of you?

Leave a Reply

Please be kind and respectful to help make the comments section excellent. (Comment Policy)

This site uses Akismet to reduce spam. Learn how your comment data is processed.