Garbage in garbage out

Google put a box of AI-generated text at the top of their search results. If you don’t know what that means, I’m talking about the slow-loading colorful text box you’re seeing lately—it’s Google’s first step toward destroying the internet entirely (as I predicted last summer).

I am yet to meet anyone who is excited about this “feature”—at best people seem to begrudgingly accept it. But that doesn’t mean there isn’t joy to be found here: the feature is telling people to do absolutely bonkers things. Apparently Google thinks you can keep cheese from falling off pizza by mixing glue into the sauce, which was based on a decade-old joke from Reddit. Google is also telling people that geologists recommend eating “at least a few rocks per day”, a suggestion it found on The Onion. I could go on, but others already have.

What I find interesting is how this happens. Hype aside, large language models aren’t sentient. They are machines that are trying to guess the next word to say based on statistical probability—that’s it. The fact that this ever results in coherent answers to questions is staggering. The problem is that it’s factually wrong a lot, and getting it to be wrong less often is a very hard problem.

The glue pizza and the rock diet are both things that were posted to the internet as jokes. The machine, as it exists right now, can’t detect these sorts of jokes, so it repeats them as though they are facts. I do not know which engineer thought it would be a good idea to include Reddit and The Onion in their training data, but including that had a predictable effect: jokes presented as facts. Garbage in, garbage out.

This has me thinking about my own mind. Now, obviously, I am not a large language model—I’m sentient. I have an understanding of myself as a person and the ability to sense whether something is a joke or not from cultural context. And when I read something, I’m not “training” my brain in the sense that a large language model is. But there’s still a level on which thinking about the “garbage in, garbage out” is useful.

I am, as a person, the average of the people I hang out with and the information that I take in. If I spend an evening watching horrible people saying horrible things, that is going to have an impact. If I spend time around horrible people, I’m going to be more likely to treat others poorly. If I use every bit of downtime to scroll on a website full of outrage, I’m going to feel constantly outraged.

So I try not to do that. I have excellent friends, friends I admire, friends who make me better. And I try to spend less time thinking about the outrageous and more time thinking about things in context. For me that has meant replacing Twitter with Mastodon and Reddit with the home page of various news outlets. For you it might mean something else. The point is to think critically about the kind of information you’re taking in, because on some level it will form who you are. You do not want to end up telling people to put glue on pizza.

10 Comments

  1. @JustinPotBlog WELL SAID! Thought we were getting better but I think we get stupider as time goes on.
    People don’t think, can’t think anymore, they all search the internet. Then believe what they read cause the internet is truth but it is actually not. I don’t believe anything I read on the internet. Like you said here – garbage in gives garbage out.

  2. @JustinPotBlog

    No, Justin, you're NOT 'an average of the people you hang out with'; the idea that you might be gives so-called 'AI' Samplebots more credit than they deserve.

    You 'parse' information & people constantly. It's what we mean by the word 'thinking'. Your judgment makes it possible to fall out with friends (perhaps based more on their treatment of YOU than on other behaviours, but still…) and to assess whether a piece of information is humorous.

    The Internet will survive…

Comments are closed.