I have many conversations with people about Large Language Models like ChatGPT and Copilot. The idea that “it makes convincing sentences, but it doesn’t know what it’s talking about” is a difficult concept to convey or wrap your head around. Because the sentences are so convincing.

Any good examples on how to explain this in simple terms?

Edit:some good answers already! I find especially that the emotional barrier is difficult to break. If an AI says something malicious, our brain immediatly jumps to “it has intent”. How can we explain this away?

  • CodeInvasion@sh.itjust.works
    link
    fedilink
    arrow-up
    28
    arrow-down
    1
    ·
    edit-2
    2 months ago

    I am an LLM researcher at MIT, and hopefully this will help.

    As others have answered, LLMs have only learned the ability to autocomplete given some input, known as the prompt. Functionally, the model is strictly predicting the probability of the next word+, called tokens, with some randomness injected so the output isn’t exactly the same for any given prompt.

    The probability of the next word comes from what was in the model’s training data, in combination with a very complex mathematical method to compute the impact of all previous words with every other previous word and with the new predicted word, called self-attention, but you can think of this like a computed relatedness factor.

    This relatedness factor is very computationally expensive and grows exponentially, so models are limited by how many previous words can be used to compute relatedness. This limitation is called the Context Window. The recent breakthroughs in LLMs come from the use of very large context windows to learn the relationships of as many words as possible.

    This process of predicting the next word is repeated iteratively until a special stop token is generated, which tells the model go stop generating more words. So literally, the models builds entire responses one word at a time from left to right.

    Because all future words are predicated on the previously stated words in either the prompt or subsequent generated words, it becomes impossible to apply even the most basic logical concepts, unless all the components required are present in the prompt or have somehow serendipitously been stated by the model in its generated response.

    This is also why LLMs tend to work better when you ask them to work out all the steps of a problem instead of jumping to a conclusion, and why the best models tend to rely on extremely verbose answers to give you the simple piece of information you were looking for.

    From this fundamental understanding, hopefully you can now reason the LLM limitations in factual understanding as well. For instance, if a given fact was never mentioned in the training data, or an answer simply doesn’t exist, the model will make it up, inferring the next most likely word to create a plausible sounding statement. Essentially, the model has been faking language understanding so much, that even when the model has no factual basis for an answer, it can easily trick a unwitting human into believing the answer to be correct.

    —-

    +more specifically these words are tokens which usually contain some smaller part of a word. For instance, understand and able would be represented as two tokens that when put together would become the word understandable.

    • HamsterRage@lemmy.ca
      link
      fedilink
      arrow-up
      8
      ·
      2 months ago

      I think that a good starting place to explain the concept to people would be to describe a Travesty Generator. I remember playing with one of those back in the 1980’s. If you fed it a snippet of Shakespeare, what it churned out sounded remarkably like Shakespeare, even if it created brand “new” words.

      The results were goofy, but fun because it still almost made sense.

      The most disappointing source text I ever put in was TS Eliot. The output was just about as much rubbish as the original text.

    • Sabata11792@kbin.social
      link
      fedilink
      arrow-up
      4
      ·
      2 months ago

      As some nerd playing with various Ai models at home with no formal training, any wisdom you think that’s worth sharing?

  • IzzyScissor@lemmy.world
    link
    fedilink
    arrow-up
    19
    arrow-down
    1
    ·
    2 months ago

    It’s your phone’s ‘predictive text’, but if it were trained on the internet.

    It can guess what the next word should be a lot of the time, but it’s also easy for it to go off the rails.

  • GamingChairModel@lemmy.world
    cake
    link
    fedilink
    arrow-up
    15
    ·
    2 months ago

    Harry Frankfurt’s influential 2005 book (based on his influential 1986 essay), On Bullshit, offered a description of what bullshit is.

    When we say a speaker tells the truth, that speaker says something true that they know is true.

    When we say a speaker tells a lie, that speaker says something false that they know is false.

    But bullshit is when the speaker says something to persuade, not caring whether the underlying statement is true or false. The goal is to persuade the listener of that underlying fact.

    The current generation of AI chat bots are basically optimized for bullshit. The underlying algorithms reward the models for sounding convincing, not necessarily for being right.

  • Deconceptualist@lemm.ee
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    2
    ·
    2 months ago

    You could maybe just share a meme like this one.

    Some folks in the comments there share actual LLM results, a few of which are sensible but plenty that aren’t far off from the joke.

  • BlameThePeacock@lemmy.ca
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    2
    ·
    2 months ago

    It’s just fancy predictive text like while texting on your phone. It guesses what the next word should be for a lot more complex topics.

    • kambusha@sh.itjust.works
      link
      fedilink
      arrow-up
      4
      ·
      2 months ago

      This is the one I got from the house to get the kids to the park and then I can go to work and then I can go to work and get the rest of the day after that I can get it to you tomorrow morning to pick up the kids at the same time as well as well as well as well as well as well as well as well as well… I think my predictive text broke

    • k110111@feddit.de
      link
      fedilink
      arrow-up
      3
      ·
      2 months ago

      Its like saying an OS is just a bunch of if then else statements. While it is true, in practice it is far far more complicated.

  • Feathercrown@lemmy.world
    link
    fedilink
    English
    arrow-up
    9
    ·
    2 months ago

    The idea that “it makes convincing sentences, but it doesn’t know what it’s talking about” is a difficult concept to convey or wrap your head around.

    I see the people you talk to aren’t familiar with politicians?

  • rubin@lemmy.sdf.org
    link
    fedilink
    arrow-up
    9
    ·
    2 months ago

    Imagine that you have a random group of people waiting in line at your desk. You have each one read the prompt, and the response so far, and then add a word themself. Then they leave and the next person in line comes and does it.

    This is why “why did you say ?” questions are nonsensical to AI. The code answering it is not the code that wrote it and there is no communication coordination or anything between the different word answerers.

    • relevants@feddit.de
      link
      fedilink
      arrow-up
      1
      ·
      2 months ago

      Ok, I like this description a lot actually, it’s a very quick and effective way to explain the effects of no backtracking. A lot of the answers here are either too reductive or too technical to actually make this behavior understandable to a layman. “It just predicts the next word” is easy to forget when the thing makes it so easy to be anthropomorphized subconsciously.

    • kaffiene@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 months ago

      In the sense that the “argument” is an intuition pump. As an anti ai argument it’s weak - you could replace the operator in the Chinese room with an operator in an individual neuron and conclude that our brains don’t know anything, either

  • Hucklebee@lemmy.worldOP
    link
    fedilink
    arrow-up
    6
    ·
    edit-2
    2 months ago

    After reading some of the comments and pondering this question myself, I think I may have thought of a good analogy that atleast helps me (even though I know fairly well how LLM’s work)

    An LLM is like a car on the road. It can follow all the rules, like breaking in front of a red light, turning, signaling etc. However, a car has NO understanding of any of the traffic rules it follows.

    A car can even break those rules, even if its behaviour is intended (if you push the gas pedal at a red light, the car is not in the wrong because it doesn’t KNOW the rules, it just acts on it).

    Why this works for me is that when I give examples of human behaviour or animal behaviour, I automatically ascribe some sort of consciousness. An LLM has no conscious (as far as I know for now). This idea is exactly what I want to convey. If I think of a car and rules, it is obvious to me that a car has no concept of rules, but still is part of those rules somehow.

    • 1rre@discuss.tchncs.de
      link
      fedilink
      arrow-up
      3
      arrow-down
      1
      ·
      edit-2
      2 months ago

      Thing is a conscience (and any emotions, and feelings in general) is just chemicals affecting electrical signals in the brain… If a ML model such as an LLM uses parameters to affect electrical signals through its nodes then is it on us to say it can’t have a conscience, or feel happy or sad, or even pain?

      Sure the inputs and outputs are different, but when you have “real” inputs it’s possible that the training data for “weather = rain” is more downbeat than “weather = sun” so is it reasonable to say that the model gets depressed when it’s raining?

      The weightings will change leading to a a change in the electrical signals, which emulates pretty closely what happens in our heads

      • Hucklebee@lemmy.worldOP
        link
        fedilink
        arrow-up
        4
        ·
        edit-2
        2 months ago

        Doesn’t that depend on your view of consciousness and if you hold the view of naturalism?

        I thought science is starting to find more and more that a 100% naturalistic worldview is hard to keep up. (E: I’m no expert on this topic and the information and podcast I listen to are probably very biased towards my own view on this. The point I’m making is that to say “we are just neurons” is more a disputed topic for debate than actual fact when you dive a little bit into neuroscience)

        I guess my initial question is almost more philosophical in nature and less deterministic.

        • huginn@feddit.it
          link
          fedilink
          arrow-up
          3
          arrow-down
          1
          ·
          2 months ago

          I’m not positive I’m understanding your term naturalistic but no neuroscientist would say “we are just neurons”. Similarly no neuroscientist would deny that neurons are a fundamental part of consciousness and thought.

          You have plenty of complex chemical processes interacting with your brain constantly - the neurons there aren’t all of who you are.

          But without the neurons there: you aren’t anyone anymore. You cease to live. Destroying some of those neurons will change you fundamentally.

          There’s no disputing this.

          • Hucklebee@lemmy.worldOP
            link
            fedilink
            arrow-up
            1
            ·
            2 months ago

            I agree with you, and you worded what I was clumsily trying to say. Thank you:)

            With naturalism I mean the philosphical idea that only natural laws and forces are present in this world. Or as an extension, the idea that here is only matter.

  • Ziggurat@sh.itjust.works
    link
    fedilink
    arrow-up
    5
    ·
    2 months ago

    have you played that game where everyone write a subjet and put it on a stack of paper, then everyone puts a verb on a different stack of paper, then everyone put an object on a third stack of paper, and you can even add a place or whatever on the next stack of paper. You end-up with fun sentences like A cat eat Kevin’s brain on the beach. It’s the kind of stuff (pre-)teen do to have a good laugh.

    Chat GPT somehow works the same way, except that instead of having 10 paper in 5 stack, it has millions of paper in thousands of stack and depending on the “context” will choose which stack it draws paper from (To take an ELI5 analogy)

    • Hucklebee@lemmy.worldOP
      link
      fedilink
      arrow-up
      1
      ·
      2 months ago

      I think what makes it hard to wrap your head around is that sometimes, this text is emotionally charged. What I notice is that it’s especially hard if an AI “goes rogue” and starts saying sinister and malicious things. Our brain immediatly jumps to “it has bad intent” when in reality it’s jus taking some reddit posts where it happened to connect some troll messages or extremist texts.

      How can we decouple emotionally when it feels so real to us?

  • rufus@discuss.tchncs.de
    link
    fedilink
    arrow-up
    5
    arrow-down
    1
    ·
    2 months ago

    It’s like your 5 year old daughter, relaying to you what she made of something she heard earlier.

    That’s my analogy. ChatGPT kind of has the intellect and ability to differentiate between facts and fiction of a 5 year old. But it combines that with the writing style of a 40 year old with a uncanny love of mixing adjectives and sounding condescending.

  • DarkThoughts@fedia.io
    link
    fedilink
    arrow-up
    4
    ·
    2 months ago

    Someone once described is as T9 on steroids. It’s like your mobile keyboard suggesting follow up words, just a lot more complex in size.

    If an AI says something malicious, our brain immediatly jumps to “it has intent”. How can we explain this away?

    The more you understand the underlying concept of LLMs, the more the magic fades away. LLMs are certainly cool and can be fun but the hype around them seems very artificial and they’re certainly not what I’d describe as “AI”. To me, an AI would be something that actually has some form of consciousness, something that actually can form its own thoughts and learn things on its own through observation or experimentation. LLMs can’t do any of those things. They’re static and always wait for your input to even do anything. For text generation you can even just regenerate an answer to the same previous text and the replies can and will vary greatly. If they say something mean or malicious, it’s simply because it is based on whatever they were trained on and whatever parameters they are following (like if you told them to roleplay a mean person).

  • Hegar@kbin.social
    link
    fedilink
    arrow-up
    4
    ·
    edit-2
    2 months ago

    Part of the problem is hyperactive agency detection - the same biological bug/feature that fuels belief in the divine.

    If a twig snaps, it could be nothing or someone. If it’s nothing and we react as if it was someone, no biggie. If it was someone and we react as if it was nothing, potential biggie. So our brains are bias towards assuming agency where there is none, to keep us alive.