• sundray@lemmus.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    AI writing, scraped by AI, producing more AI writing…

    So not “gray goo” exactly, but “gray slop”?

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      And they’re overlooking that radionuclide contamination of steel actually isn’t much of a problem any more, since the surge in background radionuclides caused by nuclear testing peaked in 1963 and has since gone down almost back to the original background level again.

      I guess it’s still a good analogy, though. People bring up Low Background Steel because they think radionuclide contamination is an unsolved problem (despite it having been basically solved), and they bring up “model collapse” because they think it’s an unsolved problem (despite it having been basically solved). It’s like newspaper stories, everyone sees the big scary front page headline but nobody pays attention to the little block of text retracting it on page 8.

  • KevonLooney@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    2 months ago

    provenance requires some way to filter the internet into human-generated and AI-generated content, which hasn’t been cracked yet

    It doesn’t need to be filtered into human / AI content. It needs to be filtered into good (true) / bad (false) content. Or a “truth score” for each.

    We don’t teach children to read by just handing them random tweets. We give them books that are made specifically for children. Our filtering mechanism for good / bad content is very robust for humans. Why can’t AI just read every piece of “classic literature”, famous speeches, popular books, good TV and movie scripts, textbooks, etc?

    • Lvxferre@mander.xyz
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      It doesn’t need to be filtered into human / AI content. It needs to be filtered into good (true) / bad (false) content. Or a “truth score” for each.

      That isn’t enough because the model isn’t able to reason.

      I’ll give you an example. Suppose that you feed the model with both sentences:

      1. Cats have fur.
      2. Birds have feathers.

      Both sentences are true. And based on vocabulary of both, the model can output the following sentences:

      1. Cats have feathers.
      2. Birds have fur.

      Both are false but the model doesn’t “know” it. All that it knows is that “have” is allowed to go after both “cats” and “birds”, and that both “feathers” and “fur” are allowed to go after “have”.

      • KevonLooney@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        It’s not just a predictive text program. That’s been around for decades. That’s a common misconception.

        As I understand it, it uses statistics from the whole text to create new text. It would be very rare to output “cats have feathers” because that phrase doesn’t ever appear in the training data. Both words “have feathers” never follow “cats”.

        • barsoap@lemm.ee
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          2 months ago

          because that phrase doesn’t ever appear in the training data.

          Eh but LLMs abstract. It has seen “<animal> have feathers” and “<animal> have fur” quite a lot of times. The problem isn’t that LLMs can’t reason at all, the problem is that they do employ techniques used in proper reasoning, in particular tracking context throughout the text (cross-attention) but lack techniques necessary for the whole thing, instead relying on confabulation to sound convincing regardless of the BS they spout. Suffices to emulate an Etonian but that’s not a high standard.

        • vrighter@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          2 months ago

          and that is exactly how a predictive text algorithm works.

          • some tokens go in

          • they are processed by a deterministic, static statistical model, and a set of probabilities (always the same, deterministic, remember?) comes out.

          • pick the word with the highest probability, add it to your initial string and start over.

          • if you want variety, add some randomness and don’t just always pick the most probable next token.

          Coincidentally, this is exactly how llms work. It’s a big markov chain, but with a novel lossy compression algorithm on its state transition table. The last point is also the reason why, if anyone says they can fix llm hallucinations, they’re lying.

          • CeeBee_Eh@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 months ago

            Coincidentally, this is exactly how llms work

            Everyone who says this doesn’t actually understand how LLMs work.

            Multivector word embeddings create emergent relationships that’s new knowledge that doesn’t exist in the training dataset.

            Computerphile did a good video on this well before the LLM craze.

      • CeeBee_Eh@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Both sentences are true. And based on vocabulary of both, the model can output the following sentences:

        1. Cats have feathers.
        2. Birds have fur

        This is not how the models are trained or work.

        Both are false but the model doesn’t “know” it. All that it knows is that “have” is allowed to go after both “cats” and “birds”, and that both “feathers” and “fur” are allowed to go after “have”.

        Demonstrably false. This isn’t how LLMs are trained or built.

        Just considering the contextual relationships between word embeddings that are created during training is evidence enough. Those relationships from the multi-vector fields are an emergent property that doesn’t exist in the dataset.

        If you want a better understanding of what I just said, take a look at this Computerphile video from four years ago. And this came out before the LLM hype and before ChatGPT 3, which was the big leap in LLMs.

    • Zos_Kia@lemmynsfw.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      That’s what smaller models do, but it doesn’t yield great performance because there’s only so much stuff available. To get to gpt4 levels you need a lot more data, and to break the next glass ceiling you’ll need even more.

      • KevonLooney@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Then these models are stupid. Humans don’t start as a blank slate. They have an inherent aptitude for language and communication. These models should start out with basics of language, so they don’t have to learn it from the ground up. That’s the next step. Right now they’re just well read idiots.

        • Zos_Kia@lemmynsfw.com
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          Then these models are stupid

          Yup that is kind of the point. They are math functions designed to approximate human tasks.

          These models should start out with basics of language, so they don’t have to learn it from the ground up. That’s the next step. Right now they’re just well read idiots.

          I’m not sure what you’re pointing at here. How they do it right now, simplified, is you have a small model designed to cut text into tokens (“knowledge of syllables”), which are fed into a larger model which turns tokens into semantic information (“knowledge of language”), which is fed to a ridiculously fat model which “accomplishes the task” (“knowledge of things”).

          The first two models are small enough that they can be trained on the kind of data you describe, classic books, movie scripts etc… A couple hundred billion words maybe. But the last one requires orders of magnitude more data, in the trillions.

  • Lvxferre@mander.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Model degeneration is an already well-known phenomenon. The article already explains well what’s going on so I won’t go into details, but note how this happens because the model does not understand what it is outputting - it’s looking for patterns, not for the meaning conveyed by said patterns.

    Frankly at this rate might as well go with a neuro-symbolic approach.

    • CeeBee_Eh@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      The issue with your assertion is that people don’t actually work a similar way. Have you ever met someone who was clearly taught "garbage’?

      • PenisDuckCuck9001@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 months ago

        I’m autistic and sometimes I feel like an ai bot spewing out garbage in social situations. If I do what people normally do and make it sound believable, maybe no one will notice.

      • Lvxferre@mander.xyz
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        The issue with your assertion is that people don’t actually work a similar way.

        I’m talking about LLMs, not about people.

        • CeeBee_Eh@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          I know you are, but the argument that an LLM doesn’t understand context is incorrect. It’s not human level understanding, but it’s been demonstrated that they do have a level of understanding.

          And to be clear, I’m not talking about consciousness or sapience.

          • CileTheSane@lemmy.ca
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 months ago

            but it’s been demonstrated that they do have a level of understanding.

            Citation needed

              • CileTheSane@lemmy.ca
                link
                fedilink
                English
                arrow-up
                0
                ·
                edit-2
                2 months ago

                A better mathematical system of storing words does not mean the LLM understands any of them. It just has a model that represents the relation between words that it uses.

                If I put 10 minus 8 into my calculator I get 2. The calculator doesn’t actually understand what 2 means, or what subtracting represents, it just runs the commands that gives the appropriate output.

                • CeeBee_Eh@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  2 months ago

                  That’s a bad analogy, because the calculator wasn’t trained using an artificial neural network literally designed by studying biological brains (aka biological neutral networks).

                  And “understand” doesn’t equate to consciousness or sapience. For example, it is entirely and factually correct to state that an LLM is capable of reasoning. That’s not even up for debate. The accuracy of an LLM’s reasoning capability is one of the fundamental benchmarks used for evaluating its quality.

                  But that doesn’t mean it’s “thinking” in the way most people consider.

          • Lvxferre@mander.xyz
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 months ago

            I know you are, but the argument that an LLM doesn’t understand context is incorrect

            Emphasis mine. I am talking about the textual output. I am not talking about context.

            It’s not human level understanding

            Additionally, your obnoxiously insistent comparison between LLMs and human beings boils down to a red herring.

            Not wasting my time further with you.

            [For others who might be reading this: sorry for the blatantly rude tone but I got little to no patience towards people who distort what others say, like the one above.]

            • CeeBee_Eh@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 months ago

              I got little to no patience towards people who distort what others say,

              My original reply was meant to be tongue-in-cheek, but I guess I forgot about Poe’s law. I’m not a layman, for the record. I’ve worked with AI for over a decade

              Not wasting my time further with you.

              Ditto. Have a nice day.

  • Anarki_@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    ⢀⣠⣾⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⠀⠀⠀⠀⣠⣤⣶⣶ ⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⠀⠀⠀⢰⣿⣿⣿⣿ ⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣧⣀⣀⣾⣿⣿⣿⣿ ⣿⣿⣿⣿⣿⡏⠉⠛⢿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⡿⣿ ⣿⣿⣿⣿⣿⣿⠀⠀⠀⠈⠛⢿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⣿⠿⠛⠉⠁⠀⣿ ⣿⣿⣿⣿⣿⣿⣧⡀⠀⠀⠀⠀⠙⠿⠿⠿⠻⠿⠿⠟⠿⠛⠉⠀⠀⠀⠀⠀⣸⣿ ⣿⣿⣿⣿⣿⣿⣿⣷⣄⠀⡀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢀⣴⣿⣿ ⣿⣿⣿⣿⣿⣿⣿⣿⣿⠏⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠠⣴⣿⣿⣿⣿ ⣿⣿⣿⣿⣿⣿⣿⣿⡟⠀⠀⢰⣹⡆⠀⠀⠀⠀⠀⠀⣭⣷⠀⠀⠀⠸⣿⣿⣿⣿ ⣿⣿⣿⣿⣿⣿⣿⣿⠃⠀⠀⠈⠉⠀⠀⠤⠄⠀⠀⠀⠉⠁⠀⠀⠀⠀⢿⣿⣿⣿ ⣿⣿⣿⣿⣿⣿⣿⣿⢾⣿⣷⠀⠀⠀⠀⡠⠤⢄⠀⠀⠀⠠⣿⣿⣷⠀⢸⣿⣿⣿ ⣿⣿⣿⣿⣿⣿⣿⣿⡀⠉⠀⠀⠀⠀⠀⢄⠀⢀⠀⠀⠀⠀⠉⠉⠁⠀⠀⣿⣿⣿ ⣿⣿⣿⣿⣿⣿⣿⣿⣧⠀⠀⠀⠀⠀⠀⠀⠈⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢹⣿⣿ ⣿⣿⣿⣿⣿⣿⣿⣿⣿⠃⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢸⣿⣿

  • tal@lemmy.today
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    Well, you’ve got a timestamped copy of much of the Web that existed up until latent-diffusion models at archive.org. That may not give you access to newer information, but it’s a pretty whopping big chunk of data to work with.

    • palordrolap@kbin.run
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      Hopefully archive.org have measures in place to stop people from yanking all their data too quickly. As least not without a hefty donation or something. As a user it can chug a bit, and I’m hoping that’s the rate-limiting I’m talking about and not that they’re swamped.

      • Grimy@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        2 months ago

        That would go against the principal of the archive imo but regardless, if you take away all means of acquiring data freely, you are just giving companies like OpenAI and Google who already have copies of it an insane advantage.

        AI isn’t going away, we need to make sure we have free access to it as to not give our whole economy to a handful of companies.