Other platforms too, but I’m on lemmy. I’m mainly talking about LLMs in this post

First, let me acknowledge that AI is not perfect, it has limitations e.g

  • tendency to hallucinate responses instead of refusing/saying it doesn’t know
  • different models/models sizes with varying capabilities
  • lack of knowledge of recent topics without explicitly searching it
  • tendency to be patternistic/repetitive
  • inability to hold on to too much context at a time etc.

But I really don’t think they’re nearly as bad as most lemmy users make them out to be. I was going to respond to all the takes but there’s so many I’ll just make some general points

  • SOTA (State of the Art) models match or beat most humans besides experts in most fields that are measurable
  • I personally find AI is better than me in most fields except ones I know well. So maybe it’s only 80-90% there, but it’s there in like every single field whereas I am in like 1-2
  • LLM’s can also do all this in like 100 languages. You and I can do it in like… 1, with limited performance in a couple others
  • Companies often use smaller/cheaper models in various products (e.g google search), which are understandably much worse. People often then use these to think all AI sucks
  • LLM’s aren’t just memorizing their training data. They can reason, as recent reasoning models more clearly show. Also, we now have near frontier models that are like 32B, or 21B GB in size. You cannot fit the entire internet in 21GB. There is clearly higher level synthesizing going on
  • People often tend to seize on superficial questions like the strawberry question (which is essentially an LLM blind spot) to claim LLM’s are dumb.
  • In the past few years, researchers have had to come up with countless newer harder benchmarks because LLMs kept blowing through previous ones (partial list here: https://r0bk.github.io/killedbyllm/)

Also I did say willfully ignorant. This is because you can go and try most models for yourself right now. There are also endless benchmarks constantly being published showing how well they are doing. Benchmarks aren’t perfect and are increasingly being gamed, but they are still decent.

  • socsa@piefed.social
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 days ago

    I tend to agree - for whatever reason tech heavy internet communities tend to be filled with Luddites.

    My organization has found LLMs really useful for interdisciplinary collaboration. Subject matter experts can generate and check code examples, while software engineers can ask questions about what the code is meant to be doing. It really lowers the friction in these kinds of interactions