• mustbe3to20signs@feddit.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    AI models can outmatch most oncologists and radiologists in recognition of early tumor stages in MRI and CT scans.
    Further developing this strength could lead to earlier diagnosis with less-invasive methods saving not only countless live and prolonging the remaining quality life time for the individual but also save a shit ton of money.

    • msage@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Wasn’t it proven that AI was having amazing results, because it noticed the cancer screens had doctors signature at the bottom? Or did they make another run with signatures hidden?

      • mustbe3to20signs@feddit.org
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        There were more than one system proven to “cheat” through biased training materials. One model used to tell duck and chicken apart because it was trained with pictures of ducks in the water and chicken on a sandy ground, if I remember correctly.
        Since multiple multiple image recognition systems are in development, I can’t imagine they’re all this faulty.

        • msage@programming.dev
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 months ago

          They are not ‘faulty’, they have been fed wrong training data.

          This is the most important aspect of any AI - it’s only as good as the training dataset is. If you don’t know the dataset, you know nothing about the AI.

          That’s why every claim of ‘super efficient AI’ need to be investigated deeper. But that goes against line-goes-up principle. So don’t expect that to happen a lot.