• MudMan@fedia.io
      link
      fedilink
      arrow-up
      2
      ·
      7 hours ago

      The idea is having tensor acceleration built into SoCs for portable devices so they can run models locally on laptops, tablets and phones.

      Because, you know, server-side ML model calculations are expensive, so offloading compute to the client makes them cheaper.

      But this gen can’t really run anything useful locally so far, as far as I can tell. Most of the demos during the ramp-up to these were thoroughly underwhelming and nowhere near what you get from server-side services.

      Of course they could have just called the “NPU” a new GPU feature and make it work closer to how this is run on dedicated GPUs, but I suppose somebody thought that branding this as a separate device was more marketable.