• Viri4thus@feddit.org
    link
    fedilink
    arrow-up
    1
    ·
    15 days ago

    Both video and DLSS use(d) diffusion to upscale images (DLSS has allegedly ttransitioned to a transformer model). AFAIK there’s no simple way to run diffusion on an NPU as of today.

    Regarding local running LLMs, well, I’ll take an NPU with 32-64 gigs of Ram over an anemic llama 1-3B model run on the GPU. And that’s before considering people using Windows and taking advantage of MS Olive. Llama3.3 70B, which has similar performance to Llama3.2 405B will run on 64GB of ram, ezpz, forget about ever running it on local PC with an NVIDIA card.

    My eyes are set on the strix halo 128GB variant, I’m going to put that through its paces.

    BTW, most of the interesting models will fail to run locally due to NVIDIA’s shit VRAM allowance, if nvidia were giving people a minimum of 16GB of VRAM I’m sure MS would happily certify it.

    • Alphane Moon@lemmy.worldOPM
      link
      fedilink
      arrow-up
      2
      ·
      15 days ago

      That’s fair. But do you see where I am coming from?

      Marketing around TOPs isn’t everything.

      Interesting is a relative term. I find upscaling older SD content interesting. You can’t just dismiss this use case because it doesn’t fit into your arguement.

      Getting a local LLM (Llama 1B is not as good as cloud LLMs of course, but it does have valid use cases) with a Nvidia GPU is extremely simple. Can you provide a 5 bullet point guide for setting up a local LLM with 32 GB RAM (64 GB RAM isn’t that common in laptops).

      • Viri4thus@feddit.org
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        15 days ago

        Install lmstudio

        Profit

        *If you want to use the NPU

        Apply for beta branch (3.6.x) at lmstudio

        Install lmstudio beta

        Profit

        Edit: Almost forgot, the AMD drivers (under review) for the latest NPU containing CPUs (7xxx and upward) should come with the spring kernel update to 6.3, fingers crossed. It’s been two years, they took their sweet time. Windows support was available on release…