• NotMyOldRedditName@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    15 hours ago

    It does work, but it’s not really fast. I upgraded to 96gb ddr4 from 32gb a year or so ago, and being able to play with the bigger models was fun, but it’s not something I could do anything productive with it was so slow.

    • Possibly linux@lemmy.zip
      link
      fedilink
      English
      arrow-up
      4
      ·
      15 hours ago

      Your bottle necked by memory bandwidth

      You need ddr5 with lots of memory channels for it to he useful

    • tal@lemmy.today
      link
      fedilink
      English
      arrow-up
      3
      ·
      15 hours ago

      You can have applications where wall clock tine time is not all that critical but large model size is valuable, or where a model is very sparse, so does little computation relative to the size of the model, but for the major applications, like today’s generative AI chatbots, I think that that’s correct.

      • NotMyOldRedditName@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        15 hours ago

        Ya, that’s fair. If I was doing something I didn’t care about time on, it did work. And we weren’t talking hours, it it could be many minutes though.