• SabinStargem@lemmy.today
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    3
    ·
    1 day ago

    It is going to take at least five years before local AI is user-friendly enough and with performant hardware circulating, that ordinary folks would consider buying an AI machine.

    I have a top-end DDR4 gaming rig. It takes a long time for 100b sized AI to give some roleplaying output, at least forty minutes for my settings via KoboldCPP with a GGUF. I don’t think a typical person would want to wait more than 2 minutes for a good response. So we will need at least DDR6 era devices before it is practical for everyday people.

    • lmuel@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      17
      arrow-down
      1
      ·
      1 day ago

      A local LLM is still an LLM… I don’t think it’s gonna be terribly useful no matter how good your hardware is

      • maus@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        1 day ago

        I have great success with local LLM in some of my workflows and automation.

        I use it for my line completion and basic functions/asks while developing that I dont want to waste tokens on.

        I also use it in automation. I run my own media server with a few dozen people with an automated request system “jellyseerr” that adds content. I have automation that leverages local LLM to look at recent media requests and automatically requests content that is similar to it.

      • luridness@lemmy.ml
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 day ago

        Local AI can be useful. But I would rather see nice implementations that used small but brilliantly tuned models for… let’s say better predictive text… it’s already somewhat AI based I just would like it to be. Better

      • xthexder@l.sw0.com
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 day ago

        The diminishing returns are kind of insane if you compare the performance and hardware requirements of a 7b and 100b model. In some cases the smaller model can even perform better because it’s more focused and won’t be as subtle about its hallucinations.
        Something is going to have to fundamentally change before we see any big improvements, because I don’t see scaling it up further ever producing AGI or even solving any of the hallucinations/ logic errors it makes.

        In some ways it’s a bit like the Crypto blockchain speculators saying it’s going to change the world. But in reality the vast majority of applications proposed would have been better implemented with a simple centralized database.