• NeilBrü@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        2 days ago

        Oof, ok, my apologies.

        I am, admittedly, “GPU rich”; I have at ~48GB of VRAM at my disposal on my main workstation, and 24GB on my gaming rig. Thus, I am using Q8 and Q6_L quantized GGUFs.

        Naturally, my experience with the “fidelity” of my LLM models re: hallucinations would be better.

    • anus@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      8
      ·
      2 days ago

      I actually think that (presently) self hosted LLMs are much worse for hallucination