• NeilBrü@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    2 days ago

    Oof, ok, my apologies.

    I am, admittedly, “GPU rich”; I have at ~48GB of VRAM at my disposal on my main workstation, and 24GB on my gaming rig. Thus, I am using Q8 and Q6_L quantized GGUFs.

    Naturally, my experience with the “fidelity” of my LLM models re: hallucinations would be better.