• NeilBrü@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    11 hours ago

    Oof, ok, my apologies.

    I am, admittedly, “GPU rich”; I have ~48GB of VRAM at my disposal on my main workstation, and 24GB on my gaming rig. Thus, I am using Q8 and Q6_L quantized .gguf files.

    Naturally, my experience with the “fidelity” of my LLM models re: hallucinations would be better.