• Ace@feddit.uk
    link
    fedilink
    English
    arrow-up
    3
    ·
    4 days ago

    sure, but I tried this and they all suck. I only have 8GB of ram so can only use the smaller versions of the models, and they’re much much worse at just making up random shit

    • NeilBrü@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      2 days ago

      Oof, ok, my apologies.

      I am, admittedly, “GPU rich”; I have ~48GB of VRAM at my disposal on my main workstation, and 24GB on my gaming rig. Thus, I am using Q8 and Q6_L quantized .gguf files.

      Naturally, my experience with the “fidelity” of my LLM models re: hallucinations would be better.