Abacus.ai:

We recently released Smaug-72B-v0.1 which has taken first place on the Open LLM Leaderboard by HuggingFace. It is the first open-source model to have an average score more than 80.

  • Miss Brainfarts
    link
    fedilink
    English
    15 months ago

    I may need to lower it a bit more, yeah. Though when I try to to use offloading, I can see that vram usage doesn’t increase at all.

    When I leave the setting at its default 100 value on the other hand, I see vram usage climb until it stops because there isn’t enough of it.

    So I guess not all models support offloading?

    • @General_Effort@lemmy.world
      link
      fedilink
      English
      45 months ago

      Most formats don’t support it. It has to be gguf format, afaik. You can usually find a conversion on huggingface. Prefer offerings by TheBloke for the detailed documentation, if nothing else.

    • @Fisch@lemmy.ml
      link
      fedilink
      English
      25 months ago

      The models you have should be .gguf files right? I think those are the only ones where that’s supported