• 0 Posts
  • 3 Comments
Joined 1 year ago
cake
Cake day: November 10th, 2023

help-circle
  • 13B and 20B Noromaid for RP/ERP.

    I am experimenting with comparing GGUF to EXL2 as well as stretching context. So far, Noromaid 13b at GGUF Q5_K_M stretches to 12k context on a 3090 without issues. Noromaid 20B at Q3_K_M stretches to 8k without issues and is in my opinion superior to the 13B. I have recently stretched Noromaid 20B to 10k using 4bpw EXL2 and it is giving coherent responses. I haven’t used it enough to assess the quality however.

    All this is to say, if you enjoy roleplay you should be giving Noromaid a look.



  • USM-Valor@alien.topBtoLocalLLaMAWhat UI do you use and why?
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Backend: 99% of the time, KoboldCPP, 1% of the time (testing EXL2 etc) Ooba

    Front End: Silly Tavern

    Why: GGUF is my preferred model type, even with a 3090. KoboldCPP is the best that I have seen at running this model type. SillyTavern should be obvious, but it is updated multiple times a day and is amazingly feature rich and modular.