I’ve been using self-hosted LLM models for roleplay purposes. But these are the worst problems I face every time, no matter what model and parameter preset I use.
I’m using :
Pygmalion 13B AWQ
Mistral 7B AWQ
SynthIA 13B AWQ [Favourite]
WizardLM 7B AWQ
-
It messes up with who’s who. Often starts to behave like the user.
-
It writes in third person perspective or Narrative.
-
Sometimes, generates the exact same reply (exactly same to same text) back to back even though new inputs were given.
-
It starts to generate more of a dialogue or screenplay script instead of creating a normal conversation.
Anyone has any solutions for these?
For 1 and 2, apply grammar sampling to force LLM to start all his sentences with:
:
This will “force” the LLM to write dialogue as the specified character. Won’t work 100% of the times, but will become a very rare event.
where did you learn about this?
I’m really struggling to wrap my head around the intuition that goes into imposing grammars on LLM generation.
Very cool trick, thanks.
I use a custom front-end and append the character name / colon to the end of all my prompts to force this, I wonder if grammar sampling would be better. Don’t really have my head around grammar sampling yet.