Nkingsy@alien.topBtoLocalLLaMA•Is LLaMA-1-65B or LLaMA-2-70B more creative at storytelling ?English
1·
1 year agoI think llama 1 had more interesting training data, but it can’t hold a plot too well
I think llama 1 had more interesting training data, but it can’t hold a plot too well
Trained on a larger # of tokens. All the llama models are under trained it appears, especially the 70b
Or the more undertrained it is, the more fat can be trimmed