• LienniTa@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    yeah people praising 7b and 13 b models here and there, but…they just hallucinate! Then 120b goliath, no matter how terrible its initial idea was, is just really good in normal conversations. Im trying to love giga praised open hermes 2.5 and other mistral finetunes, but they are just better next-token-predictors, unlike larger models which are actually able to reason.