multiverse_fan@alien.topB to LocalLLaMAEnglish · 1 year agoAnyone have a 1B or 3B model that is mostly coherent?message-squaremessage-square16fedilinkarrow-up11arrow-down10file-text
arrow-up11arrow-down1message-squareAnyone have a 1B or 3B model that is mostly coherent?multiverse_fan@alien.topB to LocalLLaMAEnglish · 1 year agomessage-square16fedilinkfile-text
I’ve tried a few of these models but it was some months ago. Have y’all seen any that can hold a conversation yet?
minus-squareNonetendo65@alien.topBlinkfedilinkEnglisharrow-up1·1 year agoI’ve found Orca-Mini to be quite helpful for simple generation tasks < 200 tokens, given it’s only 2.0GB it’s quite powerful and easy to deploy on consumer hardware. Orca is the famous dataset that the wonderful Mistral 7B was trained on :)
I’ve found Orca-Mini to be quite helpful for simple generation tasks < 200 tokens, given it’s only 2.0GB it’s quite powerful and easy to deploy on consumer hardware. Orca is the famous dataset that the wonderful Mistral 7B was trained on :)