minus-squarervitor@alien.topBtoLocalLLaMA•When training an LLM how do you decide to use a 7b, 30b, 120b, etc model (assuming you can run them all)?linkfedilinkEnglisharrow-up1·1 year agofor training sometimes is better to pick a small model to do some tests and get faster feedback, then you can train in a larger model if you want to, and see how it goes. linkfedilink
for training sometimes is better to pick a small model to do some tests and get faster feedback, then you can train in a larger model if you want to, and see how it goes.