Hi. I am not behind the model in any capacity nor those who are asked me to do so, before anyone asks.
I am just a normal LLM enjoyer that wants better 13B models in the near future, because at the moment, they’re being plummeted onto the ground by many Mistral 7B finetunes and since we don’t have any Mistral 13B base model…
The Model in question is this one right here, which seems to be flying under the radar for some reason:
https://huggingface.co/sequelbox/DaringFortitude
TheBloke already did his magic on it, just search his profile on Hugging Face with Ctrl+F.
The reason as to why I am doing this is: I honestly think this is a really, really good (I did some little testing, but my machine is garbage to test any further) and useful Base Model for further finetuning/merging and etc…
There is very little info.
It seems to be instruction finetuned, but what template? ChatML? There is no mention of anything. Posting it this way is pretty bad.
53 GB?
it’s in FP32 rather than FP16
This model is primarily recommended as a superior-to-Llama-2 baseline for additional finetuning,
According to the model, its not really supposed to compete with something like Vicuna. Sounds like they’re trying to be an upgraded foundational model.
What have you found it useful for? The model card is pretty vague.
Really nice, I had a dreamz we need to find a way to iterate over base models so every finetune is closer to sota :D
its model average on the openllm leaderboard is 51.
I really wonder who this TheBlok is. What a legend.
I can’t speak to the quality of sequelbox/DaringFortitude but I can wholeheartedly recommend sequelbox/StellarBright. I have been using StellarBright in some experimental 70b model merges and it’s phenomenal. I imagine 13b merges using DaringFortitude, or finetunes on top of it, would be quite good.