Cradawx@alien.topB to LocalLLaMAEnglish · 2 years agoShareGPT4V - New multi-modal model, improves on LLaVAsharegpt4v.github.ioexternal-linkmessage-square17linkfedilinkarrow-up11arrow-down10
arrow-up11arrow-down1external-linkShareGPT4V - New multi-modal model, improves on LLaVAsharegpt4v.github.ioCradawx@alien.topB to LocalLLaMAEnglish · 2 years agomessage-square17linkfedilink
minus-squarejustletmefuckinggo@alien.topBlinkfedilinkEnglisharrow-up1·2 years agoim new here. but is this true multimodality, or is it the llm communicating with a vision model? and what are those 4 models being benchmark tested here for exactly?
im new here. but is this true multimodality, or is it the llm communicating with a vision model?
and what are those 4 models being benchmark tested here for exactly?