minus-squareOldAd9530@alien.topBtoLocalLLaMA•is there any other tools like vLLM or TensorRT that can be used to speed up LLM inference?linkfedilinkEnglisharrow-up1·1 year agoDo you have any idea why MLC isn’t a more used format? It seems so much faster than GGUF or ExLlama architectures, yet everyone defaults to those linkfedilink
Do you have any idea why MLC isn’t a more used format? It seems so much faster than GGUF or ExLlama architectures, yet everyone defaults to those