Has anyone already read this new article on ArXiv? https://arxiv.org/abs/2311.10770
Looks very promising, potential inference acceleration of PyTorch x30, and when implemented on native CUDA x117, and also an estimate of the maximum acceleration x341 times.
As far as I understand, this is achieved by replacing traditional forward propagation layers with so-called fast forward propagation layers.
Is there anyone here with real experience of contributing to the development of PyTorch, llama.cpp or releasing open models, what do you say to this?
" we provide high-level CPU code achieving 78x speedup over the optimized baseline feedforward implementation"
Big if true, we wouldn’t need to buy 3090 cards anymore to get sufficiant memory, just buying more RAM would suffice
Huge, if true.
you might want to read here: https://www.reddit.com/r/MachineLearning/comments/1815a05/r_exponentially_faster_language_modelling/
I say:
- It has a performance hit, but it remains to be seen if going with a much larger model can compensate for that.
- The model needs to be trained from scratch, you cannot finetune an existing model for this apparently…