• 0 Posts
  • 7 Comments
Joined 1 year ago
cake
Cake day: November 9th, 2023

help-circle

  • I’ve switched to ExLlamav2 + EXL2 as that lets me run 120B models entirely in 48 GB VRAM (2x 3090 GPUs) at 20 T/s. And even if it’s just 3-bit

    Wow, can I ask how you got this? Because I used 2x3090 the same (x16/x16, no nvlink,fresh new text generation webui,120b goliath-rpcal 3bpw exl2) and only got 10t/s

    I got 6~8t/s when using 3x3090 load 4.5bpw, and another person also got the same speed, it seems your speed is almost twice as fast as mine, which make my brain explosion



  • NEW FEATURE: Context Shifting (A.K.A. EvenSmarterContext) - This feature utilizes KV cache shifting to automatically remove old tokens from context and add new ones without requiring any reprocessing. So long as you use no memory/fixed memory and don’t use world info, you should be able to avoid almost all reprocessing between consecutive generations even at max context. This does not consume any additional context space, making it superior to SmartContext.

    do you tried this koblodcpp new feature?


  • Newbie question, but is there a way to have 4*A100 40G cards run as one, with 160G VRAM in total?

    Yes, that will work but lose some performance

    edit) If this is possible, can I run 8*3090 24G cards as one also?

    Yes and no, yes you can do this but no unless you need run a 176GB model, more gpu with a model only lose performance but not increase it

    for example, if I run a 13b gptq 4bit model on 1*3090 it get 45t/s, if i run it on 2*3090 it will slow down to 30t/s but 3*3090 is same

    and also, I dont have any idea why people always saying about nvlink, 2*pciex16 3090 give same speed compare with 2*pciex1, not sure if there will change for higher number but it just not help at all for 2*3090

    so basically the real matter is what model you want to run, not what and how many GPU you can use, you need run your model with minimum number gpu to get best performance, if 3090 have 36Gvram but not only 24Gvram, 36G*2 will way more faster than 24G * 3 even the total Vram is same