• 0 Posts
  • 20 Comments
Joined 1 year ago
cake
Cake day: October 25th, 2023

help-circle












  • ThisGonBHard@alien.topBtoLocalLLaMA30,000 AI models
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    You dont talk about the “usuals”.

    My go to models were for a long time Stable Beluga 2 13B and 70B.

    Then, 13B got replaced by Minstral, 70B by LZLV, and Airoboros Yi 34B came out that worked great for me.

    As a rule: 7B - CPU inferencing on 2-4 cores while using GPU.

    34B and 70B, GPU inferencing, models trade blows despite size diff, as they are different base models. (Llama vs Yi).


  • ThisGonBHard@alien.topBtoLocalLLaMABuild advice.
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    512MB RAM

    When did they thaw you out of ice?!

    Jokes aside, you probably mean 512 GB of RAM. That platform is slow and old, and that is at best DDR3 1333 dual channel, much worse than even bottom barrel DDR4 dual channel.

    A 3090 will not care about it as long as you are doing pure GPU inferencing and not touching the GPU, if not DDR3 and PCI-E2 will kill the performance.


  • ThisGonBHard@alien.topBtoLocalLLaMABuild advice.
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    512MB RAM

    When did they thaw you out of ice?!

    Jokes aside, you probably mean 512 GB of RAM. That platform is slow and old, and that is at best DDR3 1333 dual channel, much worse than even bottom barrel DDR4 dual channel.

    A 3090 will not care about it as long as you are doing pure GPU inferencing and not touching the GPU, if not DDR3 and PCI-E2 will kill the performance.


  • Both are bad choices:

    IDK about the 3090 and the PSU, that thing can spike HARD, in the 1.6KW range, and if you PSU is a lower end one, it will kill it. I head a lot of people on 120V countries complain because it actually cause their lights to flicker.

    The 4060 Ti is limited by PCI-E 8X at 2.0 speeds, but that is a XX50 chip masquerading as a 60, so it sips power.

    3090 would be better, but you dont have the PC for it. Get a cheap used AMD B550 board for PCI-E 4.0, 64 GB of RAM and whatever CPU is within your remaining budget, all the way down from R5 3600 to R7 5800X3D, AM4 is really well segmented for price, even new. You will get better gaming performance too, and can run a lot of stuff in CPU with 64 GB of RAM.



  • OpenAI was not gonna release shit for consumers either way, doomers are to scared of shadows to do it, and GPT3.5 was too advanced to make public by Illya.

    Because Microsoft has GPT4 too, I am pretty sure they are just gonna continue working on what they were before as if nothing happened, under Microsoft, just now are not shackled and can go full steam ahead.

    The doomers lost, because now the acceleration side is free and unshackled. At best, they bought 4 months, but progress might come 3x faster after those.