https://huggingface.co/deepnight-research

I’m not affiliated with this group at all, I was just randomly looking for any new big merges and found these.

100B model: https://huggingface.co/deepnight-research/saily_100B

220B model: https://huggingface.co/deepnight-research/Saily_220B

600B model: https://huggingface.co/deepnight-research/ai1

They have some big claims about the capabilities of their models, but the two best ones are unavailable to download. Maybe we can help convince them to release them publicly?

    • iCantHack@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      I wonder if there’s any real demand for even 48GB 4090s enough to incentives somebody to do it. I bet the hardware/electronics part of it is trivial, tho.

        • BangkokPadang@alien.topB
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          Honestly, a 4bit quantized version of the 220B model should run on a 192GB M2 Studio, assuming these models could even work with a current transformer/loader.

    • LocoMod@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      We need some hero to develop an app that downloads more GPU memory like those apps back in the 90’s. /s