So one thing that had really bothered me was that recent Arxiv paper claiming that despite GPT 3 being 175B, and GPT 4 being around 1.7T, somehow 3.5 Turbo was 20b.

This had been on my mind for the past couple of days because it just made no sense to me, so this evening I went to go check out the paper again, and noticed that I could not download the PDF or postscript. Then I saw this update comment on the Arxiv page, added yesterday:

Contains inappropriately sourced conjecture of OpenAI’s ChatGPT parameter count from this http URL, a citation which was omitted. The authors do not have direct knowledge or verification of this information, and relied solely on this article, which may lead to public confusion

That link leads to a Forbes article, from before GPT 4 even released, that claims that ChatGPT in general is 20b parameters.

It seems like the chatbot application was one of the most popular ones, so ChatGPT came out first. ChatGPT is not just smaller (20 billion vs. 175 billion parameters) and therefore faster than GPT-3, but it is also more accurate than GPT-3 when solving conversational tasks—a perfect business case for a lower cost/better quality AI product.

So it would appear that they sourced that knowledge from Forbes, and after everyone got really confused they realized that it might not actually be correct, and the paper got modified.

So, before some wild urban legend forms that GPT 3.5 is 20b, just thought I’d mention that lol.

  • ttkciar@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Perhaps someone heard “10x reduction in footprint” and didn’t realize that meant a reduction in bytes, not a reduction in parameters, and concluded it had a tenth as many parameters?

    • Tight_Range_5690@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      looking at huggingface models, a raw 20b is ~42gb, not a lot of space to fit big model quants. Q4KM of 70b llama fits in that (q2 is 30gb). and the smallest falcon 180b quantization is 74gb

      that would make more sense while still being really impressive. not sure if someone wants to math it out, but what’s the biggest B model that would fit in that on the lowest quants (q2-q3)?

      disclaimer: bees are not everything, maybe they have great dataset/money/lies

    • ambient_temp_xeno@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      So they, as big-shot microsoft scientists, just decided that was good enough to stick it in a table in their paper?