Found this in a children’s book of riddles:

Six brothers were spending their time together.

The first brother was reading a book.
The second brother was playing chess.
The third brother was solving a crossword.
The fourth brother was watering the lawn.
The fifth brother was drawing a picture.

Question: what was the sixth brother doing?

I cant get ChatGPT to answer correctly with the usual tricks, even after hinting to consider one and two-person activities and emphasizing the word “together”.

After a bunch of CoT turns we arrive to a conclusion that this is an open ended question and not a riddle :)

After trying 3 times with fresh promots, I got a correct response once, but when prompted to provide supporting reasoning the model backtracked and started apologizing.

Cant test gpt 4 r/n…

  • Hugi_R@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Open-ended question are the best for evaluating LLM, because they require common sense/world knowledge/doxa/human like behavior.

    Saying “I don’t know” is just a cop out response. At least it should say something like “It could be X but …”, be a little creative.

    Another (less?) open-ended question with the same premise would be “Where are they?” and I expect the answer to be “In a garden”.

    GPT-4 Turbo (with custom instruction) answer very well https://chat.openai.com/share/c305568e-f89e-4e71-bb97-79f7710c441a

    • Be-Kind_Always-Learn@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Another (less?) open-ended question with the same premise would be “Where are they?” and I expect the answer to be “In a garden”.

      Perhaps there’s a language barrier here, but none of those activities hint to a garden? In my locale, a garden is a small patch used to grow veggies, herbs, and/or flowers. So I would answer this with “their back yard.”

      This is a much better riddle for children IMO, because it’s barely open-ended at all. The original has almost infinite answers without any leaps or tricks, but yours has a very limited domain: a yard/garden. Though if someone were extra clever, the problem space does open back to nearly infinity (if brother 4 is playing a video game).

      Open-ended question are the best for evaluating LLM, because they require common sense/world knowledge/doxa/human like behavior.

      For personal testing, that’s certainly a valid opinion! But it’s not very productive from an objective standpoint because it can’t be graded and tests a “gotcha” path of thinking, when we’re still focusing on fundamentals like uniform context attention, consistency over time, etc.

    • laca_komputilulo@alien.topOPB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Thank you, bud Mind trying the same prompt on the cheapo 3.5 model? I suspect it will hit it on the nail with your custom instructions, given that it was hit and miss for me with my weaker prompting judjitsu

      • Hugi_R@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        3.5 never suspect the 6th playing chess

        https://chat.openai.com/share/b7e6b24d-44db-4abf-9a81-5325f836bca5 (the === are artifacts of the custom system prompt, 3.5 sucks at following it)

        I asked it for candidate activity, and mostly offered different ones. It’s weird, I would expect a LLM to list activities that were already mentioned in the conversation. Maybe the repetition penalty is set too high?