I’ve been saying this for about a year since seeing the Othello GPT research, but it’s nice to see more minds changing as the research builds up.

  • bionicjoey@lemmy.ca
    link
    fedilink
    English
    arrow-up
    0
    ·
    10 months ago

    You can see clearly that it has no understanding of the wordplay. Though I’ll concede it’s impressive that it got the right answer at all.

    • kromem@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      10 months ago

      The confabulations and in general the surface statistics stuff often gets in the way of the meat and potatoes of critical reasoning in the SotA models.

      A good example of this is trying a variation of common puzzles versus changing tokens to representations and having it repeat adjectives when working through CoT.

      Often as soon as it makes a mistake and has that mistake in context, it just has no way of correcting course. A lot of my current work is related to that and using a devil’s advocate approach to self-correction.

      But in reality, we won’t see a significant jump in things like being able to identify self-ignorance until hardware shifts in the next few years.