Which examples do you think of when you say this? (Not necessarily disagreeing, I’m just interested in the different interpretations of ‘LLMs are poor at general reasoning’ ).
I also think that LLM reasoning can be significantly boosted with scaffolding—i.e: most hard reasoning problems can be split up into a a handful of easier reasoning problems; this can be done recursively until your LLM can solve a subproblem, then build back up the full solution. So whilst scale might not get us to a level of general reasoning that qualifies as AGI, perhaps GPT-5 (or 6) plus scaffolding can.
Appreciate the concreteness in the predictions!
Which examples do you think of when you say this? (Not necessarily disagreeing, I’m just interested in the different interpretations of ‘LLMs are poor at general reasoning’ ).
I also think that LLM reasoning can be significantly boosted with scaffolding—i.e: most hard reasoning problems can be split up into a a handful of easier reasoning problems; this can be done recursively until your LLM can solve a subproblem, then build back up the full solution. So whilst scale might not get us to a level of general reasoning that qualifies as AGI, perhaps GPT-5 (or 6) plus scaffolding can.