But they don’t ask why it is not a much larger, newer model. My answer is that OpenAI has tried and does not yet have the ability to build anything much bigger and more capable relative to GPT-4, despite two years and untold billions of investment.
I’m not sure this is true. Two key points are made in the Sam Hammond tweet:
OpenAI has made better models internally, they just haven’t been released.
There wasn’t a big increase in compute for GPT5 because this compute isn’t yet available. Big compute projects take time. Maybe we’re in a bit of a compute slump now, but it isn’t clear this will always be the case.
I’m not sure this is true. Two key points are made in the Sam Hammond tweet:
OpenAI has made better models internally, they just haven’t been released.
There wasn’t a big increase in compute for GPT5 because this compute isn’t yet available. Big compute projects take time. Maybe we’re in a bit of a compute slump now, but it isn’t clear this will always be the case.