AGI: Yay or Nay?

There’s a lot of hype going around social media surrounding OpenAI’s final announcement of its “12 Days of OpenAI” extravaganza. Following the release of o1, codenamed strawberry, earlier this year, they have now announced o3. (o2 was not immediately available for comment)

Here’s arguably the most important tweet regarding o3, from an AI researcher who designed one of the best known dataset/tasks meant to measure true reasoning ability:

This is a big milestone within AI research, but it has also led to some hyperbole:

As well as some ridiculous nonsense:

But there are those who have pushed back on the hype:

We still don’t know exactly how o1/o3 work, but most assume it’s a form of search combined with GPT-4.5 LLM(s). Inference costs to solve the most challenging problems are quite high ($1,000+).

Ultimately, my opinion is that while this is a big step forward for AI, we still have a long way to go to get to true AGI. I’ve believed for years now that more advanced inference algorithms would be necessary to get closer to human-level intelligence, and this appears to be one successful way of using more test-time compute to solve tougher problems. Progress will continue, but my guess is that we are still a decade or so away from genuine human-level intelligence.

Leave a Reply