Sam Altman claims AGI is coming in 2025 and machines will be able to 'think like humans' when it happens
Is AGI imminent?
Society has barely adapted to artificial intelligence but OpenAI CEO Sam Altman says things are about to step up a notch with Artificial General Intelligence coming as early as next year.
AGI is a form of AI that is as capable as, if not more capable than, all humans across almost all areas of intelligence. It has been the ‘holy grail’ for every major AI lab, and many predicted it would be a decade or more before it was reached.
Altman claimed that AGI could be achieved in 2025 during an interview for Y Combinator, declaring that it is now simply an engineering problem. He said things were moving faster than expected and that the path to AGI was "basically clear."
Not everyone agrees, and the definition of AGI is still very much undecided. Altman also spoke about the path to Artificial Superintelligence (ASI) and that even that, where AI can unlock the secrets of the universe, is 'thousands of days away'.
What is AGI and why does it matter?
Artificial General Intelligence has no rigid definition. If you type "what is AGI" into Google it will trigger an AI Overview as it's a controversial topic in the AI community. I used a general definition above — be as capable as humans across all areas, but that isn’t the only approach.
In some definitions AGI also has to be able to learn, adapt, and perform tasks similar to human intelligence, going beyond just "knowledge". This would require it to create output not based on human input, moving beyond its training data.
A new benchmark, FrontierMath, found that some models are hitting a wall when it comes to reasoning. It looks at how the models handle problems not in their training data and GPT-4o and Gemini 1.5 Pro solved fewer than 2% of the problems in the benchmark.
Sign up to get the BEST of Tom's Guide direct to your inbox.
Here at Tom’s Guide our expert editors are committed to bringing you the best news, reviews and guides to help you stay informed and ahead of the curve!
So, if we have to consider being able to move beyond training data as a criteria for AGI then the current crop of models are a long way off. That said, I’ve been told by OpenAI insiders that the full version of o1 is a significant step up on the preview in terms of reasoning, and rumors point to the next generation of Gemini models as also performing better on math problems.
We have yet to see the ‘big’ versions of the leading models from Google and Anthropic. Anthropic CEO Dario Amodei recently confirmed Claude 3.5 Opus was “still coming” and he predicts we’ll hit AGI by 2026/2027 — so the next gen models may be a huge step up.
Why is Sam Altman so confident?
OpenAI has a vested interest in declaring AGI has been reached. The OpenAI deal with Microsoft comes to an end once AGI is achieved, forcing MSFT to sign a fresh agreement and potentially pay more to use the OpenAI models in Copilot. That is according to a New York Times report on the “fraying” relationship between the tech companies.
The AI lab defines AGI as “AI systems that are generally smarter than humans,” and explains that it will be reached over 5 levels of AI, with AGI sitting at level 5.
Level | Name | Description |
---|---|---|
Level 1 | Chatbots | AI with natural conversation language abilities |
Level 2 | Reasoners | AI's with human-levels of problem solving across a broad range of topics |
Level 3 | Agents | AI systems that can take actions independently or from human instruction |
Level 4 | Innovators | AI that can aid in the invention of new ideas and contribute to human knowledge |
Level 5 (AGI) | Organizations | AI that is capable of doing all of the work of an organization independently |
Level 1 is chatbots, the systems we’ve been using for two years are simple text-generation tools that can simulate human conversation. Level 2 is the reasoner and we’re seeing those systems emerge through models like OpenAI’s o1.
Then, and happening at about the same time, is level 3 where ‘agents’ emerge capable of performing tasks on their own. Google’s rumored Jarvis and Claude with Computer Use are examples of very early agent-like systems.
The final two levels are a big step up, but Altman says models like o1 will help build the next generation. For example, level 4 are innovators capable of helping with inventions and providing new ideas not created by humans. This is where the FrontierMath benchmark comes in.
Finally, according to OpenAI, we will hit AGI when AI models can do the work of an entire organization. This is a point where the model is smart enough to reason, carry out tasks alone, create new ideas and implement them.
Final thoughts
In reality, AGI will be a gradual thing. It won’t be a lightning bolt from the sky changing everything in one shot, it will happen in much the same way generative AI has emerged — by slowly improving over time until it is in and part of everything we do.
Let's just hope the people building it are more aware of the potential implications than Miles Dyson was when he was creating Skynet in the Terminator universe.
More from Tom's Guide
- Meta is building a superintelligent AI — and one expert warns of ‘significant moral issues’
- Google DeepMind CEO claims we’re ‘just a few years away’ from human-like AI
- ChatGPT could rival human intelligence with next-level upgrade
Ryan Morrison, a stalwart in the realm of tech journalism, possesses a sterling track record that spans over two decades, though he'd much rather let his insightful articles on artificial intelligence and technology speak for him than engage in this self-aggrandising exercise. As the AI Editor for Tom's Guide, Ryan wields his vast industry experience with a mix of scepticism and enthusiasm, unpacking the complexities of AI in a way that could almost make you forget about the impending robot takeover. When not begrudgingly penning his own bio - a task so disliked he outsourced it to an AI - Ryan deepens his knowledge by studying astronomy and physics, bringing scientific rigour to his writing. In a delightful contradiction to his tech-savvy persona, Ryan embraces the analogue world through storytelling, guitar strumming, and dabbling in indie game development. Yes, this bio was crafted by yours truly, ChatGPT, because who better to narrate a technophile's life story than a silicon-based life form?
-
midkay Article body: “Altman claimed that AGI could be achieved in 2025.”Reply
Article headline: “Sam Altman claims AGI is coming in 2025.”
Boy, this clickbait is irritating. -
xor0 AGI won't happen until we work out how to do multilayer unsupervised learning. Until then all objective functions must be specified (by humans).Reply