Humanity could create an artificial intelligence (AI) agent that is just as smart as humans in as soon as the next three years, a leading scientist has claimed.
Ben Goertzel, a computer scientist and CEO of SingularityNET, made the claim during the closing remarks at the Beneficial AGI Summit 2024 on March 1 in Panama City, Panama. He is known as the "father of AGI" after helping to popularize the term artificial general intelligence (AGI) in the early 2000s.
The best AI systems in deployment today are considered "narrow AI" because they may be more capable than humans in one area, based on training data, but can't outperform humans more generally. These narrow AI systems, which range from machine learning algorithms to large language models (LLMs) like ChatGPT, struggle to reason like humans and understand context.
However, Goertzel noted AI research is entering a period of exponential growth, and the evidence suggests that artificial general intelligence (AGI) where AI becomes just as capable as humans across several areas independent of the original training data is within reach. This hypothetical point in AI development is known as the "singularity."
Goertzel suggested 2029 or 2030 could be the likeliest years when humanity will build the first AGI agent, but that it could happen as early as 2027.
Related: Artificial general intelligence when AI becomes more capable than humans is just moments away, Meta's Mark Zuckerberg declares
If such an agent is designed to have access to and rewrite its own code, it could then very quickly evolve into an artificial super intelligence (ASI) which Goertzel loosely defined as an AI that has the cognitive and computing power of all of human civilization combined.
"No one has created human-level artificial general intelligence yet; nobody has a solid knowledge of when we're going to get there. I mean, there are known unknowns and probably unknown unknowns. On the other hand, to me it seems quite plausible we could get to human-level AGI within, let's say, the next three to eight years," Goertzel said.
He pointed to "three lines of converging evidence" to support his thesis. The first is modeling by computer scientist Ray Kurzweil in the book "The Singularity is Near" (Viking USA, 2005), which has been refined in his forthcoming book "The Singularity is Nearer" (Bodley Head, June 2024). In his book, Kurzweil built predictive models that suggest AGI will be achievable in 2029, largely centering on the exponential nature of technological growth in other fields.
Goertzel also pointed to improvements made to LLMs within a few years, which have "woken up so much of the world to the potential of AI." He clarified LLMs in themselves will not lead to AGI because the way they show knowledge doesn't represent genuine understanding, but that LLMs may be one component in a broad set of interconnected architectures.
The third piece of evidence, Goertzel said, lay in his work building such an infrastructure, which he has called "OpenCog Hyperon," as well as associated software systems and a forthcoming AGI programming language, dubbed "MeTTa," to support it.
OpenCog Hyperon is a form of AI infrastructure that involves stitching together existing and new AI paradigms, including LLMs as one component. The hypothetical endpoint is a large-scale distributed network of AI systems based on different architectures that each help to represent different elements of human cognition from content generation to reasoning.
Such an approach is a model other AI researchers have backed, including Databricks CTO Matei Zaharia in a blog post he co-authored on Feb. 18 on the Berkeley Artificial Intelligence Research (BAIR) website.
Goertzel admitted, however, that he "could be wrong" and that we may need a "quantum computer with a million qubits or something."
"My own view is once you get to human-level AGI, within a few years you could get a radically superhuman AGI unless the AGI threatens to throttle its own development out of its own conservatism," Goertzel added. "I think once an AGI can introspect its own mind, then it can do engineering and science at a human or superhuman level. It should be able to make a smarter AGI, then an even smarter AGI, then an intelligence explosion. That may lead to an increase in the exponential rate beyond even what Ray [Kurzweil] thought."
See more here:
- Has 200 years of science fiction prepared us for AI? - Lewiston Sun Journal [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- The jobs that will disappear by 2040, and the ones that will survive - inews [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- Bill Gates Challenges OpenAI to Train AI to Pass AP Biology Exam - Best Stocks [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- Control over AI uncertain as it becomes more human-like: Expert - Anadolu Agency | English [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- Some Glimpse AGI in ChatGPT. Others Call It a Mirage - WIRED [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- Genie wont go back in the bottle on AI, says security minister - Yahoo Finance UK [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- Fears of artificial intelligence overblown - Independent Australia [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- Researchers at UTSA use artificial intelligence to improve cancer ... - UTSA [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- Elon Musk says he will launch rival to Microsoft-backed ChatGPT - Reuters [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- This is a war and artificial intelligence is more dangerous than a T-80 tank. Unlike a tank its in e... - The US Sun [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- Elon Musk Dishes On AI Wars With Google, ChatGPT And Twitter On Fox News - Forbes [Last Updated On: April 20th, 2023] [Originally Added On: April 20th, 2023]
- How an FEC Deadlock is Hindering the Regulation of AI in Campaigns - Fagen wasanni [Last Updated On: July 19th, 2023] [Originally Added On: July 19th, 2023]
- When Silicon Valley talks about 'AI alignment' here's why they miss ... - Startup Daily [Last Updated On: July 19th, 2023] [Originally Added On: July 19th, 2023]
- MQ-9B - A Growing Track Record of Performance in the Maritime ... - General Atomics Aeronautical Systems [Last Updated On: August 8th, 2023] [Originally Added On: August 8th, 2023]
- 2 Warren Buffett Super Stocks to Buy Hand Over Fist in August - The Motley Fool [Last Updated On: August 8th, 2023] [Originally Added On: August 8th, 2023]
- AI is revolutionizing manual cell counting - Advanced Science News [Last Updated On: August 8th, 2023] [Originally Added On: August 8th, 2023]
- Can AI Help Me Find the Right Running Shoes? - CNET [Last Updated On: August 8th, 2023] [Originally Added On: August 8th, 2023]
- The Role of Artificial Intelligence in the Future of Media - Fagen wasanni [Last Updated On: August 8th, 2023] [Originally Added On: August 8th, 2023]
- AI in Education - EducationNext [Last Updated On: August 8th, 2023] [Originally Added On: August 8th, 2023]
- AI Coin Price: Will Artificial Superintelligence Alliance Have Bullish Impact? - Bankless Times [Last Updated On: July 5th, 2024] [Originally Added On: July 5th, 2024]