Video thumbnail for Are We at the End of Ai Progress— With Gary Marcus

AI Skeptic: Is the AI Revolution Over? Gary Marcus Interview

Summary

Quick Abstract

Is the AI scaling era truly over? Leading AI skeptic Gary Marcus joins us to dissect the recent admissions from industry giants about diminishing returns in large language model (LLM) scaling. He explains why simply throwing more compute and data at AI isn't yielding the exponential improvements once promised and what the implications are for the future of AI, the market, and valuations.

Quick Takeaways:

  • Scaling laws, predicting linear improvements with more data, are failing, leading to disappointment and backpedaling.

  • Efforts to build a "GPT5-level" model have fallen short, signaling a significant hurdle in AI advancement.

  • "Reasoning" in current models is primarily mimicry, not genuine abstraction, resulting in persistent errors and hallucinations.

  • Open-source AI raises safety concerns due to potential misuse by malicious actors, while closed AI companies could monetize user data.

Marcus argues that the current path of giant black-box LLMs is a wrong hypothesis and suggests exploring neuro-symbolic AI, combining neural networks with classical AI elements, is the future path. He also forecasts a potential financial crash due to overvalued AI companies. Is this the end of the AI boom, or just a critical turning point?

Introduction

The AI landscape is shifting. As research institutions grapple with the limitations of scaling, questions arise about the future of the industry. Leading AI skeptic, author, and founder Gary Marcus joined the Big Technology Podcast to discuss these challenges. This article summarizes his insights and predictions.

The Scaling Laws and Diminishing Returns

The Original Promise

The initial premise behind large language models (LLMs) was that increasing compute and data would lead to predictable, often exponential, improvements in performance. Papers from OpenAI and others seemed to confirm this relationship, leading to significant investment in the field.

The Reality Check

However, recent results have disappointed many. Researchers are starting to acknowledge that scaling is no longer delivering the same returns. Industry figures like Mustafa Suleyman, Thomas Kurian, and Yann LeCun have alluded to these diminishing returns.

Marcus's Perspective

Gary Marcus, who predicted this slowdown in a 2022 paper, feels vindicated. He notes that the "mathematical laws" governing scaling were merely generalizations that held true for a limited time, not fundamental laws of nature.

GPT-4 and the Quest for GPT-5

GPT-4's Success

GPT-4 represented a significant leap forward compared to its predecessors. The improvements were easily noticeable. This success fueled expectations for an even more advanced GPT-5.

Project Orion's Failure

OpenAI's attempt to build GPT-5 (Project Orion) fell short of expectations and was ultimately released as GPT 4.5. The model did not meet the performance targets dictated by the established scaling laws.

Redefining Scaling

Marcus argues that the industry is now redefining scaling to mask the limitations. Adding more data may yield some improvement, but the exponential growth predicted by the original scaling laws is no longer observed.

The Limits of Current Approaches

Billion-Dollar Experiments

Companies invested heavily in scaling, expecting commensurate performance gains. However, the billion-dollar experiments failed to deliver the anticipated results.

Diminishing Returns Explained

Diminishing returns mean that each additional unit of data or compute yields a smaller performance increase. The gains are not as significant or generalizable as they once were.

Test Time Compute and Reasoning

The Idea of Reasoning

To overcome the scaling limitations, researchers are exploring "test time compute" or "reasoning." This involves having the model check its progress and take iterative steps toward finding an answer.

Limited Success

While test time compute can improve performance on certain problems, particularly in math and programming, it's not a universal solution. The gains are often limited to closed domains where synthetic data can be generated and verified.

Mimicry vs. True Reasoning

Marcus believes these models are primarily copying patterns of human reasoning, rather than engaging in true, abstract reasoning. They still make frequent and obvious mistakes.

Hallucinations

Intriguingly, some newer models, such as 03, exhibit more hallucinations than their predecessors. This highlights the limited understanding of these "black box" systems.

The Black Box Problem and Interpretability

Lack of Understanding

Current AI models are often described as "black boxes" because their internal workings are opaque. While inputs and outputs are known, the process by which the system arrives at a solution is unclear. This lack of understanding makes it difficult to diagnose and fix problems like hallucinations.

The Need for Interpretability

Interpretability, the ability to understand how an AI system works, is crucial. Unlike interpretable AI systems like GPS navigation, LLMs lack transparency.

A Call for New Approaches

Marcus advocates for a fundamentally different approach to AI, arguing that further progress within the "black box" paradigm is unlikely.

Are We There Yet?

Limited Progress Since GPT-4

Marcus suggests that significant progress in AI has stalled since the release of GPT-4. While incremental improvements exist, they haven't delivered the transformative breakthroughs many anticipated.

Improvement but not a Quantum Leap

Although there are improvements on benchmarks, issues with data contamination and limited generalization remain. The progress is not as groundbreaking as expected. Hallucinations and reasoning errors persist.

Real-world Limitations

These systems still make subtle errors that often go unnoticed, and struggle with tasks that require novel coding or debugging.

Financial Implications and Predictions

Diminishing Returns and Investment

The diminishing returns from scaling raise questions about the future of investment in AI. If larger GPU data centers don't yield substantial improvements, will companies continue to pour billions of dollars into them?

Financial Collapse?

Marcus forecasts a potential financial collapse in the AI sector, arguing that current valuations are unsustainable. He doesn't foresee OpenAI being worth hundreds of billions of dollars.

NVIDIA's Risk

NVIDIA, a major beneficiary of the AI boom, faces a significant risk. If the next generation of GPUs fails to deliver substantial performance gains, demand could plummet, leading to a crash in the company's stock price.

Price Wars and Lack of Moats

The increasing number of companies pursuing the same approach (building bigger LLMs) is likely to lead to price wars and a lack of sustainable competitive advantages (moats).

Privacy and Surveillance Concerns

Data Monetization

OpenAI's vast collection of user data could lead the company to monetize this information through surveillance and hyper-targeted advertising.

Echoes of Facebook

The potential for AI companies to exploit user data raises concerns similar to those surrounding Facebook's data practices.

Potential for Abuse

The sensitive information users share with AI systems could be vulnerable to extortion or other forms of abuse.

The Path Forward: Neuro-Symbolic AI

Limitations of Current AI

Current AI systems excel at fast, statistical processing (System One thinking) but struggle with abstract reasoning and critical thinking (System Two thinking).

Neuro-Symbolic AI

Marcus advocates for a neuro-symbolic approach that combines the strengths of neural networks (learning from data) with classical AI (explicit knowledge and formal reasoning). He suggests the best AI approach will combine both of these systems together.

A Potential Solution

This approach, exemplified by AlphaFold, could lead to more robust and reliable AI systems.

Was this summary helpful?