The Paradox of Scaling Law in AI: Does Stronger Reinforcement Learning (RL) Push Us Further from AGI?

Deep News
2025/12/24

In the race toward Artificial General Intelligence (AGI), the current emphasis on Reinforcement Learning (RL) may be leading us astray—ironically, the stronger RL becomes, the farther we might be from achieving true AGI.

On December 24, prominent tech blogger and host of the Dwarkesh Podcast, Dwarkesh Patel, released a thought-provoking video challenging the industry's prevailing optimism around Scaling Law and RL. Patel presents a counterintuitive argument: excessive reliance on RL may not be a shortcut to AGI but rather a clear indicator of its distant horizon.

Patel's core argument centers on the contradiction in current AI development. Leading labs are investing heavily in RL to "pre-bake" specific skills—like Excel manipulation or web browsing—into large models by training them on verifiable outcomes. However, Patel argues that this approach inherently conflicts with the essence of AGI. "If we were truly close to creating a human-like learner," he asserts, "this entire method of training on verifiable outcomes would be doomed to fail."

According to Patel, this "pre-baked" skill paradigm exposes a fundamental flaw in current models. Human value in the workplace stems from our ability to learn and adapt without needing specialized training loops for every minor task. A truly intelligent agent should learn autonomously through experience and feedback, not rehearsed scripts. If AI cannot achieve this, its generality remains limited, and AGI remains out of reach.

Patel contends that the real driver of advanced AI is not endless RL but "Continual Learning"—the ability to learn from experience, much like humans. He predicts that solving continual learning won't be a singular breakthrough but a gradual evolution, akin to improvements in "in-context learning" capabilities. This process may take "5 to 10 years to mature," ruling out the possibility of any single model gaining a runaway advantage by cracking the problem first.

Key Takeaways: 1. **The Pre-Baked Skill Paradox**: Current models' reliance on pre-programmed skills (e.g., Excel or browser use) proves they lack human-like general learning abilities, suggesting AGI is not imminent. 2. **Robotics as an Algorithmic Problem**: If human-like learning existed, robotics would already be solved without needing millions of repetitive training sessions. 3. **The "Diffusion Takes Time" Fallacy**: The argument that slow adoption is due to "technology diffusion" is a cope—truly human-like AI would be rapidly integrated due to its lower risk and training costs. 4. **The Income-Ability Gap**: Global knowledge workers generate trillions in value, while AI model revenues remain orders of magnitude lower, indicating models haven’t reached human-replacement thresholds. 5. **Continual Learning as the Bottleneck**: AGI's true hurdle is continual learning, not RL compute power. Achieving AGI may require another 10–20 years.

Patel's insights challenge the AI community to rethink its trajectory, emphasizing that AGI's arrival hinges not on scaling RL but on unlocking autonomous, experience-driven learning.

免責聲明:投資有風險,本文並非投資建議,以上內容不應被視為任何金融產品的購買或出售要約、建議或邀請,作者或其他用戶的任何相關討論、評論或帖子也不應被視為此類內容。本文僅供一般參考,不考慮您的個人投資目標、財務狀況或需求。TTM對信息的準確性和完整性不承擔任何責任或保證,投資者應自行研究並在投資前尋求專業建議。

熱議股票

  1. 1
     
     
     
     
  2. 2
     
     
     
     
  3. 3
     
     
     
     
  4. 4
     
     
     
     
  5. 5
     
     
     
     
  6. 6
     
     
     
     
  7. 7
     
     
     
     
  8. 8
     
     
     
     
  9. 9
     
     
     
     
  10. 10