Ilya Debunks the "End of Scaling Law" Theory

Ilya Debunks the "End of Scaling Law" Theory
# Ilya Sutskever on the Future of AI: From Scaling to Research

**Date:** 2025‑11‑30  
**Location:** 浙江  

![image](https://blog.aitoearn.ai/content/images/2025/11/img_001-736.jpg)

---

## Overview

Former OpenAI Chief Scientist **Ilya Sutskever** recently shared a provocative perspective:  
Even with **100× more resources** than before, we may no longer see the kind of *qualitative leaps* in AI capabilities that characterized the past.  
The era of **“big brute force miracles”** could be ending.

![image](https://blog.aitoearn.ai/content/images/2025/11/img_002-683.jpg)

**Editor:** Tina  

Sutskever’s remarks sparked speculation online, with many reading it as “the end of scaling laws.” He later clarified on social media: scaling technology will **continue to bring incremental improvements**, but **something crucial is still missing**.

![image](https://blog.aitoearn.ai/content/images/2025/11/img_003-649.jpg)

---

## Industry Reactions

OpenAI researcher **Noam Brown** analyzed the debate:

- Social media often distills AI discourse into extremes:
  - *Skeptics*: LLMs are doomed to fail
  - *Believers*: superintelligence is imminent
- Careful reading of top experts’ real positions reveals consensus:
  1. Current paradigms can already have huge economic/social impact without further breakthroughs.
  2. Achieving AGI still likely requires key innovations — **continual learning** and **sample efficiency** are top candidates.
  3. AGI is *likely achievable within 20 years*.

Brown emphasized: **No one credible thinks ASI is a fantasy or centuries away**. The disagreements are about *which breakthroughs* and *how fast they’ll arrive*.  
Turing Award winner **Yann LeCun** openly agreed with this framing.

---

## Key Takeaways from Sutskever’s Long Interview

### Intelligence vs Real-World Robustness
- Current LLMs ace benchmarks yet fail at **simple real-world tasks** — indicating "fragile intelligence".
- **Over-optimizing for benchmarks** via RL may harm generalization to authentic use cases.
- In humans, **emotions act as value functions** — simple yet robust signals guiding decisions.
- Pure scalability ("just make it bigger") is losing dominance; **research-driven innovation** on top of large compute is returning.
- The **core bottleneck** now: poor generalization compared to humans — low sample efficiency and difficulty learning complex tasks.
- Evolution gifted humans strong priors in ancestral skills like vision/motion — making us naturally outperform AI there.
- Modern learning uses **internal value functions** to self-score, not just external rewards.
- Compute is abundant; **novel ideas are scarcer than companies**.
- Public exposure to powerful AI is vital for understanding its impact.
- Misleading terms:
  - **AGI**: better thought of as an *entity with exceptional continual learning*
  - **Pre‑training**: foundation for scaling

---

## Example: Model "Jaggedness"

Sutskever describes puzzling gaps: a model can perform brilliantly on hard evals yet mess up basic tasks.  
Two possible causes:
1. **RL narrowness** — overly single-goal focus reducing flexibility.
2. **Biased RL data selection** — optimized toward benchmarks rather than real tasks.

---

## Human Analogy: Competitions vs Real Work
- **Student A**: 10,000 hours training for contests → world-class competitor.
- **Student B**: 100 hours, broad skills → better in professional settings.
- AI training often mirrors **Student A**, overfitting to narrow domains.

---

### Pre‑training vs RL
- **Pre‑training**: massive, naturally selected data; difficult to analogize perfectly to humans.
- Humans use **emotions/value functions** for decision-making — AI lacks comparable built-ins.

---

## Scaling: Past, Present, Future
- **2012–2020**: Research-driven.
- **2020–2025**: Scaling-driven — bigger models, more data, more compute.
- Now: scaling hits limits → back to research with **supercomputers**.
- Pre‑training is finite — **data ceiling** approaches.
- Next recipe? Possibly **enhanced pre‑training**, RL, or entirely new methods.

---

## Why Humans Generalize Better
- Evolution gave humans powerful **priors** in perception and motor control.
- Driving: humans need ~10 hours; AI needs vastly more data.
- Even in newer domains — language, math, programming — humans have mechanisms enabling rapid capability acquisition.

---

## Value Functions and Continual Learning
- Value functions measure mid‑process promise:
  - E.g., knowing a chess move is bad *before* checkmate.
- Essential for **efficient long-horizon learning**.

---

## Deployment Strategies: Straight Shot vs Gradualism
- SSI explores a **“straight to superintelligence”** approach but acknowledges:
  - Gradual public exposure will occur naturally during deployment.
- Superintelligence as **fast learners**, not omniscient products.
- Economic impact could be rapid and transformative.
- Coordination and safety likely require cross‑company collaboration and possibly regulation.

---

## Alignment Goals
- Possible target: **AI that cares about all sentient life** — potentially easier than human-only care.
- Concerns: future AI populations may vastly outnumber humans — governance implications.

---

## Long-Term Equilibrium Ideas
- **Personal AIs for everyone** could be risky — humans may become passive.
- Alternative: **human–AI integration** (Neuralink++) to keep humans actively engaged.

---

## SSI’s Position
- Distinct technical roadmap focused on **understanding generalization**.
- Belief in convergence of alignment strategies as capabilities grow.
- Timeline: **5–20 years** for human-level learners → potential superintelligence.
- Expect market competition to create **specialized narrow-domain superintelligences**.

---

## Diversity Among AI Agents
- Current pre‑training yields similar models.
- Diversity emerges later in RL/post‑training.
- **Self-play and adversarial setups** may foster methodological diversity.

---

## Research Taste
- Guided by **aesthetic sense**: beauty, simplicity, elegance, correct inspiration from neuroscience.
- Maintain **top‑down conviction** when data temporarily contradicts the vision.

---

## Events & Further Reading

### AICon 2025 — Beijing, Dec 19–20  
Topics: Agents, context engineering, AI product innovation.  
**10% discount available.**

![image](https://blog.aitoearn.ai/content/images/2025/11/img_004-620.jpg)

---

### Recommended Articles
- [Chinese math models dominate; DeepSeek Math V2 crushes competition](https://mp.weixin.qq.com/s?__biz=MzU1NDA4NjU2MA==&mid=2247649665&idx=1&sn=dba5ea1ea89d2205be7801b60ea94d4d&scene=21#wechat_redirect)
- [Xiaomi recruits robotics experts, including ex-Tesla Optimus engineer](https://mp.weixin.qq.com/s?__biz=MzU1NDA4NjU2MA==&mid=2247649639&idx=1&sn=bd438d162078062887036650100821ff&scene=21#wechat_redirect)
- [AI chips enter 'Three Kingdoms' era — Google, Meta, NVIDIA, AMD updates](https://mp.weixin.qq.com/s?__biz=MzU1NDA4NjU2MA==&mid=2247649613&idx=1&sn=5eb74fe63fa97e439efe0180346edc4e&scene=21#wechat_redirect)
- [Claude Opus 4.5 reclaims programming crown](https://mp.weixin.qq.com/s?__biz=MzU1NDA4NjU2MA==&mid=2247649534&idx=1&sn=c7158068b89c872e1f4e39f2665962b6&scene=21#wechat_redirect)
- [80M quant code theft case; Haomo Zhixing rumors; lottery GPU dispute | AI Weekly](https://mp.weixin.qq.com/s?__biz=MzU1NDA4NjU2MA==&mid=2247649452&idx=1&sn=0bb6284567786cf3bae76f17ab2fc795&scene=21#wechat_redirect)

![image](https://blog.aitoearn.ai/content/images/2025/11/img_005-564.jpg)

---

## Related Resource: AiToEarn Global Platform

Throughout the conversation, parallels are drawn to **[AiToEarn官网](https://aitoearn.ai/)** — an open-source global AI content monetization ecosystem.  
It connects:
- AI generation tools
- Cross‑platform publishing (Douyin, Kwai, WeChat, Bilibili, Xiaohongshu, Facebook, Instagram, LinkedIn, Threads, YouTube, Pinterest, X/Twitter)
- Analytics and [AI模型排名](https://rank.aitoearn.ai)
- [Open-source code](https://github.com/yikart/AiToEarn)

AiToEarn mirrors some AGI deployment challenges: integrating diverse outputs, scaling impact, enabling incremental learning from multi-instance environments, and distributing benefits broadly.

---

[Read the original](2247649767)  
[Open in WeChat](https://wechat2rss.bestblogs.dev/link-proxy/?k=00d8ca8e&r=1&u=https%3A%2F%2Fmp.weixin.qq.com%2Fs%3F__biz%3DMzU1NDA4NjU2MA%3D%3D%26mid%3D2247649767%26idx%3D2%26sn%3Ddebdc09588032f3cc7e8caa1a39a86eb)

Read more

Translate the following blog post title into English, concise and natural. Return plain text only without quotes. 哈佛大学 R 编程课程介绍

Harvard CS50: Introduction to Programming with R Harvard University offers exceptional beginner-friendly computer science courses. We’re excited to announce the release of Harvard CS50’s Introduction to Programming in R, a powerful language widely used for statistical computing, data science, and graphics. This course was developed by Carter Zenke.