Self-Improving Robots and the Importance of Data

·Sergey Levine··

There is a popular idea that, with 10,000 hours of experience, a person could achieve mastery at whatever task they set themselves to. A bit of back-of-the-envelope calculation tells us that so far, our machine learning models are quite a bit more data-hungry. For example, GPT-3 is trained on over 10 terabytes of data, which corresponds to some billions of pages of text that, at average human reading speed, would take hundreds of millions of hours to read (about a hundred lifetimes). State-of-th...

Read full article →

Related Articles

Accelerating Gemma 4: faster inference with multi-token prediction drafters
amrrs · Hacker News · 3d ago
ProgramBench: Can language models rebuild programs from scratch?
jonbaer · Hacker News · 1d ago
ZAYA1-8B matches DeepSeek-R1 on math with less than 1B active parameters
steveharing1 · Hacker News · 1d ago
OpenAI’s o1 correctly diagnosed 67% of ER patients vs. 50-55% by triage doctors
donsupreme · Hacker News · 6d ago
A couple million lines of Haskell: Production engineering at Mercury
unignorant · Hacker News · 6d ago