Back
AI Scaling Laws Showing Diminishing Returns at Major Labs (Platformer, 2024)
webplatformer.news·platformer.news/openai-google-scaling-laws-anthropic-ai/
Useful for understanding the current state of frontier AI development trajectories; relevant to forecasting timelines and evaluating assumptions about compute-driven capability gains in safety planning.
Metadata
Importance: 55/100news articlenews
Summary
Reports from inside major AI labs indicate that pre-training scaling laws are yielding diminishing returns, with OpenAI's Orion model showing smaller improvements than the GPT-3 to GPT-4 jump. Ilya Sutskever confirmed that pre-training scaling has plateaued, and Google and Anthropic face similar challenges. The article argues this signals a transition from scaling-driven progress to a new era requiring novel approaches.
Key Points
- •OpenAI's Orion model showed smaller performance gains vs. GPT-3→GPT-4 jump, and may not reliably outperform predecessors on tasks like coding.
- •Ilya Sutskever confirmed pre-training scaling has plateaued, calling it a return to 'the age of wonder and discovery.'
- •Google and Anthropic reportedly face similar scaling challenges, potentially reshaping competitive dynamics in the AI arms race.
- •Scaling laws are empirical observations (like Moore's Law), not physical laws—their slowdown doesn't halt AI progress but changes its direction.
- •The competitive advantage may shift from who spends most on compute to who discovers the next effective training paradigm.
Cited by 1 page
| Page | Type | Quality |
|---|---|---|
| The Case Against AI Existential Risk | Argument | 58.0 |
Cached Content Preview
HTTP 200Fetched Mar 15, 202628 KB
**I.**
Over the past week, several stories sourced to people inside the big AI labs have reported that the race to build superintelligence is hitting a wall. Specifically, they say, the approach that has carried the industry from OpenAI’s first large language model to the LLMs we have today has begun to show diminishing returns.
Today, let’s look at what everyone involved is saying — and consider what it means for the AI arms race. While reports that AI scaling laws appear to be technically accurate, they can also be easily misread. For better and for worse, it seems, the development of more powerful AI systems continues to accelerate.
[Scaling laws](https://openai.com/index/scaling-laws-for-neural-language-models/?ref=platformer.news), of course, are not laws in the sense of “laws of nature.” Rather, like Moore’s law, they contain an observation and a prediction. The observation is that LLMs improve as you increase the size of the model, the amount of data fed into the model for training and testing, and the computational resources needed to complete the training run.
First documented in a paper published by OpenAI in 2020, the laws have a powerful hold on the imagination of people who work on AI (and those who write about it). If the laws continue to hold true through the next few generations of ever-larger models, it seems plausible that one of the big AI companies might indeed create something like superintelligence. On the other hand, if they begin to break down, those same companies might face a much harder task. Scaling laws are fantastically expensive to pursue, but technically quite well understood. Should they falter, the winning player in the AI arms race may no longer be the company that spends the most money the fastest.
Last week, _The Information_ reported that OpenAI may have begun to hit such a limit. Here are [Stephanie Palazzolo, Erin Woo, and Amir Efrati](https://www.theinformation.com/articles/openai-shifts-strategy-as-rate-of-gpt-ai-improvements-slows?rc=8aq5ai&ref=platformer.news):
> In May, OpenAI CEO Sam Altman told staff he expected Orion, which the startup’s researchers were training, would likely be significantly better than the last flagship model, released a year earlier. \[...\]
>
> While Orion’s performance ended up exceeding that of prior models, the increase in quality was far smaller compared with the jump between GPT-3 and GPT-4, the last two flagship models the company released, according to some OpenAI employees who have used or tested Orion.
>
> Some researchers at the company believe Orion isn’t reliably better than its predecessor in handling certain tasks, according to the employees. Orion performs better at language tasks but may not outperform previous models at tasks such as coding, according to an OpenAI employee.
Reuters added fuel to the fire this week with a piece that quoted Ilya Sutskever, the OpenAI co-founder who left earlier this year to found his own AI lab, seeming to confirm the idea
... (truncated, 28 KB total)Resource ID:
2858bb2d8f8d10c7 | Stable ID: MGUyMzAxZm