Close Menu
World Forbes – Business, Tech, AI & Global Insights
  • Home
  • AI
  • Billionaires
  • Business
  • Cybersecurity
  • Education
    • Innovation
  • Money
  • Small Business
  • Sports
  • Trump
What's Hot

Meesho Becomes India’s Best IPO This Year, Mints New Billionaire

December 17, 2025

Good news: How AP journalists covered some of 2025’s happiest stories

December 17, 2025

Most Americans think cash gifts are acceptable, poll shows

December 17, 2025
Facebook X (Twitter) Instagram
Trending
  • Meesho Becomes India’s Best IPO This Year, Mints New Billionaire
  • Good news: How AP journalists covered some of 2025’s happiest stories
  • Most Americans think cash gifts are acceptable, poll shows
  • School attendance plummeted during Texas measles outbreak
  • Bartenders embrace maximalism with vibrant cocktails
  • Rome opens long-awaited Colosseum subway station, with displays of unearthed artifacts
  • Creating a simple garden sanctuary can bring year-round relaxation
  • Thousands bid farewell to last 2 pandas in Japan before their return to China
World Forbes – Business, Tech, AI & Global InsightsWorld Forbes – Business, Tech, AI & Global Insights
Wednesday, December 17
  • Home
  • AI
  • Billionaires
  • Business
  • Cybersecurity
  • Education
    • Innovation
  • Money
  • Small Business
  • Sports
  • Trump
World Forbes – Business, Tech, AI & Global Insights
Home » Researchers say they’ve discovered a new method of ‘scaling up’ AI, but there’s reason to be skeptical
AI

Researchers say they’ve discovered a new method of ‘scaling up’ AI, but there’s reason to be skeptical

By adminMarch 19, 2025No Comments3 Mins Read
Facebook Twitter Pinterest LinkedIn Tumblr WhatsApp Telegram Email
Share
Facebook Twitter LinkedIn Pinterest Email
Post Views: 115


Have researchers discovered a new AI “scaling law”? That’s what some buzz on social media suggests — but experts are skeptical.

AI scaling laws, a bit of an informal concept, describe how the performance of AI models improves as the size of the datasets and computing resources used to train them increases. Until roughly a year ago, scaling up “pre-training” — training ever-larger models on ever-larger datasets — was the dominant law by far, at least in the sense that most frontier AI labs embraced it.

Pre-training hasn’t gone away, but two additional scaling laws, post-training scaling and test-time scaling, have emerged to complement it. Post-training scaling is essentially tuning a model’s behavior, while test-time scaling entails applying more computing to inference — i.e. running models — to drive a form of “reasoning” (see: models like R1).

Google and UC Berkeley researchers recently proposed in a paper what some commentators online have described as a fourth law: “inference-time search.”

Inference-time search has a model generate many possible answers to a query in parallel and then select the “best” of the bunch. The researchers claim it can boost the performance of a year-old model, like Google’s Gemini 1.5 Pro, to a level that surpasses OpenAI’s o1-preview “reasoning” model on science and math benchmarks.

Our paper focuses on this search axis and its scaling trends. For example, by just randomly sampling 200 responses and self-verifying, Gemini 1.5 (an ancient early 2024 model!) beats o1-Preview and approaches o1. This is without finetuning, RL, or ground-truth verifiers. pic.twitter.com/hB5fO7ifNh

— Eric Zhao (@ericzhao28) March 17, 2025

“[B]y just randomly sampling 200 responses and self-verifying, Gemini 1.5 — an ancient early 2024 model — beats o1-preview and approaches o1,” Eric Zhao, a Google doctorate fellow and one of the paper’s co-authors, wrote in a series of posts on X. “The magic is that self-verification naturally becomes easier at scale! You’d expect that picking out a correct solution becomes harder the larger your pool of solutions is, but the opposite is the case!”

Several experts say that the results aren’t surprising, however, and that inference-time search may not be useful in many scenarios.

Matthew Guzdial, an AI researcher and assistant professor at the University of Alberta, told TechCrunch that the approach works best when there’s a good “evaluation function” — in other words, when the best answer to a question can be easily ascertained. But most queries aren’t that cut-and-dry.

“[I]f we can’t write code to define what we want, we can’t use [inference-time] search,” he said. “For something like general language interaction, we can’t do this […] It’s generally not a great approach to actually solving most problems.”

Mike Cook, a research fellow at King’s College London specializing in AI, agreed with Guzdial’s assessment, adding that it highlights the gap between “reasoning” in the AI sense of the word and our own thinking processes.

“[Inference-time search] doesn’t ‘elevate the reasoning process’ of the model,” Cook said. “[I]t’s just a way of us working around the limitations of a technology prone to making very confidently supported mistakes […] Intuitively if your model makes a mistake 5% of the time, then checking 200 attempts at the same problem should make those mistakes easier to spot.”

That inference-time search may have limitations is sure to be unwelcome news to an AI industry looking to scale up model “reasoning” compute-efficiently. As the co-authors of the paper note, reasoning models today can rack up thousands of dollars of computing on a single math problem.

It seems the search for new scaling techniques will continue.





Source link

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
admin
  • Website

Related Posts

After Klarna, Zoom’s CEO also uses an AI avatar on quarterly call

May 23, 2025

Anthropic CEO claims AI models hallucinate less than humans

May 22, 2025

Anthropic’s latest flagship AI sure seems to love using the ‘cyclone’ emoji

May 22, 2025

A safety institute advised against releasing an early version of Anthropic’s Claude Opus 4 AI model

May 22, 2025

Anthropic’s new AI model turns to blackmail when engineers try to take it offline

May 22, 2025

Meta adds another 650 MW of solar power to its AI push

May 22, 2025
Add A Comment
Leave A Reply

Don't Miss
Billionaires

Meesho Becomes India’s Best IPO This Year, Mints New Billionaire

December 17, 2025

Vidit Aatrey, CEO of Meesho, speaks during the company’s listing ceremony at the National Stock…

MacKenzie Scott’s Latest Gifts Make Her America’s Third Most Generous Philanthropist

December 11, 2025

Indonesian Billionaires Cash In On Gold Surge

December 10, 2025

Kalshi’s Cofounder Is Now World’s Youngest Self-Made Woman Billionaire

December 2, 2025
Our Picks

Meesho Becomes India’s Best IPO This Year, Mints New Billionaire

December 17, 2025

Good news: How AP journalists covered some of 2025’s happiest stories

December 17, 2025

Most Americans think cash gifts are acceptable, poll shows

December 17, 2025

School attendance plummeted during Texas measles outbreak

December 17, 2025

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

About Us
About Us

Welcome to World-Forbes.com
At World-Forbes.com, we bring you the latest insights, trends, and analysis across various industries, empowering our readers with valuable knowledge. Our platform is dedicated to covering a wide range of topics, including sports, small business, business, technology, AI, cybersecurity, and lifestyle.

Our Picks

After Klarna, Zoom’s CEO also uses an AI avatar on quarterly call

May 23, 2025

Anthropic CEO claims AI models hallucinate less than humans

May 22, 2025

Anthropic’s latest flagship AI sure seems to love using the ‘cyclone’ emoji

May 22, 2025

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Facebook X (Twitter) Instagram Pinterest
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA Policy
  • Privacy Policy
  • Terms & Conditions
© 2025 world-forbes. Designed by world-forbes.

Type above and press Enter to search. Press Esc to cancel.