Close Menu
Karachi Chronicle
  • Home
  • AI
  • Business
  • Entertainment
  • Fashion
  • Politics
  • Sports
  • Tech
  • World

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

Apple may use Intel to make M-series and A-series chips starting in 2027, analysts suggest

New shows and movies available to stream on Netflix, Hulu, Prime Video, Apple TV, and more

Anduril faces intense investigation after multiple drones crash during US Air Force test

Facebook X (Twitter) Instagram
  • Home
  • About us
  • Advertise
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
Facebook X (Twitter) Instagram Pinterest Vimeo
Karachi Chronicle
  • Home
  • AI
  • Business
  • Entertainment
  • Fashion
  • Politics
  • Sports
  • Tech
  • World
Karachi Chronicle
You are at:Home » AI hit “peak data”. Google researchers think they’ve found a solution.
AI

AI hit “peak data”. Google researchers think they’ve found a solution.

Adnan MaharBy Adnan MaharJanuary 5, 2025No Comments6 Mins Read0 Views
Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
Share
Facebook Twitter LinkedIn Pinterest WhatsApp Email


OpenAI co-founder Ilya Satskeva recently said that the AI ​​industry has reached “data peak.” DeepMind researchers believe that the output from new “inference” models is a source of fresh AI training data. A new AI technology known as test-time computing will be tested in 2025.

OpenAI co-founder Ilya Sutskever announced something at a recent conference that should have horrified the AI ​​industry.

“We have reached peak data and there is no more,” he exclaimed in a speech at the annual New Lips event in December.

All useful data on the internet is already being used to train AI models. This process, known as pre-training, has given rise to many of the features of recent generative AI, including ChatGPT. But improvements have slowed, and Sutskever said this era “will definitely come to an end.”

This is a frightening prospect, as trillions of dollars in stock market value and investment in AI rest on ever-improving models.

But most AI experts don’t seem too worried. why?

Calculations during inference

There may be a way around this data wall. This relates to a relatively new technology that helps AI models “think” about difficult tasks for long periods of time.

This approach, called test-time or inference-time computing, breaks the query into smaller tasks and transforms each into a new prompt for the model to work on. Each step requires executing a new request. This is known as the inference stage of AI.

This generates a series of inferences that address each part of the problem. The model does not move on to the next stage until each part is correct and ultimately yields a better final response.

OpenAI released a model called o1 in September that uses inference-time computing. This was quickly followed by Google and Chinese AI research institute DeepSeek, which rolled out similar “inference” models.

“Iterative loop of self-improvement”

Benchmark-based tests of these new models show that they often produce better output than previous top AI crops, especially on math questions and similar tasks that yield definitive, unambiguous answers. Ta.

Related articles

Here’s where things get interesting. What if these high-quality outputs were used for new training data? This large amount of new information could be fed back into training runs of other AI models to yield even better results.

Google DeepMind researchers published a study on test-time computing in August, proposing the technique as a potential way to overcome peak data barriers and continue improving large-scale language models.

“We believe that in the future, the results of applying additional test-time calculations could be distilled back to the base LLM, allowing for iterative self-improvement loops,” the researchers wrote. . “To this end, future work should extend our findings and study how the results of applying test-time calculations can be used to improve the based LLM itself. there is.”

Chat with researchers during testing

Authors: Charlie Snell, Jaehoon Lee, Kelvin Xu, Aviral Kumar. Xu is still at Google and Kumar spends some of his time at DeepMind, while Lee left to join OpenAI rival Anthropic.

Snell co-authored the paper while interning at Google DeepMind. He’s now back at the University of California, Berkeley, so I called him to ask what inspired his research.

“I was motivated by several factors that are holding back the expansion of pre-training, particularly the finite supply of data,” he told me in a recent interview.

“If an AI model can use extra inference time computations to improve its output, that’s the way to generate better synthetic data,” he added. “This is a useful new source of training data. This seems like a promising way to avoid pre-training data bottlenecks.”

Satya was satisfied

In a recent video podcast, Microsoft CEO Satya Nadella seemed unperturbed and rather perky when asked about the slow pace of improving AI models and the lack of new, high-quality training data.

He described calculating inference time as “another scaling law.”

“So you have pre-training, and then you effectively do this test-time sampling, and then you create a token that you can go back to pre-training to create an even more powerful model that runs in inference. ” he explained.

“I think this is a great way to improve the capabilities of the models,” Nadella added with a smile.

Sutskever also mentioned test-time computing as one possible solution to the peak data problem during his Neurips talk in early December.

Test time for test time computing

In 2025, this approach will be put to the test. Snell is optimistic, but this is not an absolute guarantee.

“Over the past three years or so, the advances in AI have seemed to become more apparent,” he said. “Right now we’re in exploration mode.”

There is one unanswered question. How generalizable is this test-time calculation method? Snell said it works well for questions where the answers are known and can be verified, such as math assignments.

“But many things that require reasoning are not easy to check, such as writing an essay. There is often no clear answer as to how good this is,” he says. I explained.

Still, there are early signs of success, and Snell suspects that output from these types of inferential AI models is already being used to train new models.

“This synthetic data could very well be better than the data publicly available on the Internet,” he said.

If the outputs from OpenAI’s o1 model are better than the startup’s previous top-of-the-line model, GPT-4, these new outputs could theoretically be reused to train future AI models, Snell said. I explained.

He shared a theoretical example. For example, if o1 scores 90% on a particular AI benchmark, you can feed those answers into GPT-4 to boost that model to 90% as well.

“If you have a large set of prompts, you can take large amounts of data from o1, create a large set of training samples, and pre-train a new model on them, or use GPT-4. You can also continue to improve your training,” Snell said. .

A TechCrunch report in late December suggested that DeepSeek may have used output from OpenAI’s o1 to train its own AI models. The latest product, called DeepSeek V3, performs well on industry benchmarks.

“They were probably the first people to recreate o1,” Snell said. “I asked the OpenAI folks what they thought about this. They say it looks the same, but I don’t know how DeepSeek managed to do this so quickly.”

OpenAI and DeepSeek did not respond to requests for comment.



Source link

Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
Previous Article3-day state-level ‘Poly Techfest’ to start in Vijayawada from January 6, 2025
Next Article “Meeting urgent national security needs”
Adnan Mahar
  • Website

Adnan is a passionate doctor from Pakistan with a keen interest in exploring the world of politics, sports, and international affairs. As an avid reader and lifelong learner, he is deeply committed to sharing insights, perspectives, and thought-provoking ideas. His journey combines a love for knowledge with an analytical approach to current events, aiming to inspire meaningful conversations and broaden understanding across a wide range of topics.

Related Posts

AI trackers: AI agents open the door to new hacking threats

November 15, 2025

FACTS IN : FACTS OUT – Join the call for truth in AI at the global stand for trusted news

November 6, 2025

Elon Musk says AI will take over all jobs and humans will be free to grow vegetables

October 22, 2025
Leave A Reply Cancel Reply

Top Posts

20 Most Anticipated Sex Movies of 2025

January 22, 2025703 Views

President Trump’s SEC nominee Paul Atkins marries multi-billion dollar roof fortune

December 14, 2024131 Views

How to tell the difference between fake and genuine Adidas Sambas

December 26, 2024112 Views

Alice Munro’s Passive Voice | New Yorker

December 23, 202481 Views
Don't Miss
AI November 15, 2025

AI trackers: AI agents open the door to new hacking threats

As AI agents evolve, cybersecurity experts warn that they could become tools for hackers and…

FACTS IN : FACTS OUT – Join the call for truth in AI at the global stand for trusted news

Elon Musk says AI will take over all jobs and humans will be free to grow vegetables

New study finds AI assistants make widespread errors when it comes to news

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

About Us
About Us

Welcome to Karachi Chronicle, your go-to source for the latest and most insightful updates across a range of topics that matter most in today’s fast-paced world. We are dedicated to delivering timely, accurate, and engaging content that covers a variety of subjects including Sports, Politics, World Affairs, Entertainment, and the ever-evolving field of Artificial Intelligence.

Facebook X (Twitter) Pinterest YouTube WhatsApp
Our Picks

Apple may use Intel to make M-series and A-series chips starting in 2027, analysts suggest

New shows and movies available to stream on Netflix, Hulu, Prime Video, Apple TV, and more

Anduril faces intense investigation after multiple drones crash during US Air Force test

Most Popular

10 things you should never say to an AI chatbot

November 10, 20040 Views

Analyst warns Salesforce investors about AI agent optimism

July 1, 20070 Views

Musk says the Xai’s Grok 3 chatbot will be released on Monday

July 1, 20070 Views
© 2025 karachichronicle. Designed by karachichronicle.
  • Home
  • About us
  • Advertise
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions

Type above and press Enter to search. Press Esc to cancel.