Close Menu
Karachi Chronicle
  • Home
  • AI
  • Business
  • Entertainment
  • Fashion
  • Politics
  • Sports
  • Tech
  • World

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

Wall Street is lifted as data, and business revenues show consumer strength

Dell employee satisfaction ratings fell by almost 50% in two years

Who will make outwardly film academy president Janet Yang successful?

Facebook X (Twitter) Instagram
  • Home
  • About us
  • Advertise
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
Facebook X (Twitter) Instagram Pinterest Vimeo
Karachi Chronicle
  • Home
  • AI
  • Business
  • Entertainment
  • Fashion
  • Politics
  • Sports
  • Tech
  • World
Karachi Chronicle
You are at:Home » This free “inference” AI model beats OpenAI’s o1 – no $20 monthly fee
AI

This free “inference” AI model beats OpenAI’s o1 – no $20 monthly fee

Adnan MaharBy Adnan MaharJanuary 14, 2025No Comments4 Mins Read0 Views
Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
Share
Facebook Twitter LinkedIn Pinterest WhatsApp Email


The team at Novasky, “a collaborative initiative led by students and advisors from the Sky Computing Lab at the University of California, Berkeley,” has accomplished something that seemed impossible just a few months ago. They created a high-performance AI inference model that cost less than $450 to train.

Unlike traditional LLMs, which simply predict the next word in a sentence, so-called “reasoning models” understand a problem, analyze different approaches to solving it, and implement the optimal solution. It is designed to. This makes these models difficult to train and configure, as they not only have to predict the best response based on a training dataset, but also have to “reason” throughout the problem-solving process.

That’s why a subscription to ChatGPT Pro, which runs the latest O3 inference models, costs $200 per month. OpenAI claims that training and running these models is expensive.

The new Novasky model, called Sky-T1, corresponds to OpenAI’s first inference model, known as o1 (also known as Strawberry), which was released in September 2024 and costs users $20 per month. By comparison, Sky-T1 is a 32 billion parameter model that can only be run locally on your home computer if you have a powerful 24 GB GPU, such as the RTX 4090 or the older 3090 Ti. And it’s free.

We’re not talking about watered down versions. Sky-T1-32B-Preview achieved 43.3% accuracy on AIME2024 math problems, outperforming OpenAI o1’s 40%. It received a score of 56.8% on LiveCodeBench-Medium compared to 54.9% on o1-preview. The model maintained good performance on other benchmarks, reaching 82.4% on the Math500 problem and scoring 81.4% on o1-preview.

The timing couldn’t be more interesting. AI inference competition has been heating up recently. OpenAI’s o3 has garnered attention by outperforming humans on general intelligence benchmarks, sparking debate over whether we’re seeing early AGI or general artificial intelligence. Meanwhile, China’s Deepseek v3 made headlines last year for outperforming OpenAI’s o1 while using fewer resources and being open source.

🚀 Introducing DeepSeek-V3!

Biggest leap forward:
⚡ 60 tokens/second (3x faster than V2!)
💪 Enhanced features
🛠 API compatibility remains the same
🌍 Completely open source models and papers

🐋1/n pic.twitter.com/p1dV9gJ2Sd

— DeepSeek (@deepseek_ai) December 26, 2024

But Berkeley’s approach is different. Rather than chasing raw power, the team focuses on making powerful inference models available to the masses as cheaply as possible, making them easy to fine-tune and use locally without expensive enterprise hardware. We built a model that can be run on a computer.

“Incredibly, Sky-T1-32B-Preview was trained for less than $450, demonstrating that high-level inference functionality can be reproduced affordably and efficiently. All code is open source. ,” Novasky said in an official blog post.

Currently, OpenAI does not provide free access to inference models, but it does provide free access to less sophisticated models.

The prospect of being able to fine-tune an inference model to achieve domain-specific excellence for less than $500 is particularly appealing to developers. This is because such specialized models can outperform more powerful general-purpose models in the domain of interest. This cost-effective specialization opens new possibilities for intensive applications across scientific disciplines.

The team trained the model in just 19 hours using Nvidia H100 GPUs, following what they called a “recipe” that most developers should be able to reproduce. Training data appears to be the biggest hit for AI challenges.

“Our final data includes 5K coding data from APP and TACO, and 10K math data from the AIME, MATH, and Olympiads subsets of the NuminaMATH dataset. Additionally, we have STILL- We maintain 1,000 scientific and puzzle data from 2,” Novasky said.

The dataset was diverse enough to allow the model to be flexible about different types of problems. Novasky used another open source inference AI model, QwQ-32B-Preview, to generate the data and fine-tune the Qwen2.5-32B-Instruct open source LLM. The result was a powerful new model with inference capabilities that later became the Sky-T1.

A key finding from the team’s research is that bigger is better when it comes to AI models. Experiments using smaller parameter versions of 7 billion and 14 billion showed only small benefits. The sweet spot turned out to be 32 billion parameters. Large enough to avoid repeating the output, but not so large as to be impractical.

If you want your own version of a model beyond OpenAI o1, you can download Sky-T1 at Hugging Face. If your GPU isn’t powerful enough, but you still want to try it out, there are quantized versions from 8-bit to 2-bit, so you can test the next best thing, sacrificing accuracy for speed. Potato PC.

Please be careful. The developers warn that such levels of quantization are “not recommended for most purposes.”

Edited by Andrew Hayward

Generally intelligent newsletter

A weekly AI journey told by Gen, a generative AI model.





Source link

Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
Previous ArticleF-22 Raptor banned from export from US
Next Article Countermeasures against Russia’s hybrid war
Adnan Mahar
  • Website

Adnan is a passionate doctor from Pakistan with a keen interest in exploring the world of politics, sports, and international affairs. As an avid reader and lifelong learner, he is deeply committed to sharing insights, perspectives, and thought-provoking ideas. His journey combines a love for knowledge with an analytical approach to current events, aiming to inspire meaningful conversations and broaden understanding across a wide range of topics.

Related Posts

Dig into Google Deepmind CEO “Shout Out” Chip Engineers and Openai CEO Sam Altman, Sundar Pichai responds with emojis

June 1, 2025

Google, Nvidia invests in AI startup Safe Superintelligence, co-founder of Openai Ilya Sutskever

April 14, 2025

This $30 billion AI startup can be very strange by a man who said that neural networks may already be aware of it

February 24, 2025
Leave A Reply Cancel Reply

Top Posts

20 Most Anticipated Sex Movies of 2025

January 22, 2025165 Views

President Trump’s SEC nominee Paul Atkins marries multi-billion dollar roof fortune

December 14, 2024106 Views

Alice Munro’s Passive Voice | New Yorker

December 23, 202467 Views

How to tell the difference between fake and genuine Adidas Sambas

December 26, 202452 Views
Don't Miss
AI June 1, 2025

Dig into Google Deepmind CEO “Shout Out” Chip Engineers and Openai CEO Sam Altman, Sundar Pichai responds with emojis

Demis Hassabis, CEO of Google Deepmind, has expanded public approval to its chip engineers, highlighting…

Google, Nvidia invests in AI startup Safe Superintelligence, co-founder of Openai Ilya Sutskever

This $30 billion AI startup can be very strange by a man who said that neural networks may already be aware of it

As Deepseek and ChatGpt Surge, is Delhi behind?

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

About Us
About Us

Welcome to Karachi Chronicle, your go-to source for the latest and most insightful updates across a range of topics that matter most in today’s fast-paced world. We are dedicated to delivering timely, accurate, and engaging content that covers a variety of subjects including Sports, Politics, World Affairs, Entertainment, and the ever-evolving field of Artificial Intelligence.

Facebook X (Twitter) Pinterest YouTube WhatsApp
Our Picks

Wall Street is lifted as data, and business revenues show consumer strength

Dell employee satisfaction ratings fell by almost 50% in two years

Who will make outwardly film academy president Janet Yang successful?

Most Popular

ATUA AI (TUA) develops cutting-edge AI infrastructure to optimize distributed operations

October 11, 20020 Views

10 things you should never say to an AI chatbot

November 10, 20040 Views

Character.AI faces lawsuit over child safety concerns

December 12, 20050 Views
© 2025 karachichronicle. Designed by karachichronicle.
  • Home
  • About us
  • Advertise
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions

Type above and press Enter to search. Press Esc to cancel.