Close Menu
Karachi Chronicle
  • Home
  • AI
  • Business
  • Entertainment
  • Fashion
  • Politics
  • Sports
  • Tech
  • World

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

The world’s largest air force with the F-35 fleet in 2025

AI systems learn from many types of scientific information and run experiments to discover new materials | MIT News

Among the most troublesome relationships in healthcare AI

Facebook X (Twitter) Instagram
  • Home
  • About us
  • Advertise
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
Facebook X (Twitter) Instagram Pinterest Vimeo
Karachi Chronicle
  • Home
  • AI
  • Business
  • Entertainment
  • Fashion
  • Politics
  • Sports
  • Tech
  • World
Karachi Chronicle
You are at:Home » Deepseek’s Cerebras CEO: Every time your computing gets cheaper, the market grows
Tech

Deepseek’s Cerebras CEO: Every time your computing gets cheaper, the market grows

Adnan MaharBy Adnan MaharFebruary 10, 2025No Comments6 Mins Read0 Views
Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
Share
Facebook Twitter LinkedIn Pinterest WhatsApp Email


Cerebras-Feldman-2024-Large

“When they’re 50 or 70 times faster than the competition, they can do things they can’t do at all,” says Andrew Feldman, CEO of Celebras.

Tiannan Ray/ZDNET

The AI ​​computer pioneer Celebras system has been “smashed down” by the demand to run Deepseek’s R1 major language model, said Andrew Feldman, co-founder and CEO of the company.

“We’re thinking about ways to meet demand. That’s a big deal,” Feldman told me in an interview last week on Zoom.

The Deepseek R1 is told as an artificial intelligence basin moment. This is because the pre-training cost of the model can achieve good or better results, while the cost of dominant models such as Openai’s GPTO1 is small.

The impact of Deepseek on the economics of AI is important, Feldman showed. But the deeper outcome is to promote even bigger AI systems.

Also: If you are confused, you can try deepseek R1 without security risk

“As we reduce computational costs, the market gets bigger and bigger,” Feldman said.

After the AI ​​model became a sense that includes not only Celebras but much larger companies like Amazon’s AWS, many AI cloud services have been rushing to provide Deepseek’s inference. (You can try out Cerebras inference services here.)

The edge of the Celebrus is speed. According to Feldman, running inference on the company’s CS-3 computer gives you 57 times faster output than other DeepSeek service providers.

Celebras also emphasizes its speed compared to other larger language models. In the demonstration of inference problems performed by Deepseek running on the O1 Mini in Cerebras versus Openai, the Cerebras machine is finished in 2.5 seconds, with O1 taking 22 seconds to complete the task.

cerebras-r1-deepseek-2025-slide-18

Celebrus System

“This speed cannot be achieved with the number of GPUs,” Feldman said, referring to the chips Nvidia, Advanced Micro Devices and Intel sells on AI.

The challenge for those hosting DeepSeek is that, like other so-called inference models such as Openai’s GPTO1, DeepSeek uses much more computing power when generating output during inference, and results at user prompts It will be difficult to deliver in a timely manner. .

“The basic GPT model makes one inference that goes through all parameters of every word in the input of the prompt,” explained Feldman.

“These inference models, or chain models of thinking, do “multiple times for each word.”

cerebras-r1-deepseek-2025-slide-15

Celebrus System

Cerebras followed one standard procedure for companies that wanted to perform DeepSeek’s inference. Embracing the R1 neural parameters (weight) and using the parameters to train a smaller open source model. Create a “distillation” for R1.

“We were able to do it very quickly, and not only did it faster than anyone else, but we were able to produce results that weren’t just a bit, but not just a bit,” Feldman said. Ta.

Also: I tested Deepseek’s R1 and V3 coding skills, but we are all not destined (still)

The results of Celebras using the Deepseek R1 distilled lama 70b are comparable to the model’s published accuracy benchmarks. Cerebras has not disclosed the Deepseek R1 distilled lama 70b pricing for inference, but said it is “competitively priced, especially to provide performance in the top industry.”

Deepseek’s breakthrough has several meanings.

One is that it is a big win for open source AI and Feldman means that he posts neural parameters for download. Many of the advancements in new AI models can be replicated by researchers who have access to weights, even if the source code is not accessible. Private models such as the GPT-4 do not disclose weight.

“Open source certainly has that much,” Feldman said. “This was the first open source inference model for top flights.”

While Deepseek’s economics has surprised the world of AI, Feldman said the advancement will lead to continued investment in AI’s cutting-edge chips and networking technologies.

Also: Is Deepseek’s new image model another win for cheap AI?

“In the last 50 years, the public market has been wrong every time,” Feldman said, implying a massive sale of shares in NVIDIA and other AI technology providers. “Every time computing becomes cheaper, they (public market investors) systematically assume that the market will be smaller. And for over 50 years, it has made the market bigger.”

Feldman cited examples of lowering the price of X86 PCs, leading to more PCs being sold and used. Today he said, “You have 25 computers in your home. You have one in your pocket, you have one in your work, you have one in your dishwasher “Your washing machine has one, your TV has each.”

Not only is it the same, but it also allows large and large AI systems to be built, allowing results to exceed the scope of product AI. This is the point Feldman has been doing since Celebrus was founded almost 10 years ago.

“When it’s 50 or 70 times faster than the competition, they can do things they can’t do at all,” he said. “At some point, differences in degrees become differences in kind.”

Also: Apple researchers reveal the secret source behind DeepseekAI

Cerebras launched its public inference service last August and showed much faster speeds than most other providers to run generation AI. It claims to be the “world’s fastest AI inference provider.”

Apart from the distilled Llama model, Cerebras currently does not offer a complete R1 in reasoning as costs are kept down for most customers.

“The 67.1 billion parameter model is an expensive model to run,” Feldman said. “What we saw on the Llama 405B was much less because there was a huge amount of interest on the 70B node and it was much more expensive on the 405B node. That’s the market today.”

Celebras has customers paying for the full Lama 405B because “added precision worth the extra cost.”

Celebras also bets that privacy and security are features that can be used to your advantage. Initial enthusiasm for Deepseek was followed by numerous reports on concerns regarding the processing of the model’s data.

“When you use the app, your data is sent to China,” says Feldman of The Android and IOS native apps Feldman of The Android and IOS. “If you use us, your data is hosted in the US. We don’t store your weights or your information.

When asked about the many security vulnerabilities researchers have published about Deepseek R1, Feldman was philosophical. He showed that as technology matures, some problems will be resolved.

Also: Security companies discover that Deepseek has a “direct link” to Chinese government servers

“The industry is moving very quickly. No one has seen anything like that,” Feldman said. “It’s been better for more than a week, over a month. But is it perfect? ​​No. Should I use LLM (large language model) to replace common sense?

Following the announcement of the R1, Cerebras announced last Thursday that it had added support for running for Le Chat, an inference prompt run by French AI startup Mistral. Running Le Chat’s “Flash Answers” feature at 1,100 tokens per second, the model becomes the world’s fastest AI assistant, saying, “It’s 10 times faster than typical models like ChatGpt 4o, Sonnet 3.5, and Deepseek R1.” I did. . ”



Source link

Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
Previous ArticleEntertainment News Live Update February 11, 2025: YouTube defeats the controversial Indian potential episode of Samay Raina and Ranveer Allahbadia
Next Article 10 Most Promising Marvel Movies That Were Tragically Canceled
Adnan Mahar
  • Website

Adnan is a passionate doctor from Pakistan with a keen interest in exploring the world of politics, sports, and international affairs. As an avid reader and lifelong learner, he is deeply committed to sharing insights, perspectives, and thought-provoking ideas. His journey combines a love for knowledge with an analytical approach to current events, aiming to inspire meaningful conversations and broaden understanding across a wide range of topics.

Related Posts

Googleबनी$ 3

September 16, 2025

Tesla engineers will resign in eight years. He points out CEO Elon Musk as the main reason, accusing him of “liing to the public and manipulating him…”

September 12, 2025

Ant Group unveils its own Tesla Optimus competitor, R1 humanoid robot

September 11, 2025
Leave A Reply Cancel Reply

Top Posts

20 Most Anticipated Sex Movies of 2025

January 22, 2025452 Views

President Trump’s SEC nominee Paul Atkins marries multi-billion dollar roof fortune

December 14, 2024122 Views

How to tell the difference between fake and genuine Adidas Sambas

December 26, 202485 Views

Alice Munro’s Passive Voice | New Yorker

December 23, 202474 Views
Don't Miss
AI September 25, 2025

AI systems learn from many types of scientific information and run experiments to discover new materials | MIT News

Machine learning models can speed up discovery of new materials by making predictions and proposing…

Among the most troublesome relationships in healthcare AI

Does access to AI become a fundamental human right? Sam Altman says, “Everyone would want…”

Google’s Gemini AI is on TV

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

About Us
About Us

Welcome to Karachi Chronicle, your go-to source for the latest and most insightful updates across a range of topics that matter most in today’s fast-paced world. We are dedicated to delivering timely, accurate, and engaging content that covers a variety of subjects including Sports, Politics, World Affairs, Entertainment, and the ever-evolving field of Artificial Intelligence.

Facebook X (Twitter) Pinterest YouTube WhatsApp
Our Picks

The world’s largest air force with the F-35 fleet in 2025

AI systems learn from many types of scientific information and run experiments to discover new materials | MIT News

Among the most troublesome relationships in healthcare AI

Most Popular

10 things you should never say to an AI chatbot

November 10, 20040 Views

Character.AI faces lawsuit over child safety concerns

December 12, 20050 Views

Analyst warns Salesforce investors about AI agent optimism

July 1, 20070 Views
© 2025 karachichronicle. Designed by karachichronicle.
  • Home
  • About us
  • Advertise
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions

Type above and press Enter to search. Press Esc to cancel.