This site is updated Hourly Every Day

Trending Featured Popular Today, Right Now

Colorado's Only Reliable Source for Daily News @ Marijuana, Psychedelics & more...

Post: What is AI distillation and what does it mean for OpenAI?

Picture of Anschutz Medical Campus

Anschutz Medical Campus

AnschutzMedicalCampus.com is an independent website not associated or affiliated with CU Anschutz Medical Campus, CU, or Fitzsimons innovation campus.

Recent Posts

Microdosing 101
Microdosing 101

Key points Microdosing should be approached thoughtfully with proper research

Anschutz Medical Campus

What is AI distillation and what does it mean for OpenAI?
Facebook
X
LinkedIn
WhatsApp
Telegram
Threads
Email

News DeepSeek’s low-cost chatbot has shaken markets.

The tech industry and stock markets have been trying to grasp how a small, relatively unknown Chinese company was able to develop a sophisticated artificial intelligence chatbot on par with OpenAI’s ChatGPT at a fraction of the cost.

One possible answer being floated in tech circles is distillation, an AI training method that uses bigger “teacher” models to train smaller but faster-operating “student” models.

DeepSeek claims to have achieved a similar level of performance as OpenAI’s o1 model at a fraction of the cost through “optimized codesign of algorithms, frameworks, and hardware.”

This sparked a sharp selloff in tech shares as investors considered whether the Chinese company’s low-budget approach signaled the end of the AI investment race and the dominance of US tech giants.

But questions soon arose, with some in the industry speculating that the company had piggybacked on OpenAI’s developments.

Such speculation was fueled when Bloomberg reported that Microsoft and OpenAI had launched a probe into whether DeepSeek improperly obtained data from OpenAI to train its own model. OpenAI told the Financial Times on January 28 that it has seen evidence of distillation, though it did not make that evidence public.

Microsoft and DeepSeek did not immediately respond to Nikkei Asia’s request for comment.

Distillation itself is not a new technique and is not necessarily controversial. Nvidia’s Minitron and Falcon 3, which was developed by the Technology Innovation Institute in the UAE, both used the technique, likely using their own families of large language models (LLMs) as the teacher. It has become increasingly popular since 2024 amid demand from businesses wanting to utilize LLMs in their services.Big LLM models, however, are “difficult to handle, and you would need a vast number of graphics processing units (GPUs) for its deployment,” said an engineer at an AI startup in Japan.GPUs are the main reason AI systems are so expensive. Nvidia’s signature H100 chips, for example, can cost USD 30,000–35,000 each. Distillation drastically cuts development time and costs, and results in models that can operate faster than their bigger counterparts.The issue for DeepSeek is whether its low-cost model […]

Leave a Reply

Your email address will not be published. Required fields are marked *

You Might Be Interested...