News

How DeepSeek used distillation to train its artificial intelligence model, and what it means for companies such as OpenAI By Jasmine Wu, CNBC and Deirdre Bosa, CNBC • Published February 21, 2025 ...
While model distillation, the method of teaching smaller, efficient models (students) from larger, more complex ones (teachers), isn't new, DeepSeek’s implementation of it is groundbreaking.
Distillation is also a victory for advocates of open models, where the technology is made freely available for developers to build upon. DeepSeek has made its recent models also open for developers.
Leading artificial intelligence firms including OpenAI, Microsoft, and Meta are turning to a process called “distillation” in the global race to create AI models that are cheaper for consumers ...
Experts say AI model distillation is likely widespread and hard to detect, but DeepSeek has not admitted to using it on its full models Reading Time:3 minutes Why you can trust SCMP 99+ ...