DeepSeek-R1 developers open-source their proprietary technologies one after another, enabling faster AI learning and inference



DeepSeek, a China-based AI development company, attracted a lot of attention when it announced the low-cost, high-performance AI model ' DeepSeek-R1 ' in January 2025. DeepSeek is open-sourcing several technologies related to AI development every day.




On February 21, 2025, DeepSeek declared that it would open source various technologies, stating, 'We are a small team working to realize AGI (artificial general intelligence). Starting next week, we will open source five repositories.' 'As part of the open source community, we believe that every piece of code we share will accelerate our journey (toward the realization of AGI).' The first repository was released on February 24, 2025, and one repository has been released each day since then.

◆Day 1: FlashMLA
The technology that was open-sourced as the first in a series of projects on Monday, February 24, 2025 is ' FlashMLA '. FlashMLA is an MLA decoding kernel developed for NVIDIA's Hopper architecture -based GPUs, which can speed up the processing of variable-length sequences.




The FlashMLA repository is available at the following link:

GitHub - deepseek-ai/FlashMLA
https://github.com/deepseek-ai/FlashMLA



◆Day 2: Deep EP
On Tuesday, February 25, 2025, DeepEP , a communication library that can speed up the training and inference of Mixture of Experts (MoE) models, was released. MoE is an architecture that uses multiple expert models as submodels and combines the outputs of each submodel to output the final result, and DeepEP is said to contribute to speeding up communication between submodels.




For more information on DeepEP, please see the links below.

GitHub - deepseek-ai/DeepEP: DeepEP: an efficient expert-parallel communication library
https://github.com/deepseek-ai/DeepEP



◆Day 3: DeepGEMM
In AI-related computational processes, a process called ' matrix multiplication (GEMM) ' is frequently performed. The technology that can speed up this GEMM is ' DeepGEMM ', which was open-sourced on Wednesday, February 26, 2025.




The source code for DeepGEMM is available at the following link.

GitHub - deepseek-ai/DeepGEMM: DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
https://github.com/deepseek-ai/DeepGEMM



Continued
DeepSeek announces open-source '3FS', a file system that accelerates AI - GIGAZINE

in Software, Posted by log1o_hf