← Back to feed

DeepGEMM

GitHub Repo Pretty sure · Deep practical engineering.
https://github.com/deepseek-ai/DeepGEMM

DeepSeek's production GEMM library that actually ships LLM inference primitives: Mega MoE, FP8/FP4 kernels, JIT compilation. Real code, real performance numbers, real problem-solving.

5%
30%
65%
Slop 5%Signal 30%Science 65%

This is legitimate systems infrastructure. DeepSeek published the actual kernels behind v3 inference—Mega MoE with overlapped NVLink communication, FP8/FP4 GEMMs hitting 1550 TFLOPS on H800, JIT compilation avoiding install-time CUDA overhead. The code is production-grade (supports SM90/SM100, handles contiguous/masked MoE layouts, includes backward passes). Not academic—they shipped this. The science score reflects strong algorithmic contribution (tensor core optimization, kernel fusion stra...

6870 stars Cuda 2026-04-17 431 days old

Become a MFer to rate — log in