LogoCookLLM Docs
LogoCookLLM Docs
HomeCookLLM

Principles

Tokenization
Tokenization BasicsBPE AlgorithmGPT TokenizersBPE Training Engineering
Model Architecture
Attention Mechanisms
Position Encoding
Position Encoding BasicsRoPE Math DerivationRoPE ImplementationLength Extrapolation
GPU Programming Basics
GPU Architecture BasicsTensor LayoutTriton Basics: Vector Add
FlashAttention
Flash Attention PrinciplesFrom Naive to Auto-TuningBlock Pointers and Multi-Dim SupportCausal Masking OptimizationGrouped Query AttentionBackward Pass

Hands-on Training

X (Twitter)

多维混合并行

Premium

ParallelContext 坐标系统与 TP+DP+PP 的工业级组合

Companion Code
👨‍🍳

Content is cooking...

We're preparing high-quality content for you. Stay tuned!

Table of Contents

为什么要混合
ParallelContext:多维坐标系统
通信组的构建
2D 混合示例:TP + DP
3D 混合:TP + DP + PP
配置建议
总结