🎯
Focusing
NVIDIA, Senior Performance Architect, Full-stack LLM Training Optimization.
-
NVIDIA
- Hangzhou, Zhejiang
- https://fanshiqing.github.io/
Pinned Loading
-
Megatron-LM
Megatron-LM PublicForked from NVIDIA/Megatron-LM
Ongoing research training transformer language models at scale, including: BERT & GPT-2
Python
-
DAPPLE
DAPPLE PublicForked from AlibabaPAI/DAPPLE
An Efficiency Pipelined Data Parallel Approach for Large Models Training
Python 3
-
grouped_gemm
grouped_gemm PublicForked from tgale96/grouped_gemm
PyTorch bindings for CUTLASS grouped GEMM.
-
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.