README » Algorithm Implementations
This directory contains various algorithm implementations for distributed computing and matrix operations.
00_load: Load operations across multiple GPUs01_store: Store operations across multiple GPUs02_all_load: Load operations where all GPUs load simultaneously03_all_store: Store operations where all GPUs store simultaneously04_atomic_add: Atomic add operations across multiple GPUs05_atomic_xchg: Atomic exchange operations across multiple GPUs
- 06_message_passing: Point-to-point message passing (load/store and put/get operations)
07_gemm_all_scatter: Matrix multiplication with all-scatter communication08_gemm_atomics_all_reduce: Matrix multiplication with all-reduce using atomics09_gemm_one_shot_all_reduce: Matrix multiplication with one-shot all-reduce10_gemm_all_scatter_wg_specialization: Matrix multiplication with all-scatter using workgroup specialization11_gemm_all_scatter_producer_consumer: Matrix multiplication with all-scatter using producer-consumer concurrent kernels12_gemm_all_scatter_bulk_synchronous: Matrix multiplication with all-scatter using the bulk synchronous parallel approach13_flash_decode: Fused Flash Decode Attention for accelerating LLM inference14_all_gather_gemm: Fused All-Gather + GEMM with Pull and Push models
benchmark: Benchmarking utilities and performance testing toolscommon: Common utilities and shared code for examples
# Example command to run distributed load operations
python examples/00_load/load_bench.py --num_ranks 8 # Load across GPUs
python examples/02_all_load/all_load_bench.py --num_ranks 8 # Simultaneous load on all GPUs
# Example command to run distributed store operations
python examples/01_store/store_bench.py --num_ranks 8 # Store across GPUs
python examples/03_all_store/all_store_bench.py --num_ranks 8 # Simultaneous store on all GPUs
# Example command to run atomic operations
python examples/04_atomic_add/atomic_add_bench.py --num_ranks 8 # Atomic add across GPUs
python examples/05_atomic_xchg/atomic_xchg_bench.py --num_ranks 8 # Atomic exchange across GPUs
# Example command to run message passing
python examples/06_message_passing/message_passing_put.py --num_ranks 8
python examples/06_message_passing/message_passing_load_store.py --num_ranks 8
# Example command to run benchmark with all-scatter algorithm
python examples/07_gemm_all_scatter/benchmark.py --benchmark --validate --num_ranks 8
# Example command to run benchmark with all-reduce algorithm
python examples/08_gemm_atomics_all_reduce/benchmark.py --benchmark --validate --num_ranks 8
# Example command to run benchmark with one-shot all-reduce algorithm
python examples/09_gemm_one_shot_all_reduce/benchmark.py --benchmark --validate --num_ranks 8
# Example command to run benchmark with all-scatter and workgroup specialization
python examples/10_gemm_all_scatter_wg_specialization/benchmark.py --benchmark --validate --num_ranks 8
# Example command to run benchmark with all-scatter producer-consumer pattern
python examples/11_gemm_all_scatter_producer_consumer/benchmark.py --benchmark --validate --num_ranks 8
# Example command to run benchmark with all-scatter bulk synchronous approach
python examples/12_gemm_all_scatter_bulk_synchronous/benchmark.py --benchmark --validate --num_ranks 8
# Flash Decode Attention - simple example run
python examples/13_flash_decode/example_run.py --num_ranks 8
# All-Gather + GEMM - Pull model
python examples/14_all_gather_gemm/example_run_pull.py --num_ranks 8
# All-Gather + GEMM - Push model
python examples/14_all_gather_gemm/example_run_push.py --num_ranks 8