Agent Skills: torch-tensor-parallelism

Guidance for implementing tensor parallelism in PyTorch, including ColumnParallelLinear and RowParallelLinear layers. This skill should be used when implementing distributed tensor parallel operations, sharding linear layers across multiple GPUs, or simulating collective operations like all-gather and all-reduce for parallel computation.

UncategorizedID: benchflow-ai/skillsbench/torch-tensor-parallelism

Install this agent skill to your local

pnpm dlx add-skill https://github.com/benchflow-ai/skillsbench/torch-tensor-parallelism

Skill Files

Browse the full folder contents for torch-tensor-parallelism.

Download Skill

Loading file tree…

Select a file to preview its contents.