Agent Skills: torch-tensor-parallelism
Guidance for implementing tensor parallelism in PyTorch, including ColumnParallelLinear and RowParallelLinear layers. This skill should be used when implementing distributed tensor parallel operations, sharding linear layers across multiple GPUs, or simulating collective operations like all-gather and all-reduce for parallel computation.
UncategorizedID: benchflow-ai/skillsbench/torch-tensor-parallelism
278174
Install this agent skill to your local
Skill Files
Browse the full folder contents for torch-tensor-parallelism.
Loading file tree…
Select a file to preview its contents.