Agent Skills: distributed-training
Use when training models across multiple GPUs or nodes, handling large models that don't fit in memory, or optimizing training throughput - covers DDP, FSDP, DeepSpeed ZeRO, model/data parallelism, and gradient checkpointingUse when ", " mentioned.
UncategorizedID: omer-metin/skills-for-antigravity/distributed-training
439
Install this agent skill to your local
Skill Files
Browse the full folder contents for distributed-training.
Loading file tree…
Select a file to preview its contents.