Agent Skills: gpu-inference-server
Set up AI inference servers on cloud GPUs. Create private LLM APIs (vLLM, TGI), image generation endpoints, embedding services, and more. All with OpenAI-compatible interfaces that work with existing tools.
deployID: gpu-cli/gpu/gpu-inference-server
Install this agent skill to your local
Skill Files
Browse the full folder contents for gpu-inference-server.
Loading file tree…
Select a file to preview its contents.