Agent Skills: gpu-inference-server

Set up AI inference servers on cloud GPUs. Create private LLM APIs (vLLM, TGI), image generation endpoints, embedding services, and more. All with OpenAI-compatible interfaces that work with existing tools.

deployID: gpu-cli/gpu/gpu-inference-server

Install this agent skill to your local

pnpm dlx add-skill https://github.com/gpu-cli/gpu/gpu-inference-server

Skill Files

Browse the full folder contents for gpu-inference-server.

Download Skill

Loading file tree…

Select a file to preview its contents.