gpu-cli — community gpu-cli, community, ide skills, Claude Code, Cursor, Windsurf

v0.14.0
GitHub

About this Skill

Ideal for AI Agents requiring remote NVIDIA GPU acceleration for compute-intensive tasks Public facing GPU cli docs and issues

gpu-cli gpu-cli
[0]
[0]
Updated: 3/5/2026

Agent Capability Analysis

The gpu-cli skill by gpu-cli is an open-source community AI agent skill for Claude Code and other IDE workflows, helping agents execute tasks with better context, repeatability, and domain-specific guidance.

Ideal Agent Persona

Ideal for AI Agents requiring remote NVIDIA GPU acceleration for compute-intensive tasks

Core Value

Empowers agents to run local commands on remote NVIDIA GPUs, stream logs, and sync outputs using the `gpu` prefix, while providing quick diagnostics via `gpu doctor`, `gpu status`, and `gpu inventory` commands, utilizing JSON output for seamless integration

Capabilities Granted for gpu-cli

Automating machine learning model training on remote GPUs
Debugging GPU-accelerated applications using `gpu doctor`
Optimizing resource allocation with `gpu inventory` and `gpu status`

! Prerequisites & Limits

  • Requires NVIDIA GPU provider keys
  • Dependent on daemon and authentication setup
Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

gpu-cli

Install gpu-cli, an AI agent skill for AI agent workflows and automation. Works with Claude Code, Cursor, and Windsurf with one-command setup.

SKILL.md
Readonly

GPU CLI

GPU CLI runs local commands on remote NVIDIA GPUs by prefixing with gpu. It provisions a pod, syncs your code, streams logs, and syncs outputs back: uv run python train.py becomes gpu run uv run python train.py.

Quick diagnostics

bash
1gpu doctor --json # Check if setup is healthy (daemon, auth, provider keys) 2gpu status --json # See running pods and costs 3gpu inventory --json # See available GPUs and pricing

Command families

Getting started

CommandPurpose
gpu loginBrowser-based authentication
gpu logout [-y]Remove session
gpu init [--gpu-type T] [--force]Initialize project config
gpu upgradeOpen subscription upgrade page

Running code

CommandPurpose
gpu run <command>Execute on remote GPU (main command)
gpu run -d <command>Run detached (background)
gpu run -a <job_id>Reattach to running job
gpu run --cancel <job_id>Cancel a running job
gpu status [--json]Show project status, pods, costs
gpu logs [-j JOB] [-f] [--tail N] [--json]View job output
gpu attach <job_id>Reattach to job output stream
gpu stop [POD_ID] [-y]Stop active pod

Key gpu run flags: --gpu-type, --gpu-count <1-8>, --min-vram, --rebuild, -o/--output, --no-output, --sync, -p/--publish <PORT>, -e <KEY=VALUE>, -i/--interactive.

GPU inventory

CommandPurpose
gpu inventory [--available] [--min-vram N] [--max-price P] [--json]List GPUs with pricing

Volumes

CommandPurpose
gpu volume list [--detailed] [--json]List network volumes
gpu volume create [--name N] [--size GB] [--datacenter DC]Create volume
gpu volume delete <VOL> [--force]Delete volume
gpu volume extend <VOL> --size <GB>Increase size
gpu volume set-global <VOL>Set default volume
gpu volume status [--volume V] [--json]Volume usage
gpu volume migrate <VOL> --to <DC>Migrate to datacenter
gpu volume sync <SRC> <DEST>Sync between volumes

Vault (encrypted storage)

CommandPurpose
gpu vault list [--json]List encrypted output files
gpu vault export <PATH> <DEST>Export decrypted file
gpu vault stats [--json]Storage usage stats

Configuration

CommandPurpose
gpu config show [--json]Show merged config
gpu config validateValidate against schema
gpu config schemaPrint JSON schema
gpu config set <KEY> <VALUE>Set global config option
gpu config get <KEY>Get global config value

Authentication

CommandPurpose
gpu auth login [--profile P]Authenticate with cloud provider
gpu auth logoutRemove credentials
gpu auth statusShow auth status
gpu auth add <HUB>Add hub credentials (hf, civitai)
gpu auth remove <HUB>Remove hub credentials
gpu auth hubsList configured hubs

Organizations

CommandPurpose
gpu org listList organizations
gpu org create <NAME>Create organization
gpu org switch [SLUG]Set active org context
gpu org invite <EMAIL>Invite member
gpu org service-account create --name NCreate service token
gpu org service-account listList service accounts
gpu org service-account revoke <ID>Revoke token

LLM inference

CommandPurpose
gpu llm run [--ollama|--vllm] [--model M] [-y]Launch LLM inference
gpu llm info [MODEL] [--url URL] [--json]Show model info

ComfyUI workflows

CommandPurpose
gpu comfyui list [--json]Browse available workflows
gpu comfyui info <WORKFLOW> [--json]Show workflow details
gpu comfyui validate <WORKFLOW> [--json]Pre-flight checks
gpu comfyui run <WORKFLOW>Run workflow on GPU
gpu comfyui generate "<PROMPT>"Text-to-image generation
gpu comfyui stop [WORKFLOW] [--all]Stop ComfyUI pod

Notebooks

CommandPurpose
gpu notebook [FILE] [--run] [--new NAME]Run Marimo notebook on GPU

Alias: gpu nb

Serverless endpoints

CommandPurpose
gpu serverless deploy [--template T] [--json]Deploy endpoint
gpu serverless status [ENDPOINT] [--json]Endpoint status
gpu serverless logs [ENDPOINT]View request logs
gpu serverless list [--json]List all endpoints
gpu serverless delete [ENDPOINT]Delete endpoint
gpu serverless warm [--cpu|--gpu]Pre-warm endpoint

Templates

CommandPurpose
gpu template list [--json]Browse official templates
gpu template clear-cacheClear cached templates

Daemon control

CommandPurpose
gpu daemon status [--json]Show daemon health
gpu daemon startStart daemon
gpu daemon stopStop daemon
gpu daemon restartRestart daemon
gpu daemon logs [-f] [-n N]View daemon logs

Tools and utilities

CommandPurpose
gpu dashboardInteractive TUI for pods and jobs
gpu doctor [--json]Diagnostic checks
gpu agent-docsPrint agent reference to stdout
gpu update [--check]Update CLI
gpu changelog [VERSION]View release notes
gpu issue ["desc"]Report issue
gpu desktopDesktop app management
gpu supportOpen community Discord

Common workflows

  1. Setup: gpu login then gpu init
  2. Run job: gpu run python train.py --epochs 10
  3. With specific GPU: gpu run --gpu-type "RTX 4090" python train.py
  4. Detached job: gpu run -d python long_training.py then gpu status --json
  5. Check status: gpu status --json
  6. View logs: gpu logs --json
  7. Stop pods: gpu stop -y
  8. LLM inference: gpu llm run --ollama --model llama3 -y
  9. ComfyUI: gpu comfyui run flux_schnell
  10. Diagnose issues: gpu doctor --json

gpu run is pod-reuse oriented: after a command completes, the next gpu run reuses the existing pod until you gpu stop or cooldown ends.

JSON output

Most commands support --json for machine-readable output. Structured data goes to stdout; human-oriented status and progress messages go to stderr.

Commands with --json: status, logs, doctor, inventory, config show, daemon status, volume list, volume status, vault list, vault stats, comfyui list, comfyui info, comfyui validate, serverless deploy, serverless status, serverless list, template list, llm info.

Exit codes

CodeMeaningRecovery
0SuccessProceed
1General errorRead stderr
2Usage errorFix command syntax
10Auth requiredgpu auth login
11Quota exceededgpu upgrade or wait
12Not foundCheck resource ID
13Daemon unavailablegpu daemon start, retry
14TimeoutRetry
15CancelledRe-run if needed
130InterruptedRe-run if needed

Configuration

  • Project config: gpu.toml, gpu.jsonc, or pyproject.toml [tool.gpu]
  • Global config: ~/.gpu-cli/config.toml (via gpu config set/get)
  • Sync model: .gitignore controls upload; outputs patterns control download
  • Secrets and credentials must stay in the OS keychain, never plaintext project files
  • CI env vars: GPU_RUNPOD_API_KEY, GPU_SSH_PRIVATE_KEY, GPU_SSH_PUBLIC_KEY

References

  • Project generation and task setup: references/create.md
  • Debugging and common failures: references/debug.md
  • Config schema and field examples: references/config.md
  • Cost and GPU selection guidance: references/optimize.md
  • Persistent storage and volumes: references/volumes.md

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is gpu-cli?

Ideal for AI Agents requiring remote NVIDIA GPU acceleration for compute-intensive tasks Public facing GPU cli docs and issues

How do I install gpu-cli?

Run the command: npx killer-skills add gpu-cli/gpu/gpu-cli. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for gpu-cli?

Key use cases include: Automating machine learning model training on remote GPUs, Debugging GPU-accelerated applications using `gpu doctor`, Optimizing resource allocation with `gpu inventory` and `gpu status`.

Which IDEs are compatible with gpu-cli?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for gpu-cli?

Requires NVIDIA GPU provider keys. Dependent on daemon and authentication setup.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add gpu-cli/gpu/gpu-cli. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use gpu-cli immediately in the current project.

Related Skills

Looking for an alternative to gpu-cli or another community skill for your workflow? Explore these related open-source skills.

View All

widget-generator

Logo of f
f

f.k.a. Awesome ChatGPT Prompts. Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.

149.6k
0
AI

flags

Logo of vercel
vercel

flags is a Next.js feature management skill that enables developers to efficiently add or modify framework feature flags, streamlining React application development.

138.4k
0
Browser

zustand

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
AI

data-fetching

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
AI