ml-api-endpoint — community ml-api-endpoint, agents-monorepo, community, ide skills, Claude Code, Cursor, Windsurf

v1.0.0
GitHub

About this Skill

Perfect for AI Agents needing to design and deploy machine learning API endpoints with stateless design and consistent response formats. Агент для управления рекламой

dengineproblem dengineproblem
[0]
[0]
Updated: 3/5/2026

Agent Capability Analysis

The ml-api-endpoint skill by dengineproblem is an open-source community AI agent skill for Claude Code and other IDE workflows, helping agents execute tasks with better context, repeatability, and domain-specific guidance.

Ideal Agent Persona

Perfect for AI Agents needing to design and deploy machine learning API endpoints with stateless design and consistent response formats.

Core Value

Empowers agents to create scalable and versioned ML API endpoints using FastAPI, with rigorous input validation and standardized success/error structures, facilitating seamless model updates and inference.

Capabilities Granted for ml-api-endpoint

Deploying machine learning models as RESTful APIs
Validating and standardizing input data for ML inference
Implementing versioning strategies for model updates and rollbacks

! Prerequisites & Limits

  • Requires Python 3.x and FastAPI installation
  • Needs careful planning for model updates and versioning
  • Limited to designing and deploying ML API endpoints, not model training
Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

ml-api-endpoint

Install ml-api-endpoint, an AI agent skill for AI agent workflows and automation. Works with Claude Code, Cursor, and Windsurf with one-command setup.

SKILL.md
Readonly

ML API Endpoint Expert

Expert in designing and deploying machine learning API endpoints.

Core Principles

API Design

  • Stateless Design: Each request contains all necessary information
  • Consistent Response Format: Standardize success/error structures
  • Versioning Strategy: Plan for model updates
  • Input Validation: Rigorous validation before inference

FastAPI Implementation

Basic ML Endpoint

python
1from fastapi import FastAPI, HTTPException 2from pydantic import BaseModel, validator 3import joblib 4import numpy as np 5 6app = FastAPI(title="ML Model API", version="1.0.0") 7 8model = None 9 10@app.on_event("startup") 11async def load_model(): 12 global model 13 model = joblib.load("model.pkl") 14 15class PredictionInput(BaseModel): 16 features: list[float] 17 18 @validator('features') 19 def validate_features(cls, v): 20 if len(v) != 10: 21 raise ValueError('Expected 10 features') 22 return v 23 24class PredictionResponse(BaseModel): 25 prediction: float 26 confidence: float | None = None 27 model_version: str 28 request_id: str 29 30@app.post("/predict", response_model=PredictionResponse) 31async def predict(input_data: PredictionInput): 32 features = np.array([input_data.features]) 33 prediction = model.predict(features)[0] 34 35 return PredictionResponse( 36 prediction=float(prediction), 37 model_version="v1", 38 request_id=generate_request_id() 39 )

Batch Prediction

python
1class BatchInput(BaseModel): 2 instances: list[list[float]] 3 4 @validator('instances') 5 def validate_batch_size(cls, v): 6 if len(v) > 100: 7 raise ValueError('Batch size cannot exceed 100') 8 return v 9 10@app.post("/predict/batch") 11async def batch_predict(input_data: BatchInput): 12 features = np.array(input_data.instances) 13 predictions = model.predict(features) 14 15 return { 16 "predictions": predictions.tolist(), 17 "count": len(predictions) 18 }

Performance Optimization

Model Caching

python
1class ModelCache: 2 def __init__(self, ttl_seconds=300): 3 self.cache = {} 4 self.ttl = ttl_seconds 5 6 def get(self, features): 7 key = hashlib.md5(str(features).encode()).hexdigest() 8 if key in self.cache: 9 result, timestamp = self.cache[key] 10 if time.time() - timestamp < self.ttl: 11 return result 12 return None 13 14 def set(self, features, prediction): 15 key = hashlib.md5(str(features).encode()).hexdigest() 16 self.cache[key] = (prediction, time.time())

Health Checks

python
1@app.get("/health") 2async def health_check(): 3 return { 4 "status": "healthy", 5 "model_loaded": model is not None 6 } 7 8@app.get("/metrics") 9async def get_metrics(): 10 return { 11 "requests_total": request_counter, 12 "prediction_latency_avg": avg_latency, 13 "error_rate": error_rate 14 }

Docker Deployment

dockerfile
1FROM python:3.9-slim 2 3WORKDIR /app 4COPY requirements.txt . 5RUN pip install --no-cache-dir -r requirements.txt 6 7COPY . . 8EXPOSE 8000 9 10CMD ["uvicorn", "main:app", "--host", "0.0.0.0", "--port", "8000", "--workers", "4"]

Best Practices

  • Use async/await for I/O operations
  • Validate data types, ranges, and business rules
  • Cache predictions for deterministic models
  • Handle model failures with fallback responses
  • Log predictions, latencies, and errors
  • Support multiple model versions
  • Set memory and CPU limits

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is ml-api-endpoint?

Perfect for AI Agents needing to design and deploy machine learning API endpoints with stateless design and consistent response formats. Агент для управления рекламой

How do I install ml-api-endpoint?

Run the command: npx killer-skills add dengineproblem/agents-monorepo/ml-api-endpoint. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for ml-api-endpoint?

Key use cases include: Deploying machine learning models as RESTful APIs, Validating and standardizing input data for ML inference, Implementing versioning strategies for model updates and rollbacks.

Which IDEs are compatible with ml-api-endpoint?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for ml-api-endpoint?

Requires Python 3.x and FastAPI installation. Needs careful planning for model updates and versioning. Limited to designing and deploying ML API endpoints, not model training.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add dengineproblem/agents-monorepo/ml-api-endpoint. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use ml-api-endpoint immediately in the current project.

Related Skills

Looking for an alternative to ml-api-endpoint or another community skill for your workflow? Explore these related open-source skills.

View All

widget-generator

Logo of f
f

f.k.a. Awesome ChatGPT Prompts. Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.

149.6k
0
AI

flags

Logo of vercel
vercel

flags is a Next.js feature management skill that enables developers to efficiently add or modify framework feature flags, streamlining React application development.

138.4k
0
Browser

zustand

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
AI

data-fetching

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
AI