Deploy ML models with FastAPI, Docker, Kubernetes. Use for serving predictions, containerization, monitoring, drift detection, or encountering latency issues, health check failures, version conflicts.
94
92%
Does it follow best practices?
Impact
100%
1.75xAverage score across 3 eval scenarios
Passed
No known issues
FastAPI ML server setup
Uses FastAPI
100%
100%
Uses joblib
100%
100%
Pydantic request model
100%
100%
Finite value validator
50%
100%
Liveness endpoint
50%
100%
Readiness endpoint
50%
100%
Batch prediction endpoint
100%
100%
model_version in response
100%
100%
latency_ms in response
0%
100%
Request logging middleware
0%
100%
X-Process-Time header
0%
100%
X-Model-Version header
0%
100%
Docker containerization
Base image python:3.11-slim
0%
100%
Multi-stage build
0%
100%
Non-root user
100%
100%
MODEL_PATH env var
100%
100%
Uvicorn CMD on port 8000
100%
100%
HEALTHCHECK directive
100%
100%
Pinned requirements
100%
100%
Versioned image tag
100%
100%
Build arg for model version
0%
100%
pip --no-cache-dir
100%
100%
WORKDIR /app
100%
100%
EXPOSE 8000
100%
100%
Kubernetes deployment and monitoring
Memory request 512Mi
0%
100%
Memory limit 1Gi
0%
100%
CPU request 500m
0%
100%
CPU limit 1000m
100%
100%
Liveness probe /health
100%
100%
Readiness probe /ready
0%
100%
Rolling update maxUnavailable 0
0%
100%
maxSurge: 1
100%
100%
KS test for input drift
100%
100%
Jensen-Shannon for prediction drift
0%
100%
Rolling window deque
100%
100%
should_retrain method
100%
100%
Periodic drift check
0%
100%
90d6bd7
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.