Inference Latency Profiler - Auto-activating skill for ML Deployment. Triggers on: inference latency profiler, inference latency profiler Part of the ML Deployment skill category.
35
3%
Does it follow best practices?
Impact
92%
1.01xAverage score across 3 eval scenarios
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./planned-skills/generated/08-ml-deployment/inference-latency-profiler/SKILL.mdProduction-ready latency load testing script
Standard latency tool
100%
100%
p50 measurement
100%
100%
p95 measurement
100%
100%
p99 measurement
100%
100%
Error/timeout handling
100%
100%
Parameterizable CLI
100%
100%
Structured results file
100%
100%
Threshold validation
0%
0%
Step-by-step structure
80%
90%
No hardcoded secrets
100%
100%
Explanatory comments
42%
42%
MLOps inference monitoring configuration
Standard monitoring tool
100%
100%
Latency histogram or summary
100%
100%
p95 or p99 alert
100%
100%
Multi-stage pipeline coverage
100%
100%
Configuration file format
100%
100%
Alerting threshold defined
100%
100%
Standard metric naming
100%
100%
Component documentation
100%
100%
Model version or env labels
100%
100%
No deprecated metric patterns
100%
100%
Step-by-step bottleneck identification and optimization
Standard profiling approach
100%
100%
Per-stage latency
100%
100%
Latency percentiles
100%
100%
Bottleneck identified
100%
100%
Production-safe code
70%
70%
Structured results file
100%
100%
Optimization recommendations
100%
100%
Latency in milliseconds
50%
50%
Step-by-step structure
100%
100%
Spike detection
100%
100%
87f14eb
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.