CtrlK
BlogDocsLog inGet started
Tessl Logo

configuring-auto-scaling-policies

Configure use when you need to work with auto-scaling. This skill provides auto-scaling configuration with comprehensive guidance and automation. Trigger with phrases like "configure auto-scaling", "set up elastic scaling", or "implement scaling".

Install with Tessl CLI

npx tessl i github:jeremylongshore/claude-code-plugins-plus-skills --skill configuring-auto-scaling-policies
What are skills?

48

Quality

37%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

Optimize this skill with Tessl

npx tessl skill review --optimize ./plugins/devops/auto-scaling-configurator/skills/configuring-auto-scaling-policies/SKILL.md
SKILL.md
Review
Evals

Configuring Auto-Scaling Policies

Overview

Configure auto-scaling policies for cloud workloads across AWS Auto Scaling Groups, GCP Managed Instance Groups, Azure VMSS, and Kubernetes Horizontal Pod Autoscaler (HPA). Generate scaling configurations based on CPU, memory, request rate, or custom metrics with appropriate thresholds, cooldown periods, and scale-in protection.

Prerequisites

  • Cloud provider CLI installed and authenticated (aws, gcloud, or az)
  • For Kubernetes HPA: kubectl configured with cluster access and metrics-server deployed
  • Baseline performance data for the target workload (average CPU, memory, request rate)
  • Understanding of traffic patterns (steady, bursty, scheduled)
  • IAM permissions to create/modify scaling policies and CloudWatch/Stackdriver alarms

Instructions

  1. Identify the scaling target: EC2 Auto Scaling Group, GCP MIG, Azure VMSS, or Kubernetes Deployment
  2. Analyze current workload metrics to establish baseline utilization and peak patterns
  3. Define scaling boundaries: minimum instances/pods, maximum instances/pods, desired count
  4. Select scaling metric(s): CPU utilization, memory, request count, queue depth, or custom metrics
  5. Set target thresholds: scale-out trigger (e.g., CPU > 70%), scale-in trigger (e.g., CPU < 30%)
  6. Configure cooldown periods to prevent flapping (typically 300s scale-out, 600s scale-in)
  7. Add scale-in protection for stateful workloads or leader nodes if needed
  8. Generate the scaling policy configuration in the appropriate format (Terraform, YAML, or CLI commands)
  9. Validate by simulating load and confirming scaling events fire correctly

Output

  • Terraform HCL for AWS ASG scaling policies with CloudWatch alarms
  • Kubernetes HPA manifests (YAML) with resource or custom metric targets
  • GCP autoscaler configurations for Managed Instance Groups
  • Scaling policy JSON/YAML for Azure VMSS
  • CloudWatch or Stackdriver alarm definitions tied to scaling actions

Error Handling

ErrorCauseSolution
No scaling activity despite high loadMetric not reaching threshold or cooldown activeVerify metric source in CloudWatch/Stackdriver; check cooldown timer with describe-scaling-activities
Scaling too aggressively (flapping)Cooldown too short or threshold too sensitiveIncrease cooldown period and widen the gap between scale-out and scale-in thresholds
Max capacity reachedInstance/pod limit hit during traffic spikeRaise max_size or implement request queuing as a backpressure mechanism
HPA unable to compute replica countMetrics server not deployed or metric unavailableInstall metrics-server and verify kubectl top pods returns data
FailedScaleUp: insufficient capacityCloud provider out of capacity in selected AZ/regionAdd multiple AZs to the ASG or use mixed instance types with allocation strategy

Examples

  • "Configure an AWS ASG with target tracking at 65% CPU, min 2 / max 20 instances, and 5-minute cooldown."
  • "Create a Kubernetes HPA for a deployment that scales from 3 to 50 pods based on requests-per-second using a custom Prometheus metric."
  • "Set up scheduled scaling for a GCP MIG: scale to 10 instances at 8am UTC and back to 2 at 10pm."

Resources

Repository
jeremylongshore/claude-code-plugins-plus-skills
Last updated
Created

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.