Use right after cub-apply returns, or any time the user asks "did it actually deploy?", "is it live?", "is it still applying?", "did argo pick it up?", "are ConfigHub and the cluster in sync?", "prove this converged", "close this release out", "show me what changed". Single skill covering the whole post-apply arc: read Unit status + latest event to classify the apply (Progressing / Completed / Failed / Aborted), drill into `cub unit-event get` for per-resource sync/ready status and the Message field when something broke, cross-check the controller (Argo/Flux) and cluster (kubectl) for runtime failures (ImagePullBackOff, CrashLoopBackOff, schema errors), optionally produce a three-way ConfigHub ↔ controller ↔ cluster agreement table, and on success surface the Revision history with its --change-desc + GUI review links. Do not load for pure ConfigHub-internal query (use cub-query), for reconciling known drift (use drift-reconcile), or when the apply has not actually run yet (use cub-apply).
89
88%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that thoroughly covers specific capabilities, includes abundant natural trigger phrases users would actually say, clearly delineates both what the skill does and when to use it, and explicitly defines boundaries against related skills to prevent conflicts. The only minor concern is its length, which borders on verbose, but the density of useful information justifies it. The description uses proper third-person voice throughout.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: read Unit status + latest event, classify apply states (Progressing/Completed/Failed/Aborted), drill into per-resource sync/ready status, cross-check controller and cluster for runtime failures (ImagePullBackOff, CrashLoopBackOff, schema errors), produce three-way agreement table, surface Revision history with change-desc and GUI review links. | 3 / 3 |
Completeness | Clearly answers both 'what' (classify apply status, drill into per-resource sync, cross-check controller/cluster, produce agreement table, surface revision history) and 'when' (right after cub-apply returns, or when user asks specific trigger questions). Also explicitly states when NOT to use it with three exclusion cases, which further strengthens the 'when' guidance. | 3 / 3 |
Trigger Term Quality | Excellent coverage of natural trigger phrases users would actually say: 'did it actually deploy?', 'is it live?', 'is it still applying?', 'did argo pick it up?', 'are ConfigHub and the cluster in sync?', 'prove this converged', 'close this release out', 'show me what changed'. Also includes technical terms like ImagePullBackOff, CrashLoopBackOff, Argo, Flux, kubectl. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive with explicit negative boundaries ('Do not load for pure ConfigHub-internal query (use cub-query), for reconciling known drift (use drift-reconcile), or when the apply has not actually run yet (use cub-apply)'). The post-apply verification niche is clearly carved out and unlikely to conflict with other skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
77%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a high-quality, operationally sophisticated skill that covers a complex post-apply verification workflow with excellent actionability and workflow clarity. The branching logic, preflight gates, stop conditions, and explicit tool boundaries make it safe and reliable. The main weakness is moderate verbosity — some sections are slightly redundant, and the inline detail (especially the three-way table and status taxonomy) could potentially be split into reference files for better progressive disclosure.
Suggestions
Consider moving the three-way agreement table template and the apply-status taxonomy into a reference file (e.g., references/verify-apply-details.md) to reduce the main skill's length while preserving discoverability.
Consolidate the 'What this answers' and 'When to use' sections — they overlap significantly and could be merged into a single trigger/scope section.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is thorough and mostly earns its length given the complexity of the multi-step verification workflow, but there's some redundancy (e.g., the 'What this answers' section partially restates the 'When to use' section, and the Evidence section largely repeats commands already shown). The taxonomy table and status values are valuable additions, though some explanatory prose could be tightened. | 2 / 3 |
Actionability | Excellent actionability throughout — every step includes exact CLI commands with flags, specific jq selectors, concrete kubectl/argocd/flux commands, and clear branching logic. The status taxonomy table maps exact field values to meanings, and error patterns name specific error strings with remediation routes. | 3 / 3 |
Workflow Clarity | The five-step loop is clearly sequenced with explicit branching (Completed → step 4/5, Progressing → step 2, Failed → step 3, Aborted → surface and ask). Validation checkpoints are built in: preflight gates before starting, three-way agreement table as a verification step, close-out preflight that blocks if any Unit is still non-terminal. Stop conditions and tool boundaries are explicit, preventing destructive actions. | 3 / 3 |
Progressive Disclosure | The skill references several companion files (references/cub-cli.md, references/filters-and-queries.md, references/revisions.md) and companion skills, which is good structure. However, no bundle files were provided to verify these exist, and the skill itself is quite long (~200+ lines) with the full three-way agreement table and detailed taxonomy inline rather than in a reference file. The inline detail is arguably justified by the workflow nature, but the agreement table and status taxonomy could be split out. | 2 / 3 |
Total | 10 / 12 Passed |
Validation
81%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 9 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
allowed_tools_field | 'allowed-tools' contains unusual tool name(s) | Warning |
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 9 / 11 Passed | |
59ea831
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.