CtrlK
BlogDocsLog inGet started
Tessl Logo

tdg-personal/continuous-agent-loop

Patterns for continuous autonomous agent loops with quality gates, evals, and recovery controls.

34

Quality

34%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Overview
Quality
Evals
Security
Files

Quality

Discovery

22%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description is too abstract and pattern-oriented, reading more like a topic heading than an actionable skill description. It lacks concrete actions, explicit trigger guidance ('Use when...'), and natural user-facing keywords. While the domain of autonomous agent loops provides some distinctiveness, the vague language significantly limits its effectiveness for skill selection.

Suggestions

Add a 'Use when...' clause with explicit triggers, e.g., 'Use when building agentic loops, implementing self-correcting workflows, or setting up autonomous task execution with stop conditions.'

Replace abstract terms like 'patterns' with concrete actions, e.g., 'Implements continuous agent loops with evaluation checkpoints, automatic error recovery, and configurable termination conditions.'

Include natural user keywords such as 'agentic workflow', 'loop until complete', 'self-healing agent', 'retry logic', or 'autonomous execution' to improve trigger term coverage.

DimensionReasoningScore

Specificity

The description uses abstract language like 'patterns', 'quality gates', 'evals', and 'recovery controls' without listing concrete actions. It describes what the skill contains conceptually but not what specific actions it enables (e.g., 'implement retry logic', 'set up evaluation checkpoints', 'configure loop termination conditions').

1 / 3

Completeness

The description partially addresses 'what' (patterns for agent loops) but is vague, and completely lacks a 'when' clause or any explicit trigger guidance. Per the rubric, a missing 'Use when...' clause caps completeness at 2, and the weak 'what' brings it down to 1.

1 / 3

Trigger Term Quality

Terms like 'autonomous agent loops', 'quality gates', and 'evals' are somewhat relevant but lean toward technical jargon. Missing natural user phrases like 'agent loop', 'agentic workflow', 'self-correcting agent', 'loop until done', or 'continuous execution'. Some terms like 'recovery controls' are niche.

2 / 3

Distinctiveness Conflict Risk

The domain of 'continuous autonomous agent loops' is somewhat specific and narrows the scope, but terms like 'evals' and 'quality gates' could overlap with testing, CI/CD, or evaluation-focused skills. The lack of explicit triggers increases conflict risk.

2 / 3

Total

6

/

12

Passed

Implementation

14%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill reads as a high-level outline or table of contents rather than an actionable skill. It names many tools and patterns but never provides concrete instructions, executable examples, or navigable references. The content would leave Claude unable to actually implement a continuous agent loop without significant external knowledge.

Suggestions

Add concrete, executable examples for at least the default 'sequential' loop pattern, showing actual code or commands Claude should run.

Provide explicit links or file paths for all referenced skills/tools (e.g., `ralphinho-rfc-pipeline`, `plankton-code-quality`) so Claude can navigate to them.

Add validation checkpoints to the Recovery section with specific criteria (e.g., 'if test pass rate < 80% after 3 iterations, freeze loop') and a feedback loop structure.

Define what each loop type (continuous-pr, rfc-dag, infinite, sequential) actually does with at least a 2-3 line description and a minimal working example.

DimensionReasoningScore

Conciseness

Reasonably brief and doesn't over-explain concepts, but some content is vague filler (e.g., 'This is the v1.8+ canonical loop skill name' and compatibility notes) that doesn't add actionable value. The failure modes list is generic.

2 / 3

Actionability

Almost entirely abstract and descriptive. No executable code, no concrete commands, no specific examples. References like `ralphinho-rfc-pipeline`, `/quality-gate`, `/harness-audit` are named but never defined or demonstrated. 'Freeze loop' and 'reduce scope to failing unit' are vague directions, not actionable instructions.

1 / 3

Workflow Clarity

The loop selection flow is a decision tree but lacks any detail on what each path actually entails. The 'Combined Pattern' is just a list of tool names without sequencing or validation checkpoints. The 'Recovery' section lists steps but has no validation, feedback loops, or concrete criteria for when to trigger recovery.

1 / 3

Progressive Disclosure

References several other skills/tools by name (e.g., `ralphinho-rfc-pipeline`, `plankton-code-quality`, `eval-harness`, `nanoclaw-repl`) but provides no links, file paths, or any way to navigate to them. There's no clear structure separating overview from detailed content.

1 / 3

Total

5

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

10

/

11

Passed

Reviewed

Table of Contents