CtrlK
BlogDocsLog inGet started
Tessl Logo

meta-results-risk-of-bias

Generates the "Risk of Bias" results section for a meta-analysis based on assessment tables and statistics. Use when the user wants to draft the risk of bias analysis text from provided data tables.

60

Quality

51%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./scientific-skills/Academic Writing/meta-results-risk-of-bias/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

75%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description is well-structured with a clear 'what' and 'when' clause, targeting a highly specific academic niche. Its main weaknesses are moderate specificity in listing concrete actions and incomplete coverage of natural trigger terms that domain users might employ (e.g., 'RoB', 'systematic review', specific assessment tools).

Suggestions

Add common trigger term variations such as 'RoB', 'systematic review', 'quality assessment', 'ROBINS-I', 'RoB 2', or 'bias domains' to improve keyword coverage.

List more specific concrete actions, e.g., 'summarizes bias domain ratings, interprets overall risk judgments, and formats results text for publication' to strengthen specificity.

DimensionReasoningScore

Specificity

It names the domain (meta-analysis risk of bias) and a primary action (generates the results section), but doesn't list multiple concrete actions beyond drafting text from data tables. It could specify more granular capabilities like summarizing bias domains, interpreting assessment scores, or formatting for publication.

2 / 3

Completeness

Clearly answers both 'what' (generates the Risk of Bias results section for a meta-analysis based on assessment tables and statistics) and 'when' (Use when the user wants to draft the risk of bias analysis text from provided data tables). The explicit 'Use when...' clause is present.

3 / 3

Trigger Term Quality

Includes relevant terms like 'Risk of Bias', 'meta-analysis', 'assessment tables', and 'data tables', which are natural for the domain. However, it misses common variations users might say such as 'RoB', 'bias assessment', 'systematic review', 'quality assessment', or specific tool names like 'RoB 2' or 'ROBINS-I'.

2 / 3

Distinctiveness Conflict Risk

This is a very specific niche — generating Risk of Bias results sections for meta-analyses. It is unlikely to conflict with other skills due to the highly specialized domain and clear trigger context involving bias assessment tables.

3 / 3

Total

10

/

12

Passed

Implementation

27%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill suffers severely from template bloat—the actual domain-specific content about Risk of Bias analysis is solid but buried under extensive generic boilerplate that adds no value. The core workflow (prompt template + formatting script) is reasonable but lacks concrete examples of input data and expected output. The duplicated sections and self-referential structure ('See Workflow above' when it's below) suggest auto-generated content that wasn't reviewed.

Suggestions

Remove all generic boilerplate sections (Validation and Safety Rules, Failure Handling, Deterministic Output Rules, Completion Checklist, Output Contract) that teach Claude things it already knows, and consolidate the duplicated 'When to Use' and validation sections.

Add a concrete example showing sample input data (a small statistics table and assessment table) and the expected output text, so Claude knows exactly what format to produce.

Show a realistic invocation of format_result.py—passing >300 words via --text flag is impractical; demonstrate file-based input or piping instead.

Fix the document structure so sections flow logically—move the core 'Risk of Bias Results Generator' content to the top and remove the generic scaffolding that precedes it.

DimensionReasoningScore

Conciseness

Extremely verbose and repetitive. The skill contains massive amounts of boilerplate ('When to Use', 'When Not to Use', 'Validation and Safety Rules', 'Failure Handling', 'Deterministic Output Rules', 'Completion Checklist') that are generic template filler Claude already knows. The actual domain-specific content (the Risk of Bias workflow) is buried under layers of redundant scaffolding. 'When to Use' appears twice with different content. Many sections restate obvious principles like 'do not fabricate results.'

1 / 3

Actionability

The core workflow section provides a reasonable prompt template with specific requirements (D1-D5 domains, >300 words, specific numbers from statistics), and there's a concrete bash command for the formatting script. However, the script usage relies on `--text` flag with generated text inline which is impractical for >300 words, no example input/output is shown, and the prompt template uses placeholders without showing what actual data looks like.

2 / 3

Workflow Clarity

The two-step workflow (Draft Text → Format Output) is clearly sequenced, and there's a validation shortcut. However, there are no validation checkpoints between steps—no guidance on verifying the drafted text meets requirements before formatting, no feedback loop if the output doesn't meet the >300 word requirement or misses domains, and no error recovery for when the formatting script fails on malformed input.

2 / 3

Progressive Disclosure

The content is a monolithic wall of text with heavily duplicated sections. There are two 'When to Use' sections, two 'Quick Validation' sections, and generic boilerplate sections that could be removed entirely. No content is split into separate reference files. The structure is confusing with an 'Implementation Details' section that says 'See Workflow above' when Workflow appears below it.

1 / 3

Total

6

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

10

/

11

Passed

Repository
aipoch/medical-research-skills
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.