CtrlK
BlogDocsLog inGet started
Tessl Logo

case-control-study-quality-assessment-nos

Clinical Research Bias Assessment - Case-Control Study (NOS) v2.3.0. Use when you need to assess the bias of a case-control study using the Newcastle-Ottawa Scale (NOS) criteria, or when evaluating the quality of a medical paper.

69

Quality

62%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./scientific-skills/Data Analysis/Case-control-study-quality-assessment-nos/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

89%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This is a well-constructed description with strong trigger terms, explicit 'Use when' guidance, and a highly distinctive niche. Its main weakness is that it doesn't enumerate the specific concrete actions it performs (e.g., scoring selection, comparability, and exposure domains of NOS), which would elevate its specificity. The second-person 'you' in 'when you need' is noted but the rubric penalizes first/second person on specificity, which is already scored at 2.

Suggestions

Add specific concrete actions the skill performs, e.g., 'Scores selection criteria, evaluates comparability of cases and controls, assesses adequacy of exposure ascertainment' to improve specificity.

Rephrase to third person voice: replace 'Use when you need to assess' with 'Use when assessing' or 'Assesses bias of case-control studies...Use when evaluating...'

DimensionReasoningScore

Specificity

The description names the domain (clinical research bias assessment) and the specific methodology (Newcastle-Ottawa Scale for case-control studies), but does not list the concrete actions performed—e.g., scoring selection criteria, evaluating comparability, assessing exposure measurement. It identifies what it is but not the specific steps it takes.

2 / 3

Completeness

The description answers both 'what' (assess bias of a case-control study using NOS criteria) and 'when' (explicitly states 'Use when you need to assess the bias of a case-control study using the Newcastle-Ottawa Scale (NOS) criteria, or when evaluating the quality of a medical paper'). The 'Use when' clause is explicit.

3 / 3

Trigger Term Quality

Includes strong natural trigger terms: 'bias assessment', 'case-control study', 'Newcastle-Ottawa Scale', 'NOS', 'quality of a medical paper'. These are terms a researcher would naturally use when seeking this type of evaluation. Good coverage of both the acronym and full name.

3 / 3

Distinctiveness Conflict Risk

Highly distinctive—it targets a very specific niche (case-control study bias assessment using NOS). The combination of study type (case-control), methodology (NOS), and domain (clinical research) makes it very unlikely to conflict with other skills. The version number further distinguishes it.

3 / 3

Total

11

/

12

Passed

Implementation

35%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill suffers significantly from boilerplate bloat—many sections contain generic, template-driven content that adds no NOS-specific value and wastes tokens. The core NOS workflow (Steps 1-4) is reasonably structured but lacks concrete examples of scored criteria, expected JSON schemas, and validation checkpoints. The skill would benefit greatly from removing redundant generic sections and adding specific NOS scoring examples and output format definitions.

Suggestions

Remove or drastically condense the generic boilerplate sections ('When to Use', 'When Not to Use', 'Required Inputs', 'Validation and Safety Rules', 'Failure Handling', 'Output Contract', 'Recommended Workflow') which contain no NOS-specific information and duplicate each other.

Add a concrete example of the expected JSON output schema showing what a scored NOS assessment looks like (e.g., selection stars, comparability stars, exposure stars with reasons).

Complete the empty 'Helper Scripts / PDF Text Extraction' section with actual usage instructions or remove it.

Add explicit validation checkpoints in the detailed workflow, such as verifying extracted text quality before assessment and validating JSON structure before formatting.

DimensionReasoningScore

Conciseness

The skill is extremely verbose and repetitive. There are multiple redundant sections (e.g., 'When to Use', 'When Not to Use', 'Required Inputs', 'Recommended Workflow', 'Output Contract', 'Validation and Safety Rules', 'Failure Handling') that contain generic boilerplate applicable to any skill rather than NOS-specific guidance. The 'Key Features' section restates the description verbatim. 'Example Usage' references a '## Usage' section that appears later, and the 'Recommended Workflow' duplicates the 'Detailed Workflow'. Much of this content is filler that Claude already knows how to do.

1 / 3

Actionability

There are some concrete elements—bash commands for running scripts, references to specific files like `nos_criteria_prompts.md`, and a JSON output format mention—but the actual NOS assessment criteria are not spelled out in the skill itself (delegated to a reference file). The detailed workflow steps (Selection, Comparability, Exposure) are described at a high level without concrete examples of what a scored item looks like or example JSON output schemas. The script invocations are concrete but incomplete (e.g., the PDF extraction section is empty).

2 / 3

Workflow Clarity

The detailed workflow has a clear 4-step sequence (Selection → Comparability → Exposure → Generate Summary Table), which is good. However, there are no explicit validation checkpoints between steps, no error recovery loops, and the 'Helper Scripts' section for PDF extraction is incomplete (empty). The 'Recommended Workflow' section duplicates and conflicts with the 'Detailed Workflow', creating confusion about which sequence to follow.

2 / 3

Progressive Disclosure

There is a reference to `references/nos_criteria_prompts.md` which is appropriate one-level-deep disclosure. However, the main file itself is bloated with generic boilerplate sections that should either be removed or consolidated. The structure has too many sections that repeat similar information, making navigation difficult. The reference to helper scripts is incomplete.

2 / 3

Total

7

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

10

/

11

Passed

Repository
aipoch/medical-research-skills
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.