Deep generative models for single-cell omics. Use when you need probabilistic batch correction (scVI), transfer learning, differential expression with uncertainty, or multi-modal integration (TOTALVI, MultiVI). Best for advanced modeling, batch effects, multimodal data. For standard analysis pipelines use scanpy.
91
82%
Does it follow best practices?
Impact
97%
1.14xAverage score across 6 eval scenarios
Passed
No known issues
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that clearly defines a specialized niche (deep generative models for single-cell omics), lists specific concrete capabilities with tool names, provides explicit 'Use when' triggers, and even includes a boundary condition directing users to an alternative skill (scanpy) for standard pipelines. The description is concise, uses third-person voice, and contains highly relevant domain-specific trigger terms.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: probabilistic batch correction (scVI), transfer learning, differential expression with uncertainty, and multi-modal integration (TOTALVI, MultiVI). Also distinguishes from standard analysis pipelines. | 3 / 3 |
Completeness | Clearly answers both 'what' (deep generative models for single-cell omics with specific capabilities listed) and 'when' (explicit 'Use when' clause with trigger scenarios, plus a 'Best for' clause and a negative boundary directing users to scanpy for standard pipelines). | 3 / 3 |
Trigger Term Quality | Includes strong natural keywords users would say: 'batch correction', 'scVI', 'transfer learning', 'differential expression', 'TOTALVI', 'MultiVI', 'multimodal', 'single-cell omics'. These are terms domain experts would naturally use when seeking this functionality. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive with a clear niche in deep generative modeling for single-cell omics. Explicitly differentiates from scanpy for standard analysis, reducing conflict risk. Tool-specific names (scVI, TOTALVI, MultiVI) make it very unlikely to trigger incorrectly. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
64%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a solid skill with excellent actionability — the code examples are complete, executable, and well-structured. The main weaknesses are moderate verbosity (explaining concepts Claude knows, overly detailed 'When to Use' section, theoretical foundations section), and missing validation/verification steps in the workflow. The progressive disclosure structure is well-designed in principle but the main file retains too much content that should live in the referenced files.
Suggestions
Remove the 'Theoretical Foundations' section entirely — Claude already understands VAEs, variational inference, and amortized inference.
Add validation checkpoints to the workflow: check training convergence via `model.history['elbo_train']`, verify latent representation quality with a quick UMAP visualization, and mention common failure modes (e.g., NaN loss from non-count data).
Trim the 'When to Use This Skill' section to 2-3 key differentiators vs scanpy rather than listing every possible use case.
Move the detailed model catalog (sections 1-5 under Core Capabilities) into a single reference file and keep only a brief summary table in the main skill.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill includes some unnecessary content like the 'Theoretical Foundations' section explaining variational inference and VAEs (concepts Claude already knows), the 'When to Use This Skill' section is overly exhaustive, and the overview restates what Claude would already understand about PyTorch/Lightning. However, the code examples and model listings are reasonably efficient. | 2 / 3 |
Actionability | The skill provides fully executable, copy-paste ready code examples for the typical workflow, differential expression, model persistence, and batch correction. The code includes specific function calls, parameter names, and realistic values that Claude can directly use. | 3 / 3 |
Workflow Clarity | The typical workflow is clearly sequenced with numbered steps and a complete code example. However, there are no validation checkpoints — no guidance on checking training convergence (e.g., inspecting ELBO loss), no verification that latent representations are reasonable, and no error recovery steps for common failure modes like GPU memory issues or data format problems. | 2 / 3 |
Progressive Disclosure | The skill references multiple files in a `references/` directory (models-scrna-seq.md, differential-expression.md, workflows.md, etc.) with clear signaling, which is good structure. However, no bundle files were provided, so we cannot verify these references exist. Additionally, the main file itself is quite long (~170 lines) with sections like 'Theoretical Foundations' and the exhaustive model catalog that could be moved to reference files to keep the overview leaner. | 2 / 3 |
Total | 9 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
metadata_version | 'metadata.version' is missing | Warning |
Total | 10 / 11 Passed | |
75c41d7
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.