Statistical models library for Python. Use when you need specific model classes (OLS, GLM, mixed models, ARIMA) with detailed diagnostics, residuals, and inference. Best for econometrics, time series, rigorous inference with coefficient tables. For guided statistical test selection with APA reporting use statistical-analysis.
85
75%
Does it follow best practices?
Impact
93%
1.09xAverage score across 6 eval scenarios
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./scientific-skills/statsmodels/SKILL.mdQuality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that clearly communicates specific capabilities (model classes, diagnostics, inference), includes rich natural trigger terms that users in statistics would use, and explicitly addresses both when to use this skill and when to use a related alternative. The disambiguation with the statistical-analysis skill is a notable strength that reduces conflict risk.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions and model classes: OLS, GLM, mixed models, ARIMA, along with specific outputs like detailed diagnostics, residuals, inference, and coefficient tables. | 3 / 3 |
Completeness | Clearly answers both what ('Statistical models library for Python' with specific model classes and outputs) and when ('Use when you need specific model classes... Best for econometrics, time series, rigorous inference'). Also includes explicit disambiguation guidance pointing to statistical-analysis for a different use case. | 3 / 3 |
Trigger Term Quality | Includes strong natural keywords users would say: 'OLS', 'GLM', 'mixed models', 'ARIMA', 'econometrics', 'time series', 'inference', 'coefficient tables', 'residuals', 'diagnostics'. These are terms a user working in statistics would naturally use. | 3 / 3 |
Distinctiveness Conflict Risk | Clearly distinguishes itself from the related 'statistical-analysis' skill by specifying its niche (specific model classes, diagnostics, econometrics) versus guided test selection with APA reporting. The explicit disambiguation clause reduces conflict risk significantly. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
50%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
The skill provides excellent actionable code examples that are executable and well-structured, but is severely undermined by verbosity. Large sections catalog statsmodels capabilities (model types, distribution families, test names) that Claude already knows, wasting significant token budget. The content would benefit greatly from moving the reference-catalog material into the referenced files and keeping SKILL.md focused on the quick-start examples, key gotchas, and workflow summaries.
Suggestions
Remove or drastically condense the 'Core Statistical Modeling Capabilities' section — move detailed model/test listings into the referenced files (references/*.md) and keep only a brief table or one-liner per category in SKILL.md.
Trim the 'When to Use This Skill' section to 2-3 lines; Claude doesn't need a bullet list of every possible use case.
Add explicit validation checkpoints to the Common Workflows (e.g., 'If diagnostics fail, go back to step N' or 'Only proceed to interpretation after residual checks pass').
Reduce 'Common Pitfalls' from 15 items to the 5-6 most non-obvious ones that Claude wouldn't already know from its training data.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Extremely verbose at ~400+ lines. Extensively lists model types, features, and capabilities that Claude already knows (e.g., what OLS is, what Poisson regression does, what AIC/BIC are). The 'Core Statistical Modeling Capabilities' section is essentially a catalog of statsmodels features that adds little actionable value. The 'Common Pitfalls' list has 15 items, many of which are basic statistical knowledge. The 'When to Use This Skill' section and repeated 'When to use' subsections are redundant with the skill description. | 1 / 3 |
Actionability | The Quick Start Guide provides fully executable, copy-paste ready code examples for OLS, Logistic Regression, ARIMA, and GLM. The Formula API section, Model Selection code, and Cross-Validation examples are all concrete and executable with proper imports and complete code blocks. | 3 / 3 |
Workflow Clarity | The four common workflows (Linear Regression, Binary Classification, Count Data, Time Series) list clear sequential steps, but they are described as numbered prose without validation checkpoints or explicit feedback loops. The GLM quick start example does show a conditional check for overdispersion, which is good, but the formal workflow sections lack explicit 'validate before proceeding' gates. | 2 / 3 |
Progressive Disclosure | References to five separate reference files are well-organized and clearly signaled with descriptions and search patterns. However, no bundle files were provided, so these references may not exist. More importantly, the SKILL.md itself contains enormous amounts of content that should be in those reference files (the entire 'Core Statistical Modeling Capabilities' section with its detailed model listings), making the main file a near-monolithic document despite having a reference structure. | 2 / 3 |
Total | 8 / 12 Passed |
Validation
81%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 9 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
skill_md_line_count | SKILL.md is long (613 lines); consider splitting into references/ and linking | Warning |
metadata_version | 'metadata.version' is missing | Warning |
Total | 9 / 11 Passed | |
cbcae7b
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.