Generate publication-quality AI illustrations for academic papers using Gemini image generation. Creates architecture diagrams, method illustrations with Claude-supervised iterative refinement loop. Use when user says "生成图表", "画架构图", "AI绘图", "paper illustration", "generate diagram", or needs visual figures for papers.
73
70%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Advisory
Suggest reviewing before use
Optimize this skill with Tessl
npx tessl skill review --optimize ./skills/paper-illustration/SKILL.mdQuality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is a strong skill description that clearly communicates its purpose, lists concrete capabilities, and provides explicit trigger guidance in both Chinese and English. The description is concise yet comprehensive, with a well-defined niche that distinguishes it from general image generation or diagramming skills. The inclusion of the specific tool (Gemini) and methodology (Claude-supervised iterative refinement) adds valuable specificity.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'Generate publication-quality AI illustrations', 'Creates architecture diagrams', 'method illustrations with Claude-supervised iterative refinement loop'. These are concrete, domain-specific capabilities. | 3 / 3 |
Completeness | Clearly answers both 'what' (generate AI illustrations for academic papers using Gemini, creates architecture diagrams with iterative refinement) and 'when' (explicit 'Use when' clause with specific trigger phrases and a general condition about needing visual figures). | 3 / 3 |
Trigger Term Quality | Includes strong natural trigger terms in both Chinese and English: '生成图表', '画架构图', 'AI绘图', 'paper illustration', 'generate diagram', 'visual figures for papers'. Good coverage of terms users would naturally say. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive niche: academic paper illustrations using Gemini image generation with Claude-supervised refinement. The combination of academic context, specific tool (Gemini), and bilingual triggers makes it very unlikely to conflict with other skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
39%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
The skill demonstrates excellent workflow design with a clear multi-stage iterative process, explicit validation checkpoints, and well-defined acceptance criteria. However, it is severely undermined by extreme verbosity — style guidelines and visual requirements are repeated 3-4 times across different sections, inflating the document to an unnecessary length. The bash scripts, while concrete, have interpolation issues that would prevent correct execution, and the entire document should be split across multiple files.
Suggestions
Eliminate redundancy by defining the CVPR style guide, color palette, and visual dos/don'ts exactly ONCE, then referencing that single definition in the prompt template and review checklist rather than repeating it verbatim.
Extract the style guide into a separate STYLE_GUIDE.md, the review checklist template into REVIEW_TEMPLATE.md, and the bash scripts into actual executable files (e.g., scripts/layout_optimize.sh, scripts/render.sh) referenced from the main skill.
Fix the bash scripts' variable interpolation — shell variables like $STYLE_SPEC inside Python heredocs and JSON payloads need proper escaping or should use temporary files consistently. Show the actual mechanism for Claude to inject dynamic content into the prompt.
Remove explanatory content Claude already knows (e.g., what CVPR style means, why arrows should be thick, what sans-serif fonts are) and reduce to a concise specification table of requirements.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Extremely verbose at ~500+ lines. Massive repetition of style guidelines (the CVPR style guide, visual appeal dos/don'ts, and color palettes are repeated nearly verbatim in the style guide section, the prompt template, the style verification step, and the review checklist). The ASCII workflow diagram, while informative, is oversized. Claude doesn't need explanations of what CVPR style means or why arrows should be thick — it needs the spec once. | 1 / 3 |
Actionability | Contains executable bash scripts with curl commands and Python extraction code, which is good. However, the scripts use shell variable interpolation inside heredocs in ways that won't work correctly (e.g., $STYLE_SPEC inside bash strings passed to curl), and the prompt template has placeholder brackets like '[Claude fills in the detailed prompt here]' rather than showing how Claude should programmatically construct and inject the prompt. The iteration counter is hardcoded to 1 with a comment 'Claude increments this' but no mechanism shown. | 2 / 3 |
Workflow Clarity | The multi-step workflow is exceptionally clear with numbered steps (0-8), explicit decision points (score ≥ 9 check), a feedback loop for refinement, maximum iteration limits, and a detailed review checklist with specific scoring thresholds. The validation step (Step 5) is thorough with explicit pass/fail criteria and the loop-back mechanism is well-defined. | 3 / 3 |
Progressive Disclosure | Everything is crammed into a single monolithic file with no references to external files. The CVPR style guide, prompt templates, review checklists, bash scripts, and scoring rubrics are all inline. The style guide alone could be a separate STYLE_GUIDE.md, the review template could be REVIEW_TEMPLATE.md, and the bash scripts could be referenced as actual script files. The content is a wall of text despite having headers. | 1 / 3 |
Total | 7 / 12 Passed |
Validation
72%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 8 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
skill_md_line_count | SKILL.md is long (693 lines); consider splitting into references/ and linking | Warning |
allowed_tools_field | 'allowed-tools' contains unusual tool name(s) | Warning |
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 8 / 11 Passed | |
700fbe2
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.