CtrlK
BlogDocsLog inGet started
Tessl Logo

paper-illustration

Generate publication-quality AI illustrations for academic papers using Gemini image generation. Creates architecture diagrams, method illustrations with Claude-supervised iterative refinement loop. Use when user says "生成图表", "画架构图", "AI绘图", "paper illustration", "generate diagram", or needs visual figures for papers.

73

Quality

70%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Advisory

Suggest reviewing before use

Optimize this skill with Tessl

npx tessl skill review --optimize ./skills/paper-illustration/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

100%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This is a strong skill description that clearly communicates its purpose, lists concrete capabilities, and provides explicit trigger guidance in both Chinese and English. The description is concise yet comprehensive, with a well-defined niche that distinguishes it from general image generation or diagramming skills. The inclusion of the specific tool (Gemini) and methodology (Claude-supervised iterative refinement) adds valuable specificity.

DimensionReasoningScore

Specificity

Lists multiple specific concrete actions: 'Generate publication-quality AI illustrations', 'Creates architecture diagrams', 'method illustrations with Claude-supervised iterative refinement loop'. These are concrete, domain-specific capabilities.

3 / 3

Completeness

Clearly answers both 'what' (generate AI illustrations for academic papers using Gemini, creates architecture diagrams with iterative refinement) and 'when' (explicit 'Use when' clause with specific trigger phrases and a general condition about needing visual figures).

3 / 3

Trigger Term Quality

Includes strong natural trigger terms in both Chinese and English: '生成图表', '画架构图', 'AI绘图', 'paper illustration', 'generate diagram', 'visual figures for papers'. Good coverage of terms users would naturally say.

3 / 3

Distinctiveness Conflict Risk

Highly distinctive niche: academic paper illustrations using Gemini image generation with Claude-supervised refinement. The combination of academic context, specific tool (Gemini), and bilingual triggers makes it very unlikely to conflict with other skills.

3 / 3

Total

12

/

12

Passed

Implementation

39%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

The skill demonstrates excellent workflow design with a clear multi-stage iterative process, explicit validation checkpoints, and well-defined acceptance criteria. However, it is severely undermined by extreme verbosity — style guidelines and visual requirements are repeated 3-4 times across different sections, inflating the document to an unnecessary length. The bash scripts, while concrete, have interpolation issues that would prevent correct execution, and the entire document should be split across multiple files.

Suggestions

Eliminate redundancy by defining the CVPR style guide, color palette, and visual dos/don'ts exactly ONCE, then referencing that single definition in the prompt template and review checklist rather than repeating it verbatim.

Extract the style guide into a separate STYLE_GUIDE.md, the review checklist template into REVIEW_TEMPLATE.md, and the bash scripts into actual executable files (e.g., scripts/layout_optimize.sh, scripts/render.sh) referenced from the main skill.

Fix the bash scripts' variable interpolation — shell variables like $STYLE_SPEC inside Python heredocs and JSON payloads need proper escaping or should use temporary files consistently. Show the actual mechanism for Claude to inject dynamic content into the prompt.

Remove explanatory content Claude already knows (e.g., what CVPR style means, why arrows should be thick, what sans-serif fonts are) and reduce to a concise specification table of requirements.

DimensionReasoningScore

Conciseness

Extremely verbose at ~500+ lines. Massive repetition of style guidelines (the CVPR style guide, visual appeal dos/don'ts, and color palettes are repeated nearly verbatim in the style guide section, the prompt template, the style verification step, and the review checklist). The ASCII workflow diagram, while informative, is oversized. Claude doesn't need explanations of what CVPR style means or why arrows should be thick — it needs the spec once.

1 / 3

Actionability

Contains executable bash scripts with curl commands and Python extraction code, which is good. However, the scripts use shell variable interpolation inside heredocs in ways that won't work correctly (e.g., $STYLE_SPEC inside bash strings passed to curl), and the prompt template has placeholder brackets like '[Claude fills in the detailed prompt here]' rather than showing how Claude should programmatically construct and inject the prompt. The iteration counter is hardcoded to 1 with a comment 'Claude increments this' but no mechanism shown.

2 / 3

Workflow Clarity

The multi-step workflow is exceptionally clear with numbered steps (0-8), explicit decision points (score ≥ 9 check), a feedback loop for refinement, maximum iteration limits, and a detailed review checklist with specific scoring thresholds. The validation step (Step 5) is thorough with explicit pass/fail criteria and the loop-back mechanism is well-defined.

3 / 3

Progressive Disclosure

Everything is crammed into a single monolithic file with no references to external files. The CVPR style guide, prompt templates, review checklists, bash scripts, and scoring rubrics are all inline. The style guide alone could be a separate STYLE_GUIDE.md, the review template could be REVIEW_TEMPLATE.md, and the bash scripts could be referenced as actual script files. The content is a wall of text despite having headers.

1 / 3

Total

7

/

12

Passed

Validation

72%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation8 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

skill_md_line_count

SKILL.md is long (693 lines); consider splitting into references/ and linking

Warning

allowed_tools_field

'allowed-tools' contains unusual tool name(s)

Warning

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

8

/

11

Passed

Repository
wanshuiyin/Auto-claude-code-research-in-sleep
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.