Verify research idea novelty against recent literature. Use when user says "查新", "novelty check", "有没有人做过", "check novelty", or wants to verify a research idea is novel before implementing.
74
70%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Advisory
Suggest reviewing before use
Optimize this skill with Tessl
npx tessl skill review --optimize ./skills/novelty-check/SKILL.mdQuality
Discovery
89%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is a well-crafted description with strong trigger terms (including bilingual coverage), a clear 'Use when' clause, and a distinct niche. Its main weakness is that the 'what' portion could be more specific about the concrete actions performed (e.g., searching databases, comparing approaches, summarizing related work). Overall, it would perform well in skill selection among a large pool of skills.
Suggestions
Expand the capability description with more specific actions, e.g., 'Searches recent papers, compares methodologies, and summarizes overlapping work to verify research idea novelty against recent literature.'
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | The description names the domain (research novelty verification) and one core action (verify research idea novelty against recent literature), but doesn't list multiple specific concrete actions like searching databases, comparing methodologies, generating novelty reports, etc. | 2 / 3 |
Completeness | Clearly answers both 'what' (verify research idea novelty against recent literature) and 'when' (explicit 'Use when' clause with specific trigger phrases and the scenario of wanting to verify novelty before implementing). | 3 / 3 |
Trigger Term Quality | Excellent coverage of natural trigger terms including both Chinese ('查新', '有没有人做过') and English ('novelty check', 'check novelty') variations, plus the natural phrasing 'verify a research idea is novel before implementing'. These are terms users would naturally say. | 3 / 3 |
Distinctiveness Conflict Risk | Very distinct niche — research novelty checking is a specific task unlikely to conflict with general literature review, code review, or other research skills. The bilingual trigger terms and specific use case make it clearly distinguishable. | 3 / 3 |
Total | 11 / 12 Passed |
Implementation
50%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill provides a well-structured four-phase workflow for novelty checking with a useful output template, but falls short on actionability (lacking executable examples for MCP calls and search queries) and workflow clarity (no validation checkpoints or feedback loops for handling ambiguous results). The content is reasonably concise but could be tightened in places, and the progressive disclosure is adequate but the inline report template adds bulk.
Suggestions
Add a concrete, copy-paste ready example of the mcp__codex__codex invocation in Phase C, including the full prompt structure and expected response format.
Provide specific example search queries for Phase B (e.g., 'arXiv search: "contrastive learning" AND "graph neural network" site:arxiv.org 2024..2026') to make the search step more actionable.
Add explicit validation checkpoints between phases, e.g., 'If fewer than 5 relevant papers found in Phase B, broaden search terms and retry before proceeding to Phase C.'
Move the output report template to a separate reference file (e.g., REPORT_TEMPLATE.md) and link to it, keeping the main skill leaner.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is mostly efficient but includes some unnecessary elaboration. The 'Important Rules' section contains useful but somewhat verbose guidance, and some phrasing could be tightened (e.g., explaining what 'applying X to Y' means). The constants section and phase descriptions are reasonably lean. | 2 / 3 |
Actionability | The skill provides a clear structured workflow with specific phases and a concrete output template, but lacks executable code examples. The MCP call in Phase C shows only a config snippet and a vague prompt description rather than a copy-paste ready invocation. The search instructions in Phase B are directional rather than concrete (e.g., no actual query templates or specific tool invocation syntax). | 2 / 3 |
Workflow Clarity | The four phases (A-D) provide a clear sequence, but there are no validation checkpoints or feedback loops. If the literature search yields ambiguous results, there's no guidance on how to iterate. If the cross-model verification disagrees with the search findings, there's no error recovery step. For a research verification workflow where false positives/negatives have significant consequences, this is a notable gap. | 2 / 3 |
Progressive Disclosure | The skill references external files (shared-references/review-tracing.md, tools/save_trace.sh) which is good progressive disclosure, but the main content is somewhat monolithic — the output template and important rules could potentially be separated. The references are one-level deep and clearly signaled, but the inline report template is lengthy and could be in a separate file. | 2 / 3 |
Total | 8 / 12 Passed |
Validation
81%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 9 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
allowed_tools_field | 'allowed-tools' contains unusual tool name(s) | Warning |
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 9 / 11 Passed | |
700fbe2
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.