CtrlK
BlogDocsLog inGet started
Tessl Logo

grant-mock-reviewer

Simulates NIH study section peer review for grant proposals. Triggers when user wants mock review, critique, or evaluation of a grant proposal before submission. Generates structured critique using official NIH scoring rubric (1-9 scale), identifies weaknesses, provides actionable revision recommendations, and produces a comprehensive review summary similar to actual NIH Summary Statement.

Install with Tessl CLI

npx tessl i github:aipoch/medical-research-skills --skill grant-mock-reviewer
What are skills?

77

Does it follow best practices?

Validation for skill structure

SKILL.md
Review
Evals

Grant Mock Reviewer

A simulated NIH study section reviewer that provides structured, rigorous critique of grant proposals using the official NIH scoring criteria and methodology.

Capabilities

  1. NIH Scoring Rubric Application: Official 1-9 scale scoring across all 5 criteria
  2. Weakness Identification: Systematic detection of common proposal flaws
  3. Critique Generation: Structured written critiques for each review criterion
  4. Summary Statement: Complete mock Summary Statement output
  5. Revision Guidance: Prioritized, actionable recommendations for improvement

Usage

Command Line

# Full mock review with Summary Statement
python3 scripts/main.py --input proposal.pdf --format pdf --output review.md

# Review Specific Aims only
python3 scripts/main.py --input aims.pdf --section aims --output aims_review.md

# Targeted review (specific criterion focus)
python3 scripts/main.py --input proposal.pdf --focus approach --output approach_critique.md

# Generate NIH-style scores only
python3 scripts/main.py --input proposal.pdf --scores-only --output scores.json

# Compare before/after revision
python3 scripts/main.py --original original.pdf --revised revised.pdf --compare

As Library

from scripts.main import GrantMockReviewer

reviewer = GrantMockReviewer()
result = reviewer.review(
    proposal_text=proposal_content,
    grant_type="R01",
    section="full"
)
print(result.summary_statement)
print(result.scores)

Parameters

ParameterTypeDefaultRequiredDescription
--inputstring-YesPath to proposal file (PDF, DOCX, TXT, MD)
--formatstringautoNoInput file format (pdf, docx, txt, md)
--sectionstringfullNoSection to review (full, aims, significance, innovation, approach)
--grant-typestringR01NoGrant mechanism (R01, R21, R03, K99, F32)
--focusstring-NoFocus on specific criterion (significance, investigator, innovation, approach, environment)
--scores-onlyflagfalseNoOutput scores only (JSON)
--output, -ostringstdoutNoOutput file path
--originalstring-NoOriginal proposal for comparison
--revisedstring-NoRevised proposal for comparison
--compareflagfalseNoEnable comparison mode

NIH Scoring System

Overall Impact Score (1-9)

The single most important score reflecting the likelihood of the project to exert a sustained, powerful influence on the research field.

ScoreDescriptorLikelihood of Funding
1ExceptionalVery High
2OutstandingHigh
3ExcellentGood
4Very GoodModerate
5GoodLow-Moderate
6SatisfactoryLow
7FairVery Low
8MarginalUnlikely
9PoorNot Fundable

Individual Criteria (1-9 each)

  1. Significance: Does the project address an important problem? Will scientific knowledge be advanced?
  2. Investigator(s): Are the PIs well-suited? Adequate experience and training?
  3. Innovation: Does it challenge current paradigms? Novel concepts, approaches, methods?
  4. Approach: Sound research design? Appropriate methods? Adequate controls? Address pitfalls?
  5. Environment: Adequate institutional support? Scientific environment conducive to success?

Score Interpretation

  • 1-3 (High Priority): Compelling, well-developed proposals with strong approach
  • 4-5 (Medium Priority): Good proposals with some weaknesses
  • 6-9 (Low Priority): Significant weaknesses that diminish enthusiasm

Review Output Format

1. Score Summary

Overall Impact: [Score] - [Descriptor]

Criterion Scores:
- Significance: [Score]
- Investigator(s): [Score]
- Innovation: [Score]
- Approach: [Score]
- Environment: [Score]

2. Strengths

Bullet-point list of major strengths by criterion

3. Weaknesses

Bullet-point list of major weaknesses by criterion

4. Detailed Critique

Paragraph-form critique for each criterion following NIH style

5. Summary Statement

Complete narrative synthesis of the review

6. Revision Recommendations

Prioritized, actionable suggestions for improvement

Common Weaknesses Detected

Significance

  • Insufficient justification for the research problem
  • Incremental rather than transformative impact
  • Unclear connection to human health/disease
  • Overstatement of clinical significance without evidence

Investigator

  • Lack of relevant expertise for proposed aims
  • Insufficient track record in key methodologies
  • PI overcommitted (excessive effort on other grants)
  • Missing key collaborator expertise

Innovation

  • Straightforward extension of published work
  • Methods are standard rather than novel
  • No challenging of existing paradigms
  • Incremental rather than breakthrough potential

Approach

  • Aims too ambitious for timeframe
  • Insufficient preliminary data
  • Inadequate experimental controls
  • No discussion of pitfalls and alternatives
  • Statistical analysis plan missing or inadequate
  • Sample size/power calculations absent

Environment

  • Inadequate institutional resources
  • Missing core facility access
  • Lack of relevant equipment
  • Insufficient collaborative environment

Technical Difficulty

High - Requires deep understanding of NIH peer review processes, ability to apply standardized scoring rubrics consistently, and generation of clinically/scientifically accurate critique across diverse research domains.

Review Required: Human verification recommended before deployment in production settings.

References

  • references/nih_scoring_rubric.md - Complete NIH scoring guidelines
  • references/review_criteria_explained.md - Detailed criterion descriptions
  • references/common_weaknesses_catalog.md - Database of typical proposal flaws
  • references/summary_statement_templates.md - NIH-style statement templates
  • references/score_calibration_guide.md - Score assignment guidelines

Best Practices for Users

  1. Provide Complete Proposals: The tool works best with full Research Strategy sections
  2. Include Preliminary Data: Approach critique depends on feasibility evidence
  3. Review Multiple Times: Use iteratively as you revise
  4. Compare Versions: Track improvement between drafts
  5. Consider Multiple Perspectives: Supplement with human reviewer feedback

Limitations

  1. Cannot access external literature to verify claims
  2. May not capture domain-specific methodological nuances
  3. Scoring is simulated and may not match actual study section scores
  4. Best used as preparatory tool, not replacement for human review

Version

1.0.0 - Initial release with NIH R01/R21/R03 support

Risk Assessment

Risk IndicatorAssessmentLevel
Code ExecutionPython/R scripts executed locallyMedium
Network AccessNo external API callsLow
File System AccessRead input files, write output filesMedium
Instruction TamperingStandard prompt guidelinesLow
Data ExposureOutput files saved to workspaceLow

Security Checklist

  • No hardcoded credentials or API keys
  • No unauthorized file system access (../)
  • Output does not expose sensitive information
  • Prompt injection protections in place
  • Input file paths validated (no ../ traversal)
  • Output directory restricted to workspace
  • Script execution in sandboxed environment
  • Error messages sanitized (no stack traces exposed)
  • Dependencies audited

Prerequisites

# Python dependencies
pip install -r requirements.txt

Evaluation Criteria

Success Metrics

  • Successfully executes main functionality
  • Output meets quality standards
  • Handles edge cases gracefully
  • Performance is acceptable

Test Cases

  1. Basic Functionality: Standard input → Expected output
  2. Edge Case: Invalid input → Graceful error handling
  3. Performance: Large dataset → Acceptable processing time

Lifecycle Status

  • Current Stage: Draft
  • Next Review Date: 2026-03-06
  • Known Issues: None
  • Planned Improvements:
    • Performance optimization
    • Additional feature support
Repository
aipoch/medical-research-skills
Last updated
Created

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.