Train and deploy neural networks in distributed E2B sandboxes with Flow Nexus
54
37%
Does it follow best practices?
Impact
79%
2.07xAverage score across 3 eval scenarios
Advisory
Suggest reviewing before use
Optimize this skill with Tessl
npx tessl skill review --optimize ./.agents/skills/flow-nexus-neural/SKILL.mdQuality
Discovery
32%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
The description is too terse and lacks a 'Use when...' clause, making it difficult for Claude to know when to select this skill over others. While it names a specific domain (neural networks in E2B sandboxes with Flow Nexus), it doesn't list enough concrete actions or include natural trigger terms users would employ. The product-specific terminology may help with distinctiveness but hurts discoverability.
Suggestions
Add an explicit 'Use when...' clause with trigger scenarios, e.g., 'Use when the user wants to train, fine-tune, or deploy machine learning models in sandboxed environments, or mentions E2B, Flow Nexus, or distributed GPU training.'
Include natural user-facing trigger terms like 'machine learning', 'deep learning', 'model training', 'GPU', 'ML deployment', 'fine-tuning' to improve discoverability.
Expand the list of concrete actions beyond 'train and deploy', e.g., 'configure distributed training jobs, monitor training progress, manage sandbox environments, deploy inference endpoints.'
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Names the domain (neural networks, distributed sandboxes) and two actions (train and deploy), but lacks comprehensive detail about specific capabilities beyond those two verbs. | 2 / 3 |
Completeness | Describes what it does (train and deploy neural networks) but completely lacks a 'Use when...' clause or any explicit trigger guidance for when Claude should select this skill. Per rubric guidelines, a missing 'Use when...' clause caps completeness at 2, and the 'what' is also thin, warranting a 1. | 1 / 3 |
Trigger Term Quality | Includes relevant terms like 'neural networks', 'distributed', and 'deploy', but 'E2B sandboxes' and 'Flow Nexus' are product-specific jargon that users may not naturally say. Missing common variations like 'deep learning', 'model training', 'ML', 'machine learning'. | 2 / 3 |
Distinctiveness Conflict Risk | The mention of 'E2B sandboxes' and 'Flow Nexus' provides some distinctiveness as product-specific identifiers, but 'train and deploy neural networks' is broad enough to overlap with other ML/AI-related skills. | 2 / 3 |
Total | 7 / 12 Passed |
Implementation
42%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
The skill is highly actionable with concrete, executable MCP tool call examples and clear response schemas, which is its primary strength. However, it is severely bloated—architecture patterns, use cases, and core capabilities repeat the same information multiple times, making it extremely token-inefficient. The lack of any bundle files means all content is crammed into one massive document with no progressive disclosure, and multi-step workflows lack explicit validation checkpoints.
Suggestions
Eliminate duplicate content: the LSTM, transformer, and feedforward architectures each appear 2-3 times across sections. Consolidate into a single architecture reference (ideally in a separate ARCHITECTURES.md file).
Move 'Common Use Cases', 'Architecture Patterns', and detailed response JSON examples into separate bundle files, keeping SKILL.md as a concise overview with links.
Add explicit validation checkpoints to the distributed training workflow (e.g., 'Verify cluster status is ready before deploying nodes', 'Confirm all nodes are active before starting training').
Remove the one-line descriptions of architectures like 'Standard fully-connected networks' — Claude already knows what a feedforward network is.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Extremely verbose at ~500+ lines. Massive amounts of repetitive code examples (e.g., LSTM architecture appears three separate times). Response JSON examples, architecture pattern summaries, and common use cases largely duplicate the core capability sections. Claude doesn't need exhaustive API surface documentation inlined. | 1 / 3 |
Actionability | Every capability includes fully concrete, copy-paste-ready MCP tool calls with complete parameter objects and expected JSON responses. The examples are specific and executable, covering training, inference, cluster management, and marketplace operations. | 3 / 3 |
Workflow Clarity | The distributed training section shows a reasonable sequence (init cluster → deploy nodes → connect → train → monitor → terminate), but there are no explicit validation checkpoints or error-recovery feedback loops. The 'Common Use Cases' section shows sequential steps but lacks 'validate before proceeding' gates. The troubleshooting section is reactive rather than integrated into workflows. | 2 / 3 |
Progressive Disclosure | Everything is inlined in a single monolithic file with no bundle files to offload content to. The architecture patterns, common use cases, and detailed API examples could easily be split into separate reference files. The 'Related Skills' and 'Resources' sections reference external links but the body itself is a wall of repetitive content that should be decomposed. | 1 / 3 |
Total | 7 / 12 Passed |
Validation
81%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 9 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
skill_md_line_count | SKILL.md is long (739 lines); consider splitting into references/ and linking | Warning |
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 9 / 11 Passed | |
9d4a9ea
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.