CtrlK
BlogDocsLog inGet started
Tessl Logo

memory-systems

Guides implementation of agent memory systems, compares production frameworks (Mem0, Zep/Graphiti, Letta, LangMem, Cognee), and designs persistence architectures for cross-session knowledge retention. Use when the user asks to "implement agent memory", "persist state across sessions", "build knowledge graph for agents", "track entities over time", "add long-term memory", "choose a memory framework", or mentions temporal knowledge graphs, vector stores, entity memory, adaptive memory, dynamic memory or memory benchmarks (LoCoMo, LongMemEval).

74

Quality

67%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./skills/memory-systems/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

100%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This is an excellent skill description that clearly defines its domain (agent memory systems), lists specific capabilities (implementation guidance, framework comparison, architecture design), and provides comprehensive trigger terms covering both natural language phrases and technical terminology. It follows best practices with third-person voice, an explicit 'Use when...' clause, and enough specificity to be clearly distinguishable from other skills.

DimensionReasoningScore

Specificity

Lists multiple specific concrete actions: 'guides implementation of agent memory systems', 'compares production frameworks' with named examples (Mem0, Zep/Graphiti, Letta, LangMem, Cognee), and 'designs persistence architectures for cross-session knowledge retention'. These are concrete, well-defined capabilities.

3 / 3

Completeness

Clearly answers both 'what' (guides implementation, compares frameworks, designs persistence architectures) and 'when' with an explicit 'Use when...' clause containing numerous specific trigger phrases and scenarios.

3 / 3

Trigger Term Quality

Excellent coverage of natural terms users would say: 'implement agent memory', 'persist state across sessions', 'build knowledge graph for agents', 'track entities over time', 'add long-term memory', 'choose a memory framework', plus technical terms like 'temporal knowledge graphs', 'vector stores', 'entity memory', and specific benchmark names (LoCoMo, LongMemEval).

3 / 3

Distinctiveness Conflict Risk

Highly distinctive niche focused specifically on agent memory systems, naming specific frameworks and benchmarks. The combination of memory persistence, knowledge graphs, and named tools like Mem0/Zep/Graphiti makes it very unlikely to conflict with other skills.

3 / 3

Total

12

/

12

Passed

Implementation

35%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill is comprehensive and well-researched but suffers significantly from verbosity — it reads more like a reference document or tutorial than a concise skill file. The framework comparison tables and benchmark data are valuable but could be in a reference file, with the main skill focusing on the decision framework and key code patterns. The actionability is moderate: code examples exist but are mixed with large amounts of advisory prose.

Suggestions

Cut the content by 50%+: remove the 'When to Activate' section (duplicates frontmatter), the 'Core Concepts' explanatory paragraph, and prose that restates what tables already show. Trust Claude to interpret tables without surrounding explanation.

Move the benchmark comparison table and detailed framework descriptions to a reference file (e.g., ./references/framework-comparison.md), keeping only the decision heuristic ('choose X when Y') in the main skill.

Add a concrete validation step after framework integration, such as 'Verify: run a test query and confirm retrieval returns expected results before proceeding to production use.'

Consolidate the 8 gotchas into the 3-4 most critical ones inline, moving the rest to a reference file — several (like embedding model mismatch and graph schema rigidity) are standard engineering knowledge Claude already possesses.

DimensionReasoningScore

Conciseness

The skill is extremely verbose at ~300+ lines. It explains concepts Claude already knows (what memory is, what persistence means, what vector stores do), includes a 'When to Activate' section that duplicates the frontmatter description, and has extensive prose explanations between tables that could be dramatically condensed. The 'Core Concepts' paragraph explaining memory as a spectrum is unnecessary context-setting.

1 / 3

Actionability

The code examples for Mem0, temporal queries, and Cognee are concrete and mostly executable, but much of the skill is comparative/advisory rather than instructional. The temporal query example uses a hypothetical `graph` API that isn't tied to any specific framework, and the consolidation code is deferred to a reference file rather than shown inline. The escalation path is helpful but still somewhat abstract.

2 / 3

Workflow Clarity

The escalation path (prototype → scale → complex reasoning → full control) provides a clear decision sequence, and the error recovery section has ordered fallback strategies. However, there are no explicit validation checkpoints — for example, after integrating a memory framework, there's no 'verify retrieval works by running X' step. The benchmarking guideline (#7) is mentioned but not integrated into the workflow as a checkpoint.

2 / 3

Progressive Disclosure

There is one reference to an implementation file and several external references with 'Read when' annotations, which is good. However, the main file itself is monolithic — the framework comparison, benchmark tables, retrieval strategies, memory layers, consolidation guidance, examples, guidelines, and gotchas are all inline when much of this (especially the benchmark tables and 8 gotchas) could be split into reference files. The content that is inline is too extensive for an overview skill.

2 / 3

Total

7

/

12

Passed

Validation

100%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation11 / 11 Passed

Validation for skill structure

No warnings or errors.

Repository
muratcankoylan/Agent-Skills-for-Context-Engineering
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.