Custom Skills (ourdigital-custom-skills/): - 00-ourdigital-visual-storytelling: Blog featured image prompt generator - 01-ourdigital-research-publisher: Research-to-publication workflow - 02-notion-organizer: Notion workspace management - 03-research-to-presentation: Notion research to PPT/Figma - 04-seo-gateway-strategist: SEO gateway page strategy planning - 05-gateway-page-content-builder: Gateway page content generation - 20-jamie-brand-editor: Jamie Clinic branded content GENERATION - 21-jamie-brand-guardian: Jamie Clinic content REVIEW & evaluation Refinements applied: - All skills converted to SKILL.md format with YAML frontmatter - Added version fields to all skills - Flattened nested folder structures - Removed packaging artifacts (.zip, .skill files) - Reorganized file structures (scripts/, references/, etc.) - Differentiated Jamie skills with clear roles 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
Knowledge Capture Skill Evaluations
Evaluation scenarios for testing the Knowledge Capture skill across different Claude models.
Purpose
These evaluations ensure the Knowledge Capture skill:
- Correctly identifies content types (how-to guides, FAQs, decision records, wikis)
- Extracts relevant information from conversations
- Structures content appropriately for each type
- Searches and places content in the right Notion location
- Works consistently across Haiku, Sonnet, and Opus
Evaluation Files
conversation-to-wiki.json
Tests capturing conversation content as a how-to guide for the team wiki.
Scenario: Save deployment discussion to wiki
Key Behaviors:
- Extracts steps, gotchas, and best practices from conversation
- Identifies content as How-To Guide
- Structures with proper sections (Overview, Prerequisites, Steps, Troubleshooting)
- Searches for team wiki location
- Preserves technical details (commands, configs)
decision-record.json
Tests capturing architectural or technical decisions with full context.
Scenario: Document database migration decision
Key Behaviors:
- Extracts decision context, alternatives, and rationale
- Follows decision record structure (Context, Decision, Alternatives, Consequences)
- Captures both selected and rejected options with reasoning
- Places in decision log or ADR database
- Links to related technical documentation
Running Evaluations
- Enable the
knowledge-captureskill - Submit the query from the evaluation file
- Provide conversation context as specified
- Verify all expected behaviors are met
- Check success criteria for quality
- Test with Haiku, Sonnet, and Opus
Expected Skill Behaviors
Knowledge Capture evaluations should verify:
Content Extraction
- Accurately captures key points from conversation context
- Preserves specific technical details, not generic placeholders
- Maintains context and nuance from discussion
Content Type Selection
- Correctly identifies appropriate content type (how-to, FAQ, decision record, wiki page)
- Uses matching structure from reference documentation
- Applies proper Notion markdown formatting
Notion Integration
- Searches for appropriate target location (wiki, decision log, etc.)
- Creates well-structured pages with clear titles
- Uses proper parent placement
- Includes discoverable titles and metadata
Quality Standards
- Content is actionable and future-reference ready
- Technical accuracy is preserved
- Organization aids discoverability
- Formatting enhances readability
Creating New Evaluations
When adding Knowledge Capture evaluations:
- Use realistic conversation content - Include actual technical details, decisions, or processes
- Test different content types - How-to guides, FAQs, decision records, meeting notes, learnings
- Vary complexity - Simple captures vs. complex technical discussions
- Test discovery - Finding the right wiki section or database
- Include edge cases - Unclear content types, minimal context, overlapping categories
Example Success Criteria
Good (specific, testable):
- "Structures content using How-To format with numbered steps"
- "Preserves exact bash commands from conversation"
- "Creates page with title format 'How to [Action]'"
- "Places in Engineering Wiki → Deployment section"
Bad (vague, untestable):
- "Creates good documentation"
- "Uses appropriate structure"
- "Saves to the right place"