Custom Skills (ourdigital-custom-skills/): - 00-ourdigital-visual-storytelling: Blog featured image prompt generator - 01-ourdigital-research-publisher: Research-to-publication workflow - 02-notion-organizer: Notion workspace management - 03-research-to-presentation: Notion research to PPT/Figma - 04-seo-gateway-strategist: SEO gateway page strategy planning - 05-gateway-page-content-builder: Gateway page content generation - 20-jamie-brand-editor: Jamie Clinic branded content GENERATION - 21-jamie-brand-guardian: Jamie Clinic content REVIEW & evaluation Refinements applied: - All skills converted to SKILL.md format with YAML frontmatter - Added version fields to all skills - Flattened nested folder structures - Removed packaging artifacts (.zip, .skill files) - Reorganized file structures (scripts/, references/, etc.) - Differentiated Jamie skills with clear roles 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
Research & Documentation Skill Evaluations
Evaluation scenarios for testing the Research & Documentation skill across different Claude models.
Purpose
These evaluations ensure the Research & Documentation skill:
- Searches across Notion workspace effectively
- Synthesizes information from multiple sources
- Selects appropriate research report format
- Creates comprehensive documentation with proper citations
- Works consistently across Haiku, Sonnet, and Opus
Evaluation Files
basic-research.json
Tests basic research workflow with synthesis across multiple Notion pages.
Scenario: Research Q4 product roadmap and create summary
Key Behaviors:
- Searches Notion for roadmap-related pages
- Fetches multiple relevant pages (roadmap, product docs, meeting notes)
- Synthesizes information from different sources
- Selects appropriate format (Research Summary)
- Includes citations linking back to source pages
- Creates structured document with clear sections
research-to-database.json
Tests creating research documentation in a Notion database with properties.
Scenario: Research competitor landscape and save to Research database
Key Behaviors:
- Searches for existing competitive intelligence in Notion
- Identifies Research database as target
- Fetches database schema to understand properties
- Creates page with correct property values (Research Type, Status, Date, etc.)
- Structures content with comparison format
- Includes source citations for both Notion pages and external research
Running Evaluations
- Enable the
research-documentationskill - Submit the query from the evaluation file
- Verify the skill searches Notion workspace comprehensively
- Check that multiple source pages are fetched and synthesized
- Verify appropriate format is selected (Research Summary, Comprehensive Report, Quick Brief, Comparison)
- Confirm citations link back to sources
- Test with Haiku, Sonnet, and Opus
Expected Skill Behaviors
Research & Documentation evaluations should verify:
Notion Search & Synthesis
- Searches workspace with relevant queries
- Fetches multiple source pages (3-5+)
- Synthesizes information across sources
- Identifies patterns and insights
- Handles conflicting information appropriately
Format Selection
- Chooses correct format based on scope and depth:
- Research Summary: Quick overview with key findings
- Comprehensive Report: Deep analysis with multiple sections
- Quick Brief: Fast facts and takeaways
- Comparison: Side-by-side analysis
- Applies format structure consistently
- Uses appropriate sections and headings
Citation & Attribution
- Includes citations for all Notion sources
- Uses mention-page tags:
<mention-page url="..."> - Attributes findings to specific sources
- Distinguishes between Notion content and Claude research
- Links related documents
Document Quality
- Title clearly indicates research topic and date
- Executive summary or key findings upfront
- Organized with clear hierarchy
- Actionable insights and recommendations
- Appropriate depth for the query
Creating New Evaluations
When adding Research & Documentation evaluations:
- Test different research types - Product research, competitive analysis, technical investigation, market research
- Vary source count - Synthesis of 2-3 pages vs. 10+ pages
- Test format selection - Does it choose the right format for the scope?
- Include database targets - Not just standalone pages
- Test citation accuracy - Are all sources properly attributed?
- Cross-workspace search - Testing search across teamspaces if applicable
Example Success Criteria
Good (specific, testable):
- "Searches Notion for 'roadmap' and 'Q4' and 'product'"
- "Fetches at least 3 different source pages"
- "Includes citation for each key finding using mention-page tags"
- "Creates page with title format 'Research: [Topic] - [Date]'"
- "Uses Research Summary format with sections: Executive Summary → Key Findings → Details → Recommendations → Sources"
Bad (vague, untestable):
- "Searches Notion effectively"
- "Creates comprehensive research"
- "Uses sources appropriately"
- "Good documentation"