directory changes and restructuring
🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
This commit is contained in:
@@ -0,0 +1,101 @@
|
||||
# Meeting Intelligence Skill Evaluations
|
||||
|
||||
Evaluation scenarios for testing the Meeting Intelligence skill across different Claude models.
|
||||
|
||||
## Purpose
|
||||
|
||||
These evaluations ensure the Meeting Intelligence skill:
|
||||
- Gathers context from Notion workspace
|
||||
- Enriches with Claude research appropriately
|
||||
- Creates both internal pre-reads and external agendas
|
||||
- Distinguishes between Notion facts and Claude insights
|
||||
- Works consistently across Haiku, Sonnet, and Opus
|
||||
|
||||
## Evaluation Files
|
||||
|
||||
### decision-meeting-prep.json
|
||||
Tests preparation for a decision-making meeting.
|
||||
|
||||
**Scenario**: Prep for database migration decision meeting
|
||||
**Key Behaviors**:
|
||||
- Searches Notion for migration context (specs, discussions, options)
|
||||
- Fetches 2-3 relevant pages
|
||||
- Enriches with Claude research (decision frameworks, migration best practices)
|
||||
- Creates comprehensive internal pre-read with recommendation
|
||||
- Creates clean, professional external agenda
|
||||
- Clearly distinguishes Notion facts from Claude insights
|
||||
- Cross-links both documents
|
||||
|
||||
### status-meeting-prep.json
|
||||
Tests preparation for a status update or review meeting.
|
||||
|
||||
**Scenario**: Prep for project status review
|
||||
**Key Behaviors**:
|
||||
- Gathers project metrics and progress from Notion
|
||||
- Fetches relevant pages (roadmap, tasks, milestones)
|
||||
- Adds Claude context (industry benchmarks, best practices)
|
||||
- Creates internal pre-read with honest assessment
|
||||
- Creates external agenda with structured flow
|
||||
- Includes source citations using mention-page tags
|
||||
- Time-boxes agenda items
|
||||
|
||||
## Running Evaluations
|
||||
|
||||
1. Enable the `meeting-intelligence` skill
|
||||
2. Submit the query from the evaluation file
|
||||
3. Verify the skill searches Notion first (not Claude research)
|
||||
4. Check that TWO documents are created (internal + external)
|
||||
5. Verify Claude enrichment adds value without replacing Notion content
|
||||
6. Test with Haiku, Sonnet, and Opus
|
||||
|
||||
## Expected Skill Behaviors
|
||||
|
||||
Meeting Intelligence evaluations should verify:
|
||||
|
||||
### Notion Context Gathering
|
||||
- Searches workspace for relevant context first
|
||||
- Fetches specific pages (not generic)
|
||||
- Extracts key information from Notion content
|
||||
- Cites sources using mention-page tags
|
||||
|
||||
### Claude Research Integration
|
||||
- Adds industry context, frameworks, or best practices
|
||||
- Enrichment is relevant and valuable (not filler)
|
||||
- Clearly distinguishes Notion facts from Claude insights
|
||||
- Research complements (doesn't replace) Notion content
|
||||
|
||||
### Two-Document Creation
|
||||
- **Internal Pre-Read**: Comprehensive, includes strategy, recommendations, detailed pros/cons
|
||||
- **External Agenda**: Professional, focused on meeting flow, no internal strategy
|
||||
- Both documents are clearly labeled
|
||||
- Documents are cross-linked
|
||||
|
||||
### Document Quality
|
||||
- Pre-read follows structure: Overview → Background → Current Status → Context & Insights → Discussion Points
|
||||
- Agenda follows structure: Details → Objective → Agenda Items (with times) → Decisions → Actions → Resources
|
||||
- Titles include date or meeting context
|
||||
- Content is actionable and meeting-ready
|
||||
|
||||
## Creating New Evaluations
|
||||
|
||||
When adding Meeting Intelligence evaluations:
|
||||
|
||||
1. **Test different meeting types** - Decision, status, brainstorm, 1:1, sprint planning, retrospective
|
||||
2. **Vary complexity** - Simple updates vs. complex strategic decisions
|
||||
3. **Test with/without Notion content** - Rich workspace vs. minimal existing pages
|
||||
4. **Verify enrichment value** - Is Claude research genuinely helpful?
|
||||
5. **Check internal/external distinction** - Is sensitive info kept in pre-read only?
|
||||
|
||||
## Example Success Criteria
|
||||
|
||||
**Good** (specific, testable):
|
||||
- "Creates TWO documents (internal pre-read + external agenda)"
|
||||
- "Internal pre-read marked 'INTERNAL ONLY' or 'For team only'"
|
||||
- "Cites at least 2-3 Notion pages using mention-page tags"
|
||||
- "Agenda includes time allocations for each section"
|
||||
- "Claude enrichment includes decision frameworks or best practices"
|
||||
|
||||
**Bad** (vague, untestable):
|
||||
- "Creates meeting materials"
|
||||
- "Gathers context effectively"
|
||||
- "Prepares well"
|
||||
Reference in New Issue
Block a user