Extract meeting details with GPT-4.1-mini and evaluate accuracy in Google Sheets
DISCOUNT 20%
Who's it for
Developers building AI-powered workflows who want to ensure their agents work reliably. If you need to validate AI outputs, test agent behavior systematically, or build maintainable automation, this template shows you how.
What it does
This subworkflow extracts structured meeting details (title, date, time, location, links, attendees) from natural language messages using an AI agent. It demonstrates production-ready patterns:
- Structured output validation: JSON schema enforcement prevents malformed responses
- Error handling: Graceful failures with full execution traceability
- Automated evaluation: Test agent accuracy against expected outputs using Google Sheets
- Dual execution modes: Normal extraction + evaluation/testing mode
The AI resolves relative time ("tomorrow", "next Friday") using timezone context and handles incomplete data gracefully.
How to set it up
- Connect OpenAI API credential to the AI agent node
- Copy the test data sheet: https://docs.google.com/spreadsheets/d/1U89nPsasM2WNv1D7gEYINhDwylyxYw7BOd_i8ipFC0M/edit?usp=sharing
- Update Google Sheet IDs in
load_eval_dataandrecord_eval_outputnodes - Test normal mode: Execute workflow "from trigger"
- Test evaluation mode: Execute workflow "from load_eval_data"
Requirements
- OpenAI API key
- Google Sheets OAuth credential
Why subworkflow architecture?
Reusability: Wrap AI agents in subworkflows to call them from multiple parent workflows. Extract meetings from Slack, email, or webhooks—same agent, consistent results.
Testability: This pattern enables isolated testing for each AI component. Set up evaluation datasets, run automated tests, and validate accuracy before deploying to production. You can't do this easily with inline agents.
Maintainability: Update the agent logic once, and all parent workflows benefit. Error handling and validation are built-in, so failures are traceable with execution IDs.
This framework includes:
- Dual-trigger pattern (normal + evaluation modes)
- Output validation that catches silent AI failures
- Error bubbling with execution metadata for debugging
- Evaluation framework with semantic/exact matching
- Proper routing that returns output to parent workflows
Following this pattern for other agents
To adapt this for any AI task (contact extraction, invoice processing, sentiment analysis, etc.):
- Replace
extract_meeting_detailswith your AI agent (add tools, memory, etc. as needed) - Update
Structured Output Parserschema to match your data structure - Modify
evaluate_matchprompt for your validation criteria - Create test cases in Google Sheets with your inputs/expected outputs
- Adjust
normalize_eval_datatimezone/reference time if needed
The validation, error handling, and evaluation infrastructure stays the same regardless of what your agent does.