Browse working recipe examples
This catalog describes all example recipes included in the collection. Each example demonstrates different patterns and use cases.
In a session (recommended): Just ask naturally - "run the code-review recipe on src/auth.py"
From CLI: Use amplifier tool invoke recipes:
amplifier tool invoke recipes operation=execute recipe_path=<recipe.yaml> context='{"key": "value"}'Note: There is no
amplifier recipesCLI command. Recipes are invoked via therecipestool.
| Recipe | Domain | Steps | Key Pattern | Agents Used |
|---|---|---|---|---|
| simple-analysis | Tutorial | 3 | Sequential basics | zen-architect |
| bash-step-example | Tutorial | 7 | Bash steps | (none - bash only) |
| conditional-workflow | Tutorial | 4 | Conditional routing | zen-architect |
| code-review | Code Quality | 4 | Multi-mode agents | zen-architect |
| comprehensive-review | Code Quality | 3 | Recipe composition | zen-architect, security-guardian |
| security-audit | Security | 4 | Domain specialist | security-guardian, zen-architect |
| test-generation | Testing | 3 | Artifact generation | zen-architect, test-coverage |
| dependency-upgrade | DevOps | 4 | Planning workflow | zen-architect, integration-specialist |
| dependency-upgrade-staged | DevOps | 5 stages | Approval gates | zen-architect, integration-specialist |
| multi-file-analysis | Analysis | 2 | Parallel foreach | zen-architect |
| parallel-analysis | Analysis | 2 | Multi-perspective | zen-architect |
| multi-level-python-code-analysis | Analysis | 15 | Multi-level deep dive | explorer, bug-hunter, zen-architect |
| repo-activity-analysis | GitHub | 6 | Bash + agent hybrid | zen-architect |
| multi-repo-activity-report | GitHub | 3 | Recipe invocation | zen-architect |
| feature-announcement | Communication | 3 | Human-first output | zen-architect |
| session-extract | Verification | 1 | Quick metadata check | session-analyst |
| session-verify | Verification | 2 | Cross-provider audit | session-analyst, zen-architect |
| adversarial-verification | Verification | 5 stages | Deep investigation with adversarial verification | explorer, zen-architect, bug-hunter |
| context-intelligence | Advanced | varies | Tiered complexity | various |
File: examples/simple-analysis-recipe.yaml
Basic analysis workflow demonstrating core recipe concepts: context variables, sequential steps, and output accumulation.
Use Cases:
- Learning recipe basics
- Template for new recipes
- Simple file analysis
Workflow:
1. Extract Summary → 3-sentence summary
2. Identify Key Points → 3-5 key points
3. Generate Report → Markdown report
Example Usage:
In session: "run simple-analysis on README.md"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=simple-analysis-recipe.yaml \
context='{"file_path": "README.md"}'
Key Learnings:
- Context variables with
{{variable}}syntax - Sequential step execution
- Output chaining between steps
File: examples/bash-step-example.yaml
Demonstrates bash step capabilities for direct shell execution without LLM overhead.
Use Cases:
- System commands in workflows
- Environment variable handling
- Exit code-based conditionals
- Working directory control
Workflow:
1. Simple Echo → Basic command
2. Environment Variables → Variable passing
3. Multi-line Command → System info
4. Check Command → Exit code capture
5. Report Success → Conditional step
6. List Tmp → Working directory
7. Summary → Results compilation
Example Usage:
In session: "run bash-step-example with project_name MyProject"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=bash-step-example.yaml \
context='{"project_name": "MyProject"}'
Key Learnings:
type: "bash"for shell commands (no LLM cost)- Environment variables via
env:section - Exit code capture with
output_exit_code - Conditional steps based on exit codes
- Working directory control with
cwd:
File: examples/conditional-workflow.yaml
Demonstrates conditional step execution based on classification results.
Use Cases:
- Dynamic workflow routing
- Complexity-based processing
- Skip unnecessary steps
Workflow:
1. Classify → Determine simple vs complex
2a. Simple Process → (if simple) Direct processing
2b. Complex Analyze → (if complex) Detailed analysis
2c. Complex Synthesize → (if complex) Synthesis
3. Report → Final summary
Example Usage:
In session: "run conditional-workflow on this complex dataset"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=conditional-workflow.yaml \
context='{"input_data": "complex dataset with multiple dimensions"}'
Key Learnings:
condition:field for conditional execution- Classification-based routing
- Multiple paths through workflow
File: examples/code-review-recipe.yaml
Multi-stage code review workflow with conditional execution based on issue severity.
Use Cases:
- Pre-merge code review
- Architecture assessment
- Technical debt identification
Workflow:
1. Analyze Structure → Extract patterns, complexity
2. Identify Issues → Find problems, assess severity
3. Suggest Improvements → (if issues found) Design solutions
4. Validate Suggestions → (if critical/high) Deep validation
Example Usage:
In session: "run code-review on src/auth.py"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=code-review-recipe.yaml \
context='{"file_path": "src/auth.py"}'
Key Learnings:
- Multi-mode agent usage (ANALYZE, ARCHITECT, REVIEW)
- Severity-based conditional execution
- Progressive refinement pattern
File: examples/comprehensive-review.yaml
Combines code review and security audit using recipe composition.
Use Cases:
- Combined quality and security analysis
- Modular workflow design
- DRY principle (reuse existing recipes)
Workflow:
1. Code Review (sub-recipe) → Invoke code-review-recipe.yaml
2. Security Audit (sub-recipe) → Invoke security-audit-recipe.yaml
3. Synthesize → Unified prioritized action plan
Example Usage:
In session: "run comprehensive-review on src/api.py"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=comprehensive-review.yaml \
context='{"file_path": "src/api.py"}'
Key Learnings:
- Recipe composition with
type: "recipe" - Context isolation for sub-recipes
- Recursion protection with
max_depth
File: examples/security-audit-recipe.yaml
Comprehensive security analysis covering vulnerabilities, configurations, and best practices.
Use Cases:
- Pre-production security review
- Vulnerability assessment
- Compliance checks
Workflow:
1. Vulnerability Scan → Static analysis
2. Configuration Review → Security settings
3. Dependency Audit → Vulnerable dependencies
4. Synthesize Findings → Prioritized remediation
Example Usage:
In session: "run security-audit on src/api.py"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=security-audit-recipe.yaml \
context='{"file_path": "src/api.py", "severity_threshold": "medium"}'
Key Learnings:
- Domain specialist agent (security-guardian)
- Multi-perspective security analysis
- Actionable remediation output
File: examples/test-generation-recipe.yaml
Analyze code and generate comprehensive test suite.
Use Cases:
- Adding tests to legacy code
- Improving test coverage
- TDD verification
Workflow:
1. Analyze Code Structure → Understand code
2. Design Test Strategy → Determine test approach
3. Generate Test Code → Create actual tests
Example Usage:
In session: "generate tests for src/utils.py using pytest"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=test-generation-recipe.yaml \
context='{"file_path": "src/utils.py", "test_framework": "pytest"}'
Key Learnings:
- Specialized agent (test-coverage)
- Artifact generation (ready-to-use test code)
- Configurable output (framework, coverage target)
File: examples/dependency-upgrade-recipe.yaml
Systematic dependency upgrade workflow (flat, non-staged).
Use Cases:
- Monthly dependency updates
- Security vulnerability remediation
- Quick upgrade planning
Workflow:
1. Audit Current Dependencies → List versions, check updates
2. Plan Upgrade Strategy → Determine order, identify risks
3. Validate Compatibility → Check breaking changes
4. Generate Upgrade Commands → Ready-to-run commands
Example Usage:
In session: "run dependency-upgrade for this project using pip"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=dependency-upgrade-recipe.yaml \
context='{"project_path": ".", "package_manager": "pip"}'
Key Learnings:
- Domain-specific agent (integration-specialist)
- Risk assessment in planning
- Actionable output (executable commands)
File: examples/dependency-upgrade-staged-recipe.yaml
Dependency upgrade with approval gates between phases.
Use Cases:
- High-stakes production upgrades
- Phased rollout (security → minor → major)
- Human-in-loop verification
Workflow (Staged):
Stage 1: Assessment (no approval)
→ Audit dependencies, plan strategy
Stage 2: Validation (approval required)
→ Check compatibility
Stage 3: Phase 1 - Critical (approval required)
→ Security fixes only
Stage 4: Phase 2 - Minor (approval required)
→ Minor version updates
Stage 5: Phase 3 - Major (approval required)
→ Major version updates (highest risk)
Example Usage:
In session: "run dependency-upgrade-staged for this project"
CLI:
# Start the recipe
amplifier tool invoke recipes operation=execute \
recipe_path=dependency-upgrade-staged-recipe.yaml \
context='{"project_path": ".", "package_manager": "uv"}'
# Check pending approvals
amplifier tool invoke recipes operation=approvals
# Approve a stage
amplifier tool invoke recipes operation=approve \
session_id=<session_id> stage_name=validation
Key Learnings:
stages:for approval gates- Human-in-loop checkpoints
- Risk-ordered execution phases
- Compare with flat version for when to use each
File: examples/multi-file-analysis.yaml
Analyze multiple files in parallel with per-file insights and consolidated summary.
Use Cases:
- Bulk file analysis
- Codebase surveys
- Parallel processing patterns
Workflow:
1. Analyze Each File (parallel foreach)
→ All files analyzed concurrently
→ Results collected into array
2. Create Summary
→ Synthesize all analyses
→ Executive summary with priorities
Example Usage:
In session: "analyze src/auth.py, src/models.py, and src/utils.py together"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=multi-file-analysis.yaml \
context='{"files": ["src/auth.py", "src/models.py", "src/utils.py"]}'
Key Learnings:
foreach:withparallel: truefor concurrent executioncollect:to gather results into arrayon_error: "continue"for fault tolerance- 3x+ speedup vs sequential
File: examples/parallel-analysis-recipe.yaml
Analyze code from multiple perspectives simultaneously.
Use Cases:
- Multi-perspective code review
- Trade-off analysis
- Comprehensive assessment
Workflow:
1. Multi-Perspective Analysis (parallel foreach)
→ Security, performance, maintainability, testability
→ All run concurrently
2. Synthesize Perspectives
→ Cross-cutting concerns
→ Prioritized recommendations
Example Usage:
In session: "analyze src/core.py from multiple perspectives"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=parallel-analysis-recipe.yaml \
context='{"file_path": "src/core.py"}'
Key Learnings:
- Same file, multiple perspectives
- Parallel foreach over perspectives list
- Synthesis of competing concerns
File: examples/multi-level-python-code-analysis.yaml
Systematic Python codebase analysis at multiple granularities (repo → tree → file) using multiple perspectives (hypothesis-driven, exploratory, architectural).
Use Cases:
- Dead code detection after refactoring
- Retcon verification (e.g., removing deprecated features)
- API migration audits
- Security pattern analysis
- Technical debt assessment
Workflow:
Phase 0: Repository Context (1 step)
→ Establish baseline understanding
Phase 1: Repository Level (3 steps)
→ Hypothesis-driven dead code detection
→ Exploratory pattern matching
→ Differential analysis
Phase 2: Directory Tree Level (4 steps)
→ Identify and prioritize trees
→ Per-tree hypothesis testing
→ Per-tree exploratory analysis
→ Tree-level differential analysis
Phase 3: File Level (4 steps)
→ Identify high-risk files
→ Hypothesis-based reachability testing
→ Detailed file inspection
→ Action plan generation
Phase 4: Synthesis (3 steps)
→ Aggregate findings
→ Generate comprehensive report
→ Create executable remediation scripts
Example Usage:
In session: "run multi-level-python-code-analysis on amplifier-core to find dead code"
CLI:
amplifier tool invoke recipes operation=execute \
recipe_path=multi-level-python-code-analysis.yaml \
context='{"repo_path": "/path/to/repo", "analysis_ask": "Find dead code after API migration", "search_patterns": ["old_api", "deprecated"]}'
Key Learnings:
- Multi-level analysis (repo → tree → file)
- Multi-perspective validation (3 agents with different approaches)
- Hypothesis-driven dead code detection
- Differential analysis (grep vs import tracing)
- Comprehensive actionable output (scripts, checklists, reports)
- Python-specific entry point analysis (main.py, init.py, pytest)
Output Artifacts:
MULTI_LEVEL_CODE_ANALYSIS_REPORT.md- Comprehensive findingsfix_critical_bugs.sh- Executable remediation scriptdelete_dead_code.sh- Safe deletion scriptverify_removal.sh- Post-deletion verification- Various JSON reports for programmatic access
Note: This recipe is Python-specific. For other languages, customize the entry point patterns in the recipe prompts.
File: examples/repo-activity-analysis.yaml
Analyze a GitHub repository for commits and PRs in a date range.
Use Cases:
- Daily/weekly activity reports
- Contribution tracking
- Release notes preparation
Workflow:
1. Parse Date Range → Convert natural language to dates
2. Detect/Parse Repo → Get owner/name from URL or cwd
3. Fetch Commits (bash) → gh CLI for commit data
4. Fetch PRs (bash) → gh CLI for PR data
5. Synthesize Report → Human-readable summary
6. Write Files (bash) → Save report to disk
Example Usage:
In session: "analyze this repo's activity for the last week"
CLI (defaults - current repo since yesterday):
amplifier tool invoke recipes operation=execute \
recipe_path=repo-activity-analysis.yaml
CLI (specific repo and date range):
amplifier tool invoke recipes operation=execute \
recipe_path=repo-activity-analysis.yaml \
context='{"repo_url": "https://github.com/microsoft/amplifier-core", "date_range": "last 7 days"}'
Key Learnings:
- Hybrid bash + agent steps
type: "bash"for deterministic commands (no LLM cost)- Natural language date parsing
_precomputedpattern for orchestration optimization
File: examples/multi-repo-activity-report.yaml
Generate activity reports across multiple GitHub repositories.
Use Cases:
- Ecosystem-wide activity tracking
- Multi-repo release coordination
- Team contribution reports
Workflow:
1. Load Repos → From JSON list or manifest file
2. Analyze Each Repo (foreach)
→ Invokes repo-activity-analysis.yaml for each
→ Parallel execution option
3. Synthesize Cross-Repo Report
→ Combined activity summary
→ Cross-repo patterns
Example Usage:
In session: "generate activity report for amplifier-core and amplifier-foundation"
CLI (with repo list):
amplifier tool invoke recipes operation=execute \
recipe_path=multi-repo-activity-report.yaml \
context='{"repos": [{"owner": "microsoft", "name": "amplifier-core", "url": "https://github.com/microsoft/amplifier-core"}]}'
CLI (with manifest file):
amplifier tool invoke recipes operation=execute \
recipe_path=multi-repo-activity-report.yaml \
context='{"repos_manifest": "./repos.json"}'
Key Learnings:
- Recipe invocation within foreach
- Manifest file support
- Cross-repo synthesis
File: examples/feature-announcement.yaml
Generate human-friendly feature announcements from Amplifier session work.
Original Author: Salil Das (https://github.com/sadlilas)
Use Cases:
- Communicate technical changes to mixed audiences
- Generate plain-text announcements for Teams/Slack
- Consistent announcement format
Workflow:
1. Analyze Changes
→ Session analysis, git history, or user description
→ Structured change summary
2. Generate Announcement
→ Human-first, plain text
→ Hook + bullets + try it + links format
3. Save Announcement
→ Write to file for easy copy-paste
Input Modes:
- Current session (default)
- Specific session ID
- Git repository history
Example Usage:
In session: "run the feature-announcement recipe"
CLI (analyze current session):
amplifier tool invoke recipes operation=execute \
recipe_path=feature-announcement.yaml
CLI (with git history):
amplifier tool invoke recipes operation=execute \
recipe_path=feature-announcement.yaml \
context='{"repo_path": ".", "git_range": "HEAD~5..HEAD"}'
CLI (with user description):
amplifier tool invoke recipes operation=execute \
recipe_path=feature-announcement.yaml \
context='{"user_description": "Added new caching layer for API responses"}'
Key Learnings:
- Human-first output (plain text, no markdown)
- Multiple input mode detection
- Structured format for scanning (hook, bullets, action, links)
File: examples/session-extract.yaml
Quick metadata extraction to see what provider/model was used in a session. Use this when you don't remember what provider a session used and want to decide what verifier to use.
Use Cases:
- Check what provider/model a session used
- Decide which provider to use for verification
- Quick session overview before full audit
Workflow:
1. Quick Extract → Provider, model, one-line summary, turn count
Example Usage:
In session: "run session-extract on session abc123"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=session-extract.yaml \
context='{"session_id": "abc123"}'
Output:
ORIGINAL_PROVIDER: anthropic
ORIGINAL_MODEL: claude-opus-4-5-20251101
One-Line Summary: Fixed authentication bug and added rate limiting
Turn Count: 8
Key Learnings:
- Lightweight extraction (single step, fast)
- Guides user to next step (session-verify)
- Part of two-recipe verification workflow
File: examples/session-verify.yaml
Cross-provider verification audit that checks if work in a session actually accomplished what was intended. Uses a DIFFERENT provider than the original to catch blind spots.
Use Cases:
- Verify work actually got done (not just claimed)
- Cross-provider auditing (Anthropic reviews OpenAI's work, etc.)
- Post-session quality assurance
- Catch gaps between intent, claims, and reality
Why Cross-Provider?
- Different models have different blind spots and biases
- Anthropic models may miss what OpenAI catches (and vice versa)
- A model is unlikely to catch its own systematic errors
Workflow:
1. Extract Session Context → Intent, claims, reality from session
2. Verification Review → Independent audit using different provider
→ Verdict: PASS or ISSUES FOUND
→ Evidence-based assessment
Example Usage:
In session: "run session-verify on session abc123 with verifier_provider openai"
CLI: amplifier tool invoke recipes operation=execute \
recipe_path=session-verify.yaml \
context='{"session_id": "abc123", "verifier_provider": "openai"}'
Inputs:
session_id(required): Session to verifyverifier_provider(required): Provider for verification (openai, anthropic, google)verifier_model(optional): Override default model for provider
Default Models per Provider:
- openai → gpt-5.2
- google → gemini-3-pro-preview
- anthropic → claude-opus-4-5-20251101
Output:
Verdict: ISSUES FOUND
Intent Coverage:
✓ VERIFIED: Fix authentication bug - Evidence confirms completion
✗ GAP: Add rate limiting - Code added but never tested
? UNCLEAR: Update docs - No evidence either way
Issues:
- Rate limiting code was written but no test execution found
- Claimed "tests pass" but no test command in tool results
Key Learnings:
- Cross-provider verification pattern
- Intent vs claims vs reality framework
- Evidence-based verdicts
- Two-recipe workflow (extract → verify)
Verification-Driven Development (VDD):
These recipes implement a simplified version of VDD - a methodology where:
- Work is done by one AI provider
- A different AI provider audits the work
- Different training = different blind spots = better coverage
The key insight: fresh context + different provider = truly independent review.
File: examples/context-intelligence/verification/adversarial-verification.yaml
Deep investigation of a technical topic with multi-round adversarial verification. Uses multiple specialized agents to investigate, then subjects findings to iterative challenge/defense rounds with escalation.
Use Cases:
- Architecture documentation with verified accuracy
- Codebase deep-dives with fact-checking
- Technical investigations requiring high confidence
- Any documentation that needs adversarial review
Workflow (Staged):
Stage 1: Pre-Check (no approval)
→ Validate inputs, check paths exist
Stage 2: Initial Investigation (no approval)
→ Parallel multi-agent investigation
→ explorer, LSP, architect agents
→ Confidence tagging on all claims
Stage 3: Synthesis (no approval)
→ Combine findings into structured document v1
→ Include confidence markers
Stage 4: Adversarial Verification (no approval)
→ Iterative rounds with 4 verifiers:
- LSP verifier (code accuracy)
- Logic verifier (consistency) ⭐ weighted 1.5x
- Flow verifier (sequences)
- Completeness verifier (omissions)
→ Escalation tiers: Standard → Aggressive → Final
→ Convergence detection (v2.2+)
→ Scope-aware exit (v2.2+)
Stage 5: Final Review (approval required)
→ Human approval of verified document
→ Auto-approve if conditions met (v2.2+)
v2.2 Features:
- Convergence detection - Auto-exits when issues plateau
- Scope-aware exit - Exits if only scope/completeness issues remain
- Dynamic agent weighting - Reduced completeness weight in continuation mode
- Continuation mode - Resume verification from previous document
- Smart auto-approve - Auto-approve when factual verifiers clean
Example Usage:
In session: "run adversarial-verification to investigate the authentication system"
CLI (fresh investigation):
amplifier tool invoke recipes operation=execute \
recipe_path=@recipes:examples/context-intelligence/verification/adversarial-verification.yaml \
context='{"topic": "How does authentication work?", "output_dir": "./docs/investigations/auth", "codebase_paths": ["./src"]}'
CLI (continuation from previous):
amplifier tool invoke recipes operation=execute \
recipe_path=@recipes:examples/context-intelligence/verification/adversarial-verification.yaml \
context='{"topic": "How does authentication work?", "output_dir": "./docs/investigations/auth-round2", "continue_from": "./docs/investigations/auth/INVESTIGATION_v3.md", "starting_version": 3}'
Key Learnings:
- Multi-agent parallel investigation
- Iterative adversarial verification with escalation
- Convergence detection for plateau handling
- Factual vs scope issue classification
- Continuation mode for extended verification
Output Artifacts:
INVESTIGATION_FINAL.md- Verified documentverification-log.md- Round-by-round audit trailmetrics.json- Machine-readable statisticsINVESTIGATION_v*.md- Version history
Directory: examples/context-intelligence/
A collection of advanced recipes for context-aware AI workflows, organized in tiers of complexity.
Structure:
context-intelligence/
├── tier1/ # Foundation patterns
├── tier2/ # Intermediate patterns
├── tier3/ # Advanced patterns
├── compression/ # Context compression techniques
├── synthesis/ # Information synthesis
├── verification/ # Output verification
├── orchestrators/ # Custom orchestration
├── workflows/ # Complete workflow examples
├── foundation/ # Core utilities
├── shared/ # Shared resources
└── test/ # Test recipes
See examples/context-intelligence/README.md for detailed documentation.
These recipes are primarily for testing recipe engine functionality:
| Recipe | Purpose |
|---|---|
test-parse-json.yaml |
Verify parse_json functionality |
ultra-minimal-test.yaml |
Minimal recipe for quick testing |
# Copy example to your recipes directory
cp examples/code-review-recipe.yaml my-recipes/custom-review.yaml
# Edit as needed, then run in a session:
# "run my-recipes/custom-review.yaml on src/mycode.py"Each example demonstrates specific patterns:
| Pattern | Examples |
|---|---|
| Sequential steps | simple-analysis |
| Bash steps | bash-step-example, repo-activity-analysis |
| Conditional execution | conditional-workflow, code-review |
| Parallel foreach | multi-file-analysis, parallel-analysis |
| Approval gates (staged) | dependency-upgrade-staged |
| Recipe composition | comprehensive-review, multi-repo-activity-report |
| Multi-mode agents | code-review |
| Domain specialists | security-audit, test-generation |
| Domain | Examples |
|---|---|
| Tutorial | simple-analysis, bash-step-example, conditional-workflow |
| Code Quality | code-review, comprehensive-review |
| Security | security-audit |
| Testing | test-generation |
| DevOps | dependency-upgrade, dependency-upgrade-staged |
| Analysis | multi-file-analysis, parallel-analysis |
| GitHub | repo-activity-analysis, multi-repo-activity-report |
| Communication | feature-announcement |
| Verification | session-extract, session-verify |
| Advanced | context-intelligence |
| Pattern | Examples |
|---|---|
| Sequential | simple-analysis, code-review |
| Conditional | conditional-workflow, code-review |
| Parallel | multi-file-analysis, parallel-analysis |
| Staged (Approval Gates) | dependency-upgrade-staged |
| Recipe Composition | comprehensive-review, multi-repo-activity-report |
| Bash Hybrid | bash-step-example, repo-activity-analysis |
| Artifact Generation | test-generation, feature-announcement |
| Cross-Provider Verification | session-extract, session-verify |
| Two-Recipe Workflow | session-extract, session-verify |
Have a useful recipe? Share it!
- Test thoroughly with multiple inputs
- Document well (context vars, usage, expected output)
- Follow best practices (see Best Practices)
- Submit PR with:
- Recipe YAML file
- Description for this catalog
- Example usage
See Contributing Guidelines for details.
- Browse examples in
examples/directory - Try an example with your data
- Customize for your needs
- Create your own using patterns learned
- Share useful recipes with community
Questions about examples? Ask in GitHub Discussions!