“Show me all data-load events from the last 24 hours”
“Were there any failed migrations this week?”
Best Practices
Use descriptive sources — Include the pipeline name in the source (e.g. //airflow/etl-pipeline)
Track both success and failure — Record events for both successful and failed jobs, using severity to distinguish
Include row counts — Add metrics like row counts or duration in the data payload for richer context
Combine with deployment tracking — Use the Deployment Tracking pattern to track code deploys alongside data pipeline events for full operational visibility
Enable incident response — When data pipelines fail, the Incident Response pattern helps AI quickly identify the root cause