Data Collection Pipeline Setup
Design and implement automated data collection pipelines for research and analytics workflows
Complexity: Complex | Duration: 30m+ | Category: Research/Intelligence
Tags: data-pipeline ETL data-engineering automation infrastructure
Workflow Diagram
Steps
Step 1: Requirements gathering
Agent: data
engineering - Define data sources, frequency, volume, and quality requirements
Step 2: Architecture design
Agent: backend
architect - Design pipeline architecture with extraction, transformation, loading
Step 3: Source integration
Agent: data
engineering - Implement data source connectors (APIs, databases, files, streams)
Step 4: Transformation logic
Agent: data
engineering - Build data cleaning, normalization, enrichment transformations
Step 5: Storage setup
Agent: backend
architect - Configure data storage (data lake, warehouse, database)
Step 6: Scheduling
Agent: devops
engineer - Implement scheduling and orchestration (Airflow, Prefect, cron)
Step 7: Monitoring
Agent: devops
engineer - Add data quality checks, error handling, alerting
Step 8: Testing
Agent: QA
automation - Test pipeline end-to-end with production-like data volumes
Usage
To execute this workflow:
/workflow research/intelligence/data-collection-pipeline-setup.workflow
Related Workflows
See other workflows in this category for related automation patterns.