Agent-almanac orchestrate-ml-pipeline
git clone https://github.com/pjt222/agent-almanac
T=$(mktemp -d) && git clone --depth=1 https://github.com/pjt222/agent-almanac "$T" && mkdir -p ~/.claude/skills && cp -r "$T/i18n/wenyan-lite/skills/orchestrate-ml-pipeline" ~/.claude/skills/pjt222-agent-almanac-orchestrate-ml-pipeline-8ab678 && rm -rf "$T"
i18n/wenyan-lite/skills/orchestrate-ml-pipeline/SKILL.mdOrchestrate ML Pipeline
See Extended Examples for complete configuration files and templates.
Build and orchestrate end-to-end machine learning pipelines with dependency management, scheduling, and monitoring.
When to Use
- Automating multi-step ML workflows from data ingestion to deployment
- Scheduling periodic model retraining on fresh data
- Coordinating distributed data processing and training tasks
- Implementing complex dependencies between ML pipeline stages
- Managing retry logic and failure recovery
- Monitoring pipeline execution and alerting on failures
- Orchestrating feature engineering, training, evaluation, and deployment
- Building reproducible ML workflows across environments
Inputs
- Required: ML pipeline components (data ingestion, preprocessing, training, evaluation)
- Required: Orchestration framework choice (Prefect, Airflow, Kubeflow)
- Required: Python environment with orchestration library installed
- Optional: Kubernetes cluster for distributed execution
- Optional: MLflow tracking server for experiment logging
- Optional: DVC for data versioning
- Optional: Slack/email for alerting
- Optional: Monitoring infrastructure (Prometheus, Grafana)
Procedure
Step 1: Choose and Install Orchestration Framework
Select appropriate framework and set up infrastructure.
# Option 1: Prefect (modern, Pythonic, simpler) pip install prefect pip install prefect-aws prefect-dask prefect-docker # Start Prefect server (local development) prefect server start # Or use Prefect Cloud (managed) # ... (see EXAMPLES.md for complete implementation)
Docker Compose for Airflow:
# docker-compose.airflow.yml version: '3.8' x-airflow-common: &airflow-common image: apache/airflow:2.8.0 environment: AIRFLOW__CORE__EXECUTOR: CeleryExecutor AIRFLOW__DATABASE__SQL_ALCHEMY_CONN: postgresql+psycopg2://airflow:airflow@postgres/airflow # ... (see EXAMPLES.md for complete implementation)
Expected: Orchestration framework installed, web UI accessible (Prefect at http://localhost:4200, Airflow at http://localhost:8080), database initialized, scheduler running.
On failure: Check port availability (
netstat -tulpn | grep 8080), verify database connection, ensure Redis running for Celery, check Python version compatibility (Airflow requires ≥3.8), verify Docker daemon for containerized setup, inspect logs for initialization errors.
Step 2: Build ML Pipeline with Prefect
Create Prefect flow with tasks for each pipeline stage.
# prefect_ml_pipeline.py from prefect import flow, task from prefect.tasks import task_input_hash from datetime import timedelta import pandas as pd import mlflow from sklearn.ensemble import RandomForestClassifier from sklearn.model_selection import train_test_split # ... (see EXAMPLES.md for complete implementation)
Deploy and schedule:
# deploy_prefect.py from prefect.deployments import Deployment from prefect.server.schemas.schedules import CronSchedule from prefect_ml_pipeline import ml_training_pipeline # Create deployment with schedule deployment = Deployment.build_from_flow( flow=ml_training_pipeline, # ... (see EXAMPLES.md for complete implementation)
Expected: Prefect flow executes all tasks in correct order, task failures trigger retries automatically, successful runs show green in UI, MLflow logs experiments, model registered and deployed.
On failure: Check task dependencies defined correctly, verify MLflow server accessible, ensure data source paths correct, check for circular dependencies, verify task timeout limits, inspect Prefect logs for detailed errors, check resource availability (memory/CPU).
Step 3: Build ML Pipeline with Airflow
Create Airflow DAG for production ML workflow.
# dags/ml_training_dag.py from airflow import DAG from airflow.operators.python import PythonOperator from airflow.providers.docker.operators.docker import DockerOperator from airflow.utils.dates import days_ago from datetime import datetime, timedelta import mlflow import pandas as pd # ... (see EXAMPLES.md for complete implementation)
Expected: DAG appears in Airflow UI, scheduled runs execute on time, task failures trigger retries and alerts, XCom passes data between tasks, MLflow integration logs experiments.
On failure: Check DAG file syntax (
python dags/ml_training_dag.py), verify imports available in Airflow environment, ensure XCom not exceeding size limits (use file paths for large data), check email configuration for alerts, verify scheduler running, inspect task logs in Airflow UI.
Step 4: Implement Advanced Features
Add dynamic DAGs, branching, and parallel execution.
# advanced_pipeline.py (Prefect) from prefect import flow, task from prefect.task_runners import DaskTaskRunner, ConcurrentTaskRunner import time @task def process_shard(shard_id: int, data: list) -> dict: """Process data shard in parallel.""" # ... (see EXAMPLES.md for complete implementation)
Airflow branching:
# Airflow branching with BranchPythonOperator from airflow.operators.python import BranchPythonOperator def check_data_quality(**context): """Decide which branch to take.""" data_path = context['ti'].xcom_pull(key='data_path') df = pd.read_csv(data_path) # ... (see EXAMPLES.md for complete implementation)
Expected: Parallel tasks execute concurrently (faster pipeline), conditional branches execute based on logic, dynamic task generation works, Dask cluster distributes work.
On failure: Check Dask cluster configured and accessible, verify task_runner specified, ensure branching returns valid task IDs, check for resource contention with parallel tasks, verify conditional logic correctness.
Step 5: Integrate Monitoring and Alerting
Add comprehensive monitoring and failure notifications.
# monitoring_integration.py from prefect.blocks.notifications import SlackWebhook from prefect import flow, task, get_run_logger from prefect.context import FlowRunContext @task(on_failure=[notify_failure]) def critical_task(): """Task with failure notification.""" # ... (see EXAMPLES.md for complete implementation)
Airflow monitoring with sensors:
# Airflow SLA and monitoring from airflow.sensors.base import BaseSensorOperator from airflow.utils.decorators import apply_defaults default_args = { 'sla': timedelta(hours=4), # Alert if task exceeds 4 hours 'on_failure_callback': slack_alert_failure, 'on_success_callback': slack_alert_success, # ... (see EXAMPLES.md for complete implementation)
Expected: Slack/email notifications sent on failures, SLA violations trigger alerts, custom metrics tracked, logs aggregated in monitoring system.
On failure: Verify Slack webhook configured correctly, check email SMTP settings, ensure notification blocks loaded properly, verify SLA values reasonable, check for network issues blocking notifications.
Step 6: Implement CI/CD for Pipelines
Version control and automate pipeline deployments.
# .github/workflows/deploy-pipeline.yml name: Deploy ML Pipeline locale: wenyan-lite source_locale: en source_commit: 82c77053 translator: "Julius Brussee homage — caveman" translation_date: "2026-04-19" on: push: branches: [main] paths: - 'pipelines/**' # ... (see EXAMPLES.md for complete implementation)
Expected: Pipeline tests pass before deployment, automated deployment to production, team notified on successful deployment, pipeline versioning tracked in Git.
On failure: Check test coverage and failures, verify Prefect Cloud credentials, ensure deployment script handles errors, check Slack webhook configuration, inspect CI logs for deployment errors.
Validation
- Orchestration framework installed and running
- Pipeline DAG defined with correct dependencies
- All tasks execute in proper order
- Retry logic functions correctly on failures
- Scheduled runs execute on time
- MLflow integration logs experiments
- DVC integration versions data
- Parallel tasks execute concurrently
- Conditional branches work correctly
- Monitoring and alerting functional
- CI/CD pipeline deploys automatically
- Pipeline reproducible across environments
Common Pitfalls
- Circular dependencies: Task A depends on B, B depends on A - carefully design DAG structure, use Airflow/Prefect validators
- Memory leaks: Long-running tasks accumulate memory - set task timeouts, monitor resource usage, restart workers periodically
- XCom size limits: Passing large data via XCom - use file paths or external storage (S3) instead of direct serialization
- Timezone confusion: Schedule runs at wrong times - always use UTC, explicitly set timezone in schedule
- Missing retries: Tasks fail permanently on transient errors - configure retries with exponential backoff
- Tight coupling: Tasks directly depend on implementation details - use clear interfaces, pass parameters explicitly
- No idempotency: Re-running tasks causes duplicates or errors - design tasks to be idempotent (safe to retry)
- Poor error handling: Failures don't provide useful context - add detailed logging, capture exceptions properly
- Resource contention: Parallel tasks overwhelm resources - limit concurrency, set resource quotas
- Version conflicts: Different tasks need incompatible dependencies - use Docker containers for task isolation
Related Skills
- Integrate MLflow tracking into pipeline taskstrack-ml-experiments
- Use DVC for data versioning in pipelinesversion-ml-data
- Materialize features as pipeline taskbuild-feature-store
- Add deployment as final pipeline stagedeploy-ml-model-serving
- Run orchestrated pipelines on Kubernetesdeploy-to-kubernetes