HomeToolsMCPHow It WorksStoriesPhilosophyArchitectureStar on GitHub
All Tools
D
DataFreemiumOpen Source

DAGSTER

Orchestrate and observe data pipelines with software-engineering discipline

Apache-2.0

ABOUT

Data and AI teams need more than cron jobs and brittle scripts to run reliable pipelines. Dagster provides a structured orchestration layer for defining data assets, dependencies, schedules, and operational checks, so teams can build maintainable pipelines, understand lineage, and monitor data workflows without stitching together their own orchestration and observability stack.

INSTALL
pip install dagster dagster-webserver dagster-dg-cli

INTEGRATION GUIDE

1. Orchestrate ETL and ELT pipelines that move data into warehouses, lakes, and feature stores 2. Schedule and monitor data preparation workflows that feed RAG, analytics, or machine learning systems 3. Manage lineage and health checks for production data assets across teams and environments 4. Coordinate batch inference, evaluation, or retraining pipelines for AI applications 5. Build local-to-production data workflows with testing, retries, and operational visibility

TAGS

orchestrationdata-pipelinesworkflowschedulinglineagedata-assetsobservabilitypython
Dagster — AI Tool | Agentic AI For Good