Blog
The Dagster Master Plan

The Dagster Master Plan

June 15, 2023
The Dagster Master Plan
The Dagster Master Plan

Elementl CEO Pete Hunt shares the three priorities that guide how we will evolve Dagster.

We didn't start the Dagster project simply to make a great orchestrator. Our plan is to accelerate the adoption of software engineering best practices by every data team on the planet.

Data and ML engineering drive important decisions that influence billions of people and trillions of dollars worldwide, and all too often, the data pipelines backing these decisions are held together with duct tape and chewing gum. Furthermore, engineering teams are drowning in both technical and organizational complexity found inside modern data-driven organizations.

We believe that embracing software engineering best practices is the only way for these teams to move faster and maximize quality and have happier developers.

As orchestration sits at the center of the increasingly-complex data platform, it is the natural place for us to drive this change.

So far, we have delivered an orchestrator that has the reputation of having a great (the best?) developer experience due to our innovative core programming model (Software-defined Assets), superior CI/CD capabilities, and first-class local development and testing. However, we have much more work to do.

Nick Schrock, our founder and CTO, recently shared his perspectives on the state of organizational complexity, and Dagster’s role in helping to tame it.  In the spirit of open-source and transparency, I would like to lay out what’s next for Dagster and the company behind it.

Priority #1: Flatten the learning curve

Dagster has a reputation for being extremely powerful. However, the learning curve is still too steep.

One reason for this is that Dagster is actually three things rolled into one: it’s a scheduler, an operational asset catalog, and a data transformation tool. We believe that these are naturally all part of the orchestrator and that this will become more obvious over time as the category matures. However, in the short term, this means that there are a lot of different concepts to learn. Historically we have introduced them all at once, which has been challenging for some newcomers.

The good news is that most people don't need to learn most of these concepts in order to be productive with Dagster. And when you do need to reach for these concepts, they layer in nicely with the rest of the system.

Over the next few quarters, we're going to focus our efforts on creating a streamlined onboarding experience. This means we're going to emphasize a subset of our features focused on Software-defined Assets. Additionally, we're going to make improvements to these essential SDA features to ensure they can scale to the broad set of needs of our users and are easy to reason about for those who are just getting started with Dagster. I want to emphasize that these will be primarily small fixes and polish work; we don't anticipate any major deprecations or fundamental changes to the system.

Priority #2: Evolve the orchestration category

One innovation that we've brought to orchestration is a unified asset graph that combines lineage, metadata, and operational history into a single system of record. We call this an operational asset catalog, and it's critical to our long-term strategy and differentiation.

We have only scratched the surface of the value that this can provide. As more metadata accrues to this graph, we can provide a ton more value to users.

  • Consumption management. We can keep track of the resources that are consumed every time an asset is materialized. Not only can we report on and monitor costs, we can put proactive guardrails in place to ensure that teams don't accidentally consume more resources than budgeted.
  • Data quality. Improved observability and monitoring will help data engineers ensure that their data pipelines are delivering high-quality data. Additionally, incorporating data quality checks into the orchestrator's decision-making will help prevent bad data from spreading down the asset graph.
  • Sandboxes. We can build a “forkable” version of the asset graph (like “git branch” for data assets). When combined with Dagster's branch deployments and branchable storage layers like Snowflake’s zero-copy clones or LakeFS, this will allow data engineers to create "copy on write" staging environments for every pull request that read from production and write to staging. This will massively improve developer velocity, especially for ML model development
  • Programmatic governance. We can integrate governance constraints directly into the orchestrator, ensuring that properties like data sovereignty and retention are programmatically enforced across the whole organization.

Priority #3: Accelerate our early commercial success

We are a venture backed open-source company, so naturally, we often get questions about how we prioritize open-source vs commercial work.

Fundamentally, the success of the Dagster OSS project is prerequisite to the success of our commercial offering. Our strategy is to build a kick-ass, category-redefining orchestrator that, over time, becomes the standard. On top of that strong foundation, we're building a company that sells a hosted version of Dagster with features that larger organizations would otherwise need to build themselves.

We have been selling Dagster+, our commercial product, for a little under a year, and our commercial traction validates our strategy. We're confident that if we're successful in evolving the orchestration category and establishing a new standard, we can build a great, venture-scale company which in turn can continue to build an amazing open-source project.

tl;dr

In essence, our master plan is as follows:

  1. Make Dagster both increasingly powerful and easier to use for all data practitioners.
  2. Integrate new types of metadata into the operational asset catalog.
  3. Deliver features above and beyond what's currently considered “orchestration” by leveraging this metadata.
  4. Roll out paid versions of these features with extra capabilities for teams and enterprises like access control, auditing, etc.
  5. Keep doing this until software engineering best practices are widely adopted by every data team in the world.

We look forward to hearing your reactions to this. Join our community and be part of the conversation.

— Pete

Have feedback or questions? Start a discussion in Slack or Github.

Interested in working with us? View our open roles.

Want more content like this? Follow us on LinkedIn.

Dagster Newsletter

Get updates delivered to your inbox

Latest writings

The latest news, technologies, and resources from our team.

Multi-Tenancy for Modern Data Platforms
Webinar

April 13, 2026

Multi-Tenancy for Modern Data Platforms

Learn the patterns, trade-offs, and production-tested strategies for building multi-tenant data platforms with Dagster.

Deep Dive: Building a Cross-Workspace Control Plane for Databricks
Webinar

March 24, 2026

Deep Dive: Building a Cross-Workspace Control Plane for Databricks

Learn how to build a cross-workspace control plane for Databricks using Dagster — connecting multiple workspaces, dbt, and Fivetran into a single observable asset graph with zero code changes to get started.

Dagster Running Dagster: How We Use Compass for AI Analytics
Webinar

February 17, 2026

Dagster Running Dagster: How We Use Compass for AI Analytics

In this Deep Dive, we're joined by Dagster Analytics Lead Anil Maharjan, who demonstrates how our internal team utilizes Compass to drive AI-driven analysis throughout the company.

Dagster 1.13: Octopus's Garden
Dagster 1.13: Octopus's Garden
Blog

April 9, 2026

Dagster 1.13: Octopus's Garden

Dagster skills, partitioned asset checks, state backed components, virtual assets, and stronger integrations.

Monorepos, the hub-and-spoke model, and Copybara
Monorepos, the hub-and-spoke model, and Copybara
Blog

April 3, 2026

Monorepos, the hub-and-spoke model, and Copybara

How we configure Copybara for bi-directional syncing to enable a hub-and-spoke model for Git repositories

Making Dagster Easier to Contribute to in an AI-Driven World
Making Dagster Easier to Contribute to in an AI-Driven World
Blog

April 1, 2026

Making Dagster Easier to Contribute to in an AI-Driven World

AI has made contributing to open source easier but reviewing contributions is still hard. At Dagster, we’re improving the contributor experience with smarter review tooling, clearer guidelines, and a focus on contributions that are easier to evaluate, merge, and maintain.

How Magenta Telekom Built the Unsinkable Data Platform
Case study

February 25, 2026

How Magenta Telekom Built the Unsinkable Data Platform

Magenta Telekom rebuilt its data infrastructure from the ground up with Dagster, cutting developer onboarding from months to a single day and eliminating the shadow IT and manual workflows that had long slowed the business down.

Scaling FinTech: How smava achieved zero downtime with Dagster
Case study

November 25, 2025

Scaling FinTech: How smava achieved zero downtime with Dagster

smava achieved zero downtime and automated the generation of over 1,000 dbt models by migrating to Dagster's, eliminating maintenance overhead and reducing developer onboarding from weeks to 15 minutes.

Zero Incidents, Maximum Velocity: How HIVED achieved 99.9% pipeline reliability with Dagster
Case study

November 18, 2025

Zero Incidents, Maximum Velocity: How HIVED achieved 99.9% pipeline reliability with Dagster

UK logistics company HIVED achieved 99.9% pipeline reliability with zero data incidents over three years by replacing cron-based workflows with Dagster's unified orchestration platform.

Modernize Your Data Platform for the Age of AI
Guide

January 15, 2026

Modernize Your Data Platform for the Age of AI

While 75% of enterprises experiment with AI, traditional data platforms are becoming the biggest bottleneck. Learn how to build a unified control plane that enables AI-driven development, reduces pipeline failures, and cuts complexity.

Download the eBook on How to Scale Data Teams
Guide

November 5, 2025

Download the eBook on How to Scale Data Teams

From a solo data practitioner to an enterprise-wide platform, learn how to build systems that scale with clarity, reliability, and confidence.

Download the eBook Primer on How to Build Data Platforms
Guide

February 21, 2025

Download the eBook Primer on How to Build Data Platforms

Learn the fundamental concepts to build a data platform in your organization; covering common design patterns for data ingestion and transformation, data modeling strategies, and data quality tips.