Blog
Dagster’s MCP Server

Dagster’s MCP Server

August 8, 2025
Dagster’s MCP Server
Dagster’s MCP Server

We are announcing the release of our MCP server, enabling AI assistants like Cursor to seamlessly integrate with Dagster projects through Model Context Protocol, unlocking composable workflows across your entire data stack.

When Anthropic announced MCP in late 2024, the initial reaction was one of cautious optimism. AI was evolving so quickly that it wasn’t immediately clear what MCP would mean in practice. But as its implications became clearer, even competitors like OpenAI adopted the protocol.

Since then, other companies have released their own MCP servers. At Dagster, we built our first MCP server earlier this year but chose to wait before releasing it, ensuring it would integrate seamlessly with our recent platform improvements.

Today, we’re excited to make the Dagster MCP server publicly available. It complements everything we’ve been building over the years and opens up new possibilities for writing better, more maintainable code.

What is MCP?

The Model Context Protocol (MCP) is an open standard developed by Anthropic to connect large language models (LLMs) with external data sources in a secure, standardized way. In Anthropic’s own words:

The Model Context Protocol is an open standard that enables developers to build secure, two-way connections between their data sources and AI-powered tools. It provides a universal, open standard for connecting AI systems with data sources, replacing fragmented integrations with a single protocol. The result is a simpler, more reliable way to give AI systems access to the data they need.

LLMs are powerful at generating coherent language, but on their own, they:

  • Lack real-time, accurate, and domain-specific knowledge.
  • Have no native ability to interact with databases, APIs, or tools.
    Rely entirely on what they were trained on which becomes outdated quickly.

To remain relevant and useful, LLMs must pull information from external sources. When an LLM “searches the web” or queries a knowledge base, it’s functioning as an AI agent, reasoning about what data it needs, where to get it, and how to integrate it into its output.

This approach is powerful but creates challenges:

  • Complexity: Each LLM must manage custom integrations for every data source.
  • Fragility: Data sources change, breaking integrations over time.

MCP standardizes the connection between AI agents and external services. Instead of every AI system learning how to talk to every data source, the external service provides an MCP interface.

Benefits of this approach:

  • Interoperability: Any MCP-compatible AI can instantly integrate with the service.
  • Reduced Maintenance: The service owner maintains the MCP interface, ensuring accuracy and reliability.
  • Future-proofing: As AI systems evolve, the MCP layer stays consistent.

MCP shifts the integration burden from the AI agent to the service, making the ecosystem simpler, more reliable, and easier to maintain.

Dagster’s MCP Server

So where does Dagster’s MCP fit into all of this, and why does it work so well with our recent features?

If you’ve been following Dagster this year, you know we’re excited about dg and Dagster Components. These abstractions sit on top of core Dagster objects (such as assets) and make it much easier to build new features or integrate with tools and workflows, often with minimal code.

This combination, a rich core library that can do just about anything, paired with opinionated tooling, is a perfect match for MCP. With a well-documented, structured library, MCP can gain a deep understanding of Dagster’s capabilities. By exposing streamlined code interfaces with tighter guardrails, you can have more confidence in the quality and safety of the code it returns.

Let’s look at a real-world example. Say you’re using Cursor and want to start a new Dagster project.

Can you scaffold me a new Dagster project named example-project

With the Dagster MCP configured, Cursor knows where to route requests and translates them to the correct `dg` CLI commands as needed. In this case Cursor would know that it is more efficient to use `uvx -U create-dagster` to generate a project than generating all the code itself.

Composability

What makes the Dagster MCP exciting is that it’s not a one-off integration, it’s composable. If you’re using other tools in your data stack, like dbt, Snowflake, Airbyte, or any service that exposes an MCP interface, your AI assistant can seamlessly interact with all of them together.

Can you add dbt to my Dagster project

Without `dg` and components, many LLMs could handle this request by generating code directly. But that approach comes with higher risk: more potential for errors, plus the added cost of processing large amounts of context.

With Dagster MCP, the AI instead guides you toward simple `dg` commands to scaffold an integration for a dbt project. This generates just a few lines of YAML needed to configure what is needed. From there, you can use the AI agent to make any necessary changes at the YAML layer, keeping your code clean, concise, and maintainable.

This interoperability unlocks a powerful new way of working: one where AI agents operate across your entire stack, coordinating multiple tools through a shared, universal protocol. Instead of stitching together brittle point-to-point integrations or ad-hoc scripts, you get a unified, extensible interface for building and automating sophisticated workflows, accelerating development while maintaining flexibility and control.

Trying it out

The best part? You don’t need to be a Cursor user to benefit from the Dagster MCP. As more AI tools adopt MCP, the value of having an MCP-compatible interface to your project will only grow. Standardization in the AI space means better interoperability, more reliable automation, and faster iteration.

You can install the MCP server right now through the dg[mcp] package, and easily enable it within your tool using dg mcp configure.

Have feedback or questions? Start a discussion in Slack or Github.

Interested in working with us? View our open roles.

Want more content like this? Follow us on LinkedIn.

Dagster Newsletter

Get updates delivered to your inbox

Latest writings

The latest news, technologies, and resources from our team.

Multi-Tenancy for Modern Data Platforms
Webinar

April 7, 2026

Multi-Tenancy for Modern Data Platforms

Learn the patterns, trade-offs, and production-tested strategies for building multi-tenant data platforms with Dagster.

Deep Dive: Building a Cross-Workspace Control Plane for Databricks
Webinar

March 24, 2026

Deep Dive: Building a Cross-Workspace Control Plane for Databricks

Learn how to build a cross-workspace control plane for Databricks using Dagster — connecting multiple workspaces, dbt, and Fivetran into a single observable asset graph with zero code changes to get started.

Dagster Running Dagster: How We Use Compass for AI Analytics
Webinar

February 17, 2026

Dagster Running Dagster: How We Use Compass for AI Analytics

In this Deep Dive, we're joined by Dagster Analytics Lead Anil Maharjan, who demonstrates how our internal team utilizes Compass to drive AI-driven analysis throughout the company.

Monorepos, the hub-and-spoke model, and Copybara
Monorepos, the hub-and-spoke model, and Copybara
Blog

April 3, 2026

Monorepos, the hub-and-spoke model, and Copybara

How we configure Copybara for bi-directional syncing to enable a hub-and-spoke model for Git repositories

Making Dagster Easier to Contribute to in an AI-Driven World
Making Dagster Easier to Contribute to in an AI-Driven World
Blog

April 1, 2026

Making Dagster Easier to Contribute to in an AI-Driven World

AI has made contributing to open source easier but reviewing contributions is still hard. At Dagster, we’re improving the contributor experience with smarter review tooling, clearer guidelines, and a focus on contributions that are easier to evaluate, merge, and maintain.

DataOps with Dagster: A Practical Guide to Building a Reliable Data Platform
DataOps with Dagster: A Practical Guide to Building a Reliable Data Platform
Blog

March 17, 2026

DataOps with Dagster: A Practical Guide to Building a Reliable Data Platform

DataOps is about building a system that provides visibility into what's happening and control over how it behaves

How Magenta Telekom Built the Unsinkable Data Platform
Case study

February 25, 2026

How Magenta Telekom Built the Unsinkable Data Platform

Magenta Telekom rebuilt its data infrastructure from the ground up with Dagster, cutting developer onboarding from months to a single day and eliminating the shadow IT and manual workflows that had long slowed the business down.

Scaling FinTech: How smava achieved zero downtime with Dagster
Case study

November 25, 2025

Scaling FinTech: How smava achieved zero downtime with Dagster

smava achieved zero downtime and automated the generation of over 1,000 dbt models by migrating to Dagster's, eliminating maintenance overhead and reducing developer onboarding from weeks to 15 minutes.

Zero Incidents, Maximum Velocity: How HIVED achieved 99.9% pipeline reliability with Dagster
Case study

November 18, 2025

Zero Incidents, Maximum Velocity: How HIVED achieved 99.9% pipeline reliability with Dagster

UK logistics company HIVED achieved 99.9% pipeline reliability with zero data incidents over three years by replacing cron-based workflows with Dagster's unified orchestration platform.

Modernize Your Data Platform for the Age of AI
Guide

January 15, 2026

Modernize Your Data Platform for the Age of AI

While 75% of enterprises experiment with AI, traditional data platforms are becoming the biggest bottleneck. Learn how to build a unified control plane that enables AI-driven development, reduces pipeline failures, and cuts complexity.

Download the eBook on how to scale data teams
Guide

November 5, 2025

Download the eBook on how to scale data teams

From a solo data practitioner to an enterprise-wide platform, learn how to build systems that scale with clarity, reliability, and confidence.

Download the e-book primer on how to build data platforms
Guide

February 21, 2025

Download the e-book primer on how to build data platforms

Learn the fundamental concepts to build a data platform in your organization; covering common design patterns for data ingestion and transformation, data modeling strategies, and data quality tips.