Blog
The Missing Half of the Enterprise Context Layer

The Missing Half of the Enterprise Context Layer

April 22, 2026
The Missing Half of the Enterprise Context Layer
The Missing Half of the Enterprise Context Layer

AI agents that only understand business definitions without knowing whether the underlying pipeline actually succeeded are confidently wrong and operational context from the orchestrator is the missing piece.

Atlan is hosting Activate on April 29, where it will make major announcements about the Enterprise Context Layer. Dagster is proud to be a Context Layer Partner. To learn more about the Dagster + Atlan integration, check out our integration docs.

There's a gap in how the industry talks about making AI agents enterprise-ready. Most of the conversation focuses solely on semantic context: business definitions, glossary terms, metric logic, and governance rules. This makes sense. An AI agent that doesn't know the difference between "net revenue" and "gross revenue" is going to produce bad answers.

But semantic context is only part of the picture. There are other dimensions that get far less attention, but are just as critical: data context (lineage, quality signals, technical metadata), knowledge context (institutional rules, edge cases, decision history), governance context (policies, sensitivity, access controls), user context (who’s asking and why), and our focus in this article, operational context. Operational context is the runtime reality of your data platform. Did the pipeline that produces net_revenue actually run today? Did it succeed? When did it last materialize? What upstream dependencies fed into it, and are they healthy?

These are questions only an orchestrator can answer. And right now, most AI agents can't ask them.

Multiple Kinds of Context, One Blind Spot

Think about how an experienced data engineer evaluates whether a number is trustworthy. They don't just check the definition. They check the pipeline, the users, and the data itself. They look at the orchestrator to see when the asset last materialized, whether the run succeeded, and whether any upstream sources are lagging. They have the tribal knowledge to spot any exceptions or anomalies. They have a mental model of the platform's operational state that informs their judgment.

AI agents don't have that mental model. They query a warehouse and get a table. They query a catalog and get a definition. Both are valuable. But neither tells them whether the data they're looking at is fresh, complete, or the product of a healthy pipeline run.

This creates a specific failure mode: the agent that gives a confident, well-defined, semantically correct answer based on stale or broken data. The definition is right. The number is wrong. And nobody knows until someone downstream gets burned.

Where Operational Context Comes From

Operational context is generated by the systems that actually execute your data workflows. It includes:

Materialization events. Did this asset get produced? When? Was it a full refresh or a partition update? Did it succeed or fail?

Dependency graphs. What upstream assets feed into this one? If a source table lags, every downstream asset may be affected. That's not something a catalog knows on its own.

Freshness signals. Is this asset within its expected SLA? A table that refreshes daily but hasn't been touched in 36 hours is a problem, but only if you know what "on time" looks like for that asset.

Run metadata. Run IDs, timestamps, partition keys, retry counts, error messages. This is the operational provenance that tells you not just what the data is, but what happened to produce it.

This metadata is ephemeral by nature. It's generated at runtime by the orchestrator. If you don't capture it somewhere durable and expose it alongside your semantic metadata, it disappears, and your context layer has a hole in it.

Why Asset-Centric Orchestration Changes the Equation

Not all orchestrators produce the same quality of operational context. Task-based systems know that "Job 47 ran at 3am." That's useful for debugging, but it doesn't connect cleanly to the data assets your AI agents and analysts actually care about.

Dagster takes a different approach. Every data asset is a first-class citizen in the system, with its own materialization history, freshness expectations, dependency relationships, and quality check results. This isn't just a different UI. It's a fundamentally richer metadata model.

When a Dagster asset materializes, the system knows exactly what was produced, what produced it, what it depends on, and whether it met its quality and freshness criteria. That's structured operational context, and it maps directly to Atlan’s Enterprise Context Layer. The operational context, along with business definitions, lineage, quality signals, governance policies, and the full data graph, gives AI agents everything they need to act with confidence.

The Context Layer Has To Be Open and Operational

A context layer that only captures what data means is incomplete. For AI agents to act reliably on enterprise data, context has to be open and operable, meaning it reflects the live state of the data estate, not a stale snapshot.

Orchestrators are context producers. Every time Dagster materializes an asset, it generates a piece of context that belongs in the Enterprise Context Layer: which pipelines ran, when, whether they succeeded, what they produced, and how upstream dependencies behaved. That operational context — alongside business semantics, ontologies, governance policies, and decision traces — is what gives AI agents the complete picture they need to reason correctly. Without it, an agent can know what a metric means but not whether it's current. That's not enough for production.

Dagster pushes operational context into Atlan's Enterprise Context Layer to provide the context enterprises actually need at inference time. Materialization events, run status, asset lineage, partition keys, freshness signals, and custom tags flow into Atlan in real time as assets materialize or fail, with no batch sync or manual export needed. Atlan unifies that operational signal with the full Enterprise Context Graph: business semantics, ontologies, ownership, governance policies, sensitivity classification, and the decision traces that explain why data is governed the way it is.

Neither system alone gives AI agents the full picture. A context layer without operational signals can tell an agent what a metric means but not whether it's current. An orchestrator without a context layer can tell you a pipeline succeeded but not what the output represents in business terms.. The value is in the combination.

The Dagster + Atlan integration makes this concrete. Dagster streams operational metadata into Atlan's context layer in real time: materialization events, run status, asset lineage, partition keys, timestamps, descriptions, and custom tags. As assets materialize (or fail), events flow automatically. No batch sync, no manual export. Atlan unifies Dagster’s operational signals with the full Enterprise Context Graph, and serves the combined context to every AI tool you run via MCP.

This means that when an AI agent or a human analyst queries an asset in Atlan, they see the full context, its business definition, column-level lineage, data quality signals, governance policies, sensitivity classification, and ownership – alongside its operational state. They know what the asset means, where it came from, whether it passed its quality checks, who’s allowed to use it for what purpose, and that it was last materialized at 6:14am with all upstream dependencies healthy. That's a context layer with actual depth.

This Gets Harder as Unstructured Data Grows

So far, most of the context layer conversation has been grounded in structured, warehouse-native data: tables, columns, metrics, and dbt models. But the share of enterprise data that's unstructured is growing fast. PDFs, images, transcripts, embeddings, and documents are fed into RAG pipelines. AI workloads are pulling from sources that don't have neat schemas or well-established catalog entries.

This makes operational context even more important. When an AI agent reasons over a vector store populated by an embedding pipeline, there's no column-level lineage to fall back on. There's no glossary defining what "quarterly_earnings_chunk_47" means. The questions that matter are operational ones: When was this data last ingested? What source documents were processed? Did the embedding pipeline complete, or did it fail halfway through and leave the index in a partial state?

The data platforms that can track these workflows end to end, across structured and unstructured sources, and produce useful metadata at every step are the ones that will keep the context layer honest as AI workloads expand. An orchestrator that treats every asset the same way, whether it's a Snowflake table, a dbt model, or a vector database containing processed documents, provides your context layer with a single, consistent source of operational truth. That consistency matters more as the stack gets more heterogeneous, not less.

Context Layers Need Operational Depth

The teams that successfully move AI from pilot to production will be the ones that treat context as infrastructure, not an afterthought. Semantic definitions tell agents what data means. Operational signals tell them whether to trust it right now.

Orchestrators are uniquely positioned to provide that signal. No other system in the data stack has the same view into what ran, what succeeded, what failed, and what's at risk. By streaming that context into a unified layer alongside business definitions, governance rules, and lineage, you give AI agents the full picture they need to act reliably.

The context layer conversation has rightly centered on business meaning. It's time to bring operational reality into it.

Have feedback or questions? Start a discussion in Slack or Github.

Interested in working with us? View our open roles.

Want more content like this? Follow us on LinkedIn.

Dagster Newsletter

Get updates delivered to your inbox

Latest writings

The latest news, technologies, and resources from our team.

Multi-Tenancy for Modern Data Platforms
Webinar

April 13, 2026

Multi-Tenancy for Modern Data Platforms

Learn the patterns, trade-offs, and production-tested strategies for building multi-tenant data platforms with Dagster.

Deep Dive: Building a Cross-Workspace Control Plane for Databricks
Webinar

March 24, 2026

Deep Dive: Building a Cross-Workspace Control Plane for Databricks

Learn how to build a cross-workspace control plane for Databricks using Dagster — connecting multiple workspaces, dbt, and Fivetran into a single observable asset graph with zero code changes to get started.

Dagster Running Dagster: How We Use Compass for AI Analytics
Webinar

February 17, 2026

Dagster Running Dagster: How We Use Compass for AI Analytics

In this Deep Dive, we're joined by Dagster Analytics Lead Anil Maharjan, who demonstrates how our internal team utilizes Compass to drive AI-driven analysis throughout the company.

The Missing Half of the Enterprise Context Layer
The Missing Half of the Enterprise Context Layer
Blog

April 22, 2026

The Missing Half of the Enterprise Context Layer

AI agents that only understand business definitions without knowing whether the underlying pipeline actually succeeded are confidently wrong and operational context from the orchestrator is the missing piece.

How to Orchestrate Across Multiple Databricks Workspaces Without Losing Your Mind
How to Orchestrate Across Multiple Databricks Workspaces Without Losing Your Mind
Blog

April 20, 2026

How to Orchestrate Across Multiple Databricks Workspaces Without Losing Your Mind

Once your pipelines span multiple Databricks workspaces, you're no longer orchestrating a single system you're coordinating a distributed one.

Dagster 1.13: Octopus's Garden
Dagster 1.13: Octopus's Garden
Blog

April 9, 2026

Dagster 1.13: Octopus's Garden

Dagster skills, partitioned asset checks, state backed components, virtual assets, and stronger integrations.

How Magenta Telekom Built the Unsinkable Data Platform
Case study

February 25, 2026

How Magenta Telekom Built the Unsinkable Data Platform

Magenta Telekom rebuilt its data infrastructure from the ground up with Dagster, cutting developer onboarding from months to a single day and eliminating the shadow IT and manual workflows that had long slowed the business down.

Scaling FinTech: How smava achieved zero downtime with Dagster
Case study

November 25, 2025

Scaling FinTech: How smava achieved zero downtime with Dagster

smava achieved zero downtime and automated the generation of over 1,000 dbt models by migrating to Dagster's, eliminating maintenance overhead and reducing developer onboarding from weeks to 15 minutes.

Zero Incidents, Maximum Velocity: How HIVED achieved 99.9% pipeline reliability with Dagster
Case study

November 18, 2025

Zero Incidents, Maximum Velocity: How HIVED achieved 99.9% pipeline reliability with Dagster

UK logistics company HIVED achieved 99.9% pipeline reliability with zero data incidents over three years by replacing cron-based workflows with Dagster's unified orchestration platform.

Modernize Your Data Platform for the Age of AI
Guide

January 15, 2026

Modernize Your Data Platform for the Age of AI

While 75% of enterprises experiment with AI, traditional data platforms are becoming the biggest bottleneck. Learn how to build a unified control plane that enables AI-driven development, reduces pipeline failures, and cuts complexity.

Download the eBook on How to Scale Data Teams
Guide

November 5, 2025

Download the eBook on How to Scale Data Teams

From a solo data practitioner to an enterprise-wide platform, learn how to build systems that scale with clarity, reliability, and confidence.

Download the eBook Primer on How to Build Data Platforms
Guide

February 21, 2025

Download the eBook Primer on How to Build Data Platforms

Learn the fundamental concepts to build a data platform in your organization; covering common design patterns for data ingestion and transformation, data modeling strategies, and data quality tips.

AI Driven Data Engineering
Course

March 19, 2026

AI Driven Data Engineering

Learn how to build Dagster applications faster using AI-driven workflows. You'll use Dagster's AI tools and skills to scaffold pipelines, write quality code, and ship data products with confidence while still learning the fundamentals.

Dagster & ETL
Course

July 11, 2025

Dagster & ETL

Learn how to ingest data to power your assets. You’ll build custom pipelines and see how to use Embedded ETL and Dagster Components to build out your data platform.

Testing with Dagster
Course

April 21, 2025

Testing with Dagster

In this course, learn best practices for testing, including unit tests, mocks, integration tests and applying them to Dagster.