Customers
Case Study: Artemis - Powering the Crypto Markets

Case Study: Artemis - Powering the Crypto Markets

July 29, 2024
Case Study: Artemis - Powering the Crypto Markets

Artemis built a data platform around Dagster+ to bring consolidated reporting to the $2.5T Cryptocurrency markets.

Alex Kan, software engineer at Artemis

We explore how Artemis, a premier institutional data platform for digital asset fundamentals, leverages Dagster+ to serve thousands of customers, from financial institutions to individual investors and crypto protocols.

We got to chat with Alex Kan, Software Engineer at Artemis, and a member of the 12-person team, half of which are heavy Dagster users, including data engineers and data scientists.      

In 2008, a paper was published called "Bitcoin: A Peer-to-Peer Electronic Cash System," by Satoshi Nakamoto.  Since then the world has witnessed the wild and sometimes rocky ascent of cryptocurrency.

From its early days on the fringes of the digital economy, crypto has established itself as a legitimate and significant part of the global financial system. As of 2024, there are over 22,000 different cryptocurrencies in existence, of which 9,124 are considered active [1]. The total market capitalization for the cryptocurrency sector now exceeds $2.5 trillion USD.

Total aggregate cryptocurrency value, worldwide, as provided by Artemis.
   Total Crypto Market Cap Chart total market cap & volume of cryptocurrencies globally, a result of 14,951 cryptocurrencies tracked across 1,161 exchanges. Source: https://www.coingecko.com/en/global-charts

Stablecoins have also seen massive growth, with transfer volume peaking at over $150B a day in May 2024.

Total Stablecoin value, as provided by Artemis.
   Source: https://app.artemis.xyz/stablecoins

The cryptocurrency protocol stack generates a lot of public (onchain) data, but accessing that data in a consumable fashion is quite complex. There is no public repository of fundamental metrics like you might find for a centralized exchange like the NASDAQ or the NYSE.

And unlike traditional capital markets, which have structured quarterly earnings reports, crypto, makes real-time reporting possible.  Anybody can do diligence on a protocol's financials by accessing the public blockchain.

Blockchain nodes are the backbone of blockchain networks, responsible for validating transactions and maintaining the blockchain ledger. This has made it easier to participate in the cryptocurrency ecosystem, , but financial reporting is still complex.

But indexing the raw data from blockchains like Ethereum and Solana is still complex.  These datasets are often dozens of TB in size.

An ecosystem of reporting firms has set about making real-time crypto data available to the masses. Artemis is one such crypto data provider.  Think of it as Bloomberg for the crypto markets.

Artemis: “The Bloomberg of Crypto”

The Artemis UI.
   The Artemis UI.  

The Artemis team aggregates data from many blockchains in the crypto space and provides the cleansed, structured, aggregated data to a range of data consumers, including institutions in traditional finance, research analysts and crypto native investors at liquid token hedge funds, and retail investors. Artemis also works closely with the crypto protocols themselves who use their data in their workflows.

Artemis delivers its data through its Artemis Terminal, an Artemis API, and data shares on Snowflake. In addition, Artemis provides Excel and Google Sheets plugin integration to pull data directly into a spreadsheet.

Most equity research analysts or quants do not have data engineering backgrounds or a data platform team, so the service Artemis provides is highly valuable to them, making crypto data instantly consumable.

Currently, Artemis provides dozens of terabytes of data to several thousand customers. The larger blockchains have thousands of transactions per second (TPS), while smaller networks may only see a dozen TPS. The architecture of each chain also determines how big each transaction record is and what data gets logged.

The complexity is not only in the volumes of data, but rather the lack of standards across blockchains. Aggregating and benchmarking the data across providers takes a lot of data processing.  Different chains provide different features (like smart contracts). And unlike traditional financial markets, there is no “market close”—crypto is traded 24/7/365.

Behind the scenes, Artemis taps into the indexing services that are now commoditized in the industry, with many providers who will run the blockchain nodes—the backbone of blockchain networks— as a service and index the data.

An MDS tech stack

Broadly speaking, Artemis leverages an MDS stack, favoring composable, often open-source technologies.

  • Most of the data is ingested directly from Snowflake Shares, Snowflake’s marketplace for datasets from a variety of data providers
  • Where data is ingested from other APIs or a node for an unsupported chain, this is done directly in Dagster
  • Data modeling is done in dbt core
  • Jobs run on AWS’s ECS service
  • Hightouch is used for reverse ETL
  • BigQuery is also used for some more niche data sources
  • Hex is used for data sandboxing and reporting
  • Django and Next.js then support the front-end app

This streamlined stack lets the team iterate and ship quickly across the entire application.

Moving off Apache Airflow

Artemis ran Airflow as a managed service for the first year of its existence.  They chose to run it on MWAA because they were scrappy and were committed to AWS. However, the system fell short on reliability (due to the Airflow scheduler), and the non-engineers who contributed to the code base (such as data scientists and researchers) found the system hard to work with.

Anthony Yim, Co-founder and CTO at Artemis, further explains: “What was also tough is managing data lineages especially when a model failed to build. We would have to rerun the whole lineage rather than a specific asset.”

“Debugging failures was hard since there was no single pane of glass to view assets and statuses.In addition, the UI on Airflow often just didn't work or was very cryptic. Finally, as an organization that needed to step up its data pipeline game, we wanted to use DBT, and Airflow didn't have a tight integration with DBT tools (at least not back then).”

Another challenge of Airflow: the lack of a dbt integration.

It was clear that Airflow would not support the team's next stage of growth and was increasingly adding drag to their development cycles. The team looked for a more fit-for-purpose system that provided better lineage, tagging, and a more intuitive framework for pipelines.

Building on Dagster+

The Dagster framework, with its asset orientation, proved much more intuitive and a better fit.

“Software-defined Assets were a novel approach.” says Alex, “Also, not having to write custom Airflow Operators bc of Dagster's builtin I/O managers was huge.”

The tagging and built-in cataloging and lineage make data collaboration much smoother, and medium-code practitioners (people with a data science or analytics background and focus who nonetheless write code) found Dagster’s framework much more accessible.

Again, because they lack a full-time DevOps team, Artemis selected a hosted option in Dagster+, providing many enhancements around user management, cataloging, and cost observability.

While Alex joined the team after the transition, he had prior experience working on and contributing to Dagster and was able to rapidly bring new value to the team. He expanded the Dagster+ setup to multiple code locations and organized assets in groups for better scheduling and alerting.

Opening the black box: open source dbt models

Analysts and investors naturally need to trust the data they are working with. To address concerns and answer analyst questions about methodology, Artemis took the bold step of publishing the open-source code to their dbt models, which you can find here.      

The open-sourced dbt models by Artemis

This approach of openness aligns with the culture of decentralized digital currency, and other projects have also openly shared their business logic, such as Dune’s Spellbook.

A platform for growth

Today the Artemis team has a streamlined data operation that will support their growth for the coming years, as digital assets continue to expand, become more mainstream, and continue to evolve. Whatever digital currency the world decides to adopt next, Alex and the Artemis team will be ready to help investors grasp the fundamentals and make informed investment decisions.

Have feedback or questions? Start a discussion in Slack or Github.

Interested in working with us? View our open roles.

Want more content like this? Follow us on LinkedIn.

Dagster Newsletter

Get updates delivered to your inbox

Latest writings

The latest news, technologies, and resources from our team.

Multi-Tenancy for Modern Data Platforms
Webinar

April 7, 2026

Multi-Tenancy for Modern Data Platforms

Learn the patterns, trade-offs, and production-tested strategies for building multi-tenant data platforms with Dagster.

Deep Dive: Building a Cross-Workspace Control Plane for Databricks
Webinar

March 24, 2026

Deep Dive: Building a Cross-Workspace Control Plane for Databricks

Learn how to build a cross-workspace control plane for Databricks using Dagster — connecting multiple workspaces, dbt, and Fivetran into a single observable asset graph with zero code changes to get started.

Dagster Running Dagster: How We Use Compass for AI Analytics
Webinar

February 17, 2026

Dagster Running Dagster: How We Use Compass for AI Analytics

In this Deep Dive, we're joined by Dagster Analytics Lead Anil Maharjan, who demonstrates how our internal team utilizes Compass to drive AI-driven analysis throughout the company.

DataOps with Dagster: A Practical Guide to Building a Reliable Data Platform
DataOps with Dagster: A Practical Guide to Building a Reliable Data Platform
Blog

March 17, 2026

DataOps with Dagster: A Practical Guide to Building a Reliable Data Platform

DataOps is about building a system that provides visibility into what's happening and control over how it behaves

Unlocking the Full Value of Your Databricks
Unlocking the Full Value of Your Databricks
Blog

March 12, 2026

Unlocking the Full Value of Your Databricks

Standardizing on Databricks is a smart strategic move, but consolidation alone does not create a working operating model across teams, tools, and downstream systems. By pairing Databricks and Unity Catalog with Dagster, enterprises can add the coordination layer needed for dependency visibility, end-to-end lineage, and faster, more confident delivery at scale.

Announcing AI Driven Data Engineering
Announcing AI Driven Data Engineering
Blog

March 5, 2026

Announcing AI Driven Data Engineering

AI coding agents are changing how data engineers work. This Dagster University course shows how to build a production-ready ELT pipeline from prompts while learning practical patterns for reliable AI-assisted development.

How Magenta Telekom Built the Unsinkable Data Platform
Case study

February 25, 2026

How Magenta Telekom Built the Unsinkable Data Platform

Magenta Telekom rebuilt its data infrastructure from the ground up with Dagster, cutting developer onboarding from months to a single day and eliminating the shadow IT and manual workflows that had long slowed the business down.

Scaling FinTech: How smava achieved zero downtime with Dagster
Case study

November 25, 2025

Scaling FinTech: How smava achieved zero downtime with Dagster

smava achieved zero downtime and automated the generation of over 1,000 dbt models by migrating to Dagster's, eliminating maintenance overhead and reducing developer onboarding from weeks to 15 minutes.

Zero Incidents, Maximum Velocity: How HIVED achieved 99.9% pipeline reliability with Dagster
Case study

November 18, 2025

Zero Incidents, Maximum Velocity: How HIVED achieved 99.9% pipeline reliability with Dagster

UK logistics company HIVED achieved 99.9% pipeline reliability with zero data incidents over three years by replacing cron-based workflows with Dagster's unified orchestration platform.