dbt as the Analytics Workflow
Diving deeper into
Julia Schottenstein, Product Manager at dbt Labs, on the business model of open source
while dbt is infrastructure, it’s also a workflow tool which is harder to copy.
Analyzed 4 sources
Reviewing context
dbt is hardest to displace where it becomes the place a team actually does its work, not just the engine that runs SQL. Warehouses can copy transformation features, but copying the daily loop of writing models, reviewing changes, running tests, scheduling jobs, and keeping business definitions consistent across dozens of people is much harder. That is why dbt has pushed beyond core transformation into IDE, CI, orchestration, catalog, and semantic layer products.
-
dbt’s commercial wedge was ease of use for analytics teams, not managed hosting alone. In large companies, many users do not want local dev environments, they want a browser based workflow with guardrails. That makes the paid product a team productivity layer, closer to GitLab’s model than pure infrastructure resale.
-
The practical moat is embedded process. dbt brings software habits like version control, tests, pull requests, and CI into analytics work, then makes them usable for analysts as well as engineers. Once a team’s review flow, documentation, and metric definitions live there, replacing dbt means retraining both tools and habits.
-
This is also the defense against clouds. Snowflake, Databricks, and hyperscalers can bundle native transformation features, but they are tied to their own stacks. dbt’s advantage is sitting above the warehouse as a cross cloud control layer, which matters more as customers run mixed environments and want one workflow across them.
The next stage is a broader data control plane. If dbt keeps owning how analytics engineers and analysts define, review, ship, and govern data work across clouds, it can grow from a transformation standard into the default operating system for enterprise analytics teams.