Immuta's Deep Connector Strategy
Zachary Friedman, associate director of product management at Immuta, on security in the modern data stack
This choice says Immuta is trying to win like infrastructure, not like a broad governance dashboard. In data security, shallow coverage across many systems breaks down at the exact moment customers need it most, when analysts want to run normal Snowflake or Databricks queries without switching tools or waiting on security teams. Going deep on a few core warehouses lets Immuta turn policy into native controls, audit trails, and low friction daily use.
-
Immuta sells to large companies running multiple cloud data platforms. Its core value is that a team can write one business policy, like who can see customer PII, and have it enforced across Snowflake, Databricks, BigQuery, Redshift, and Starburst instead of rebuilding rules store by store. That only works if each connector is very deep.
-
The practical standard for depth is zero change management. Instead of forcing users through a proxy or separate query path, Immuta uses native warehouse features like row and column level controls so analysts keep querying the same tables in the same tool, but only see the rows and fields they are allowed to see.
-
This also explains the contrast with broader data security players. BigID is centered on finding and classifying sensitive data across many systems, while Immuta started with enforcing live access rules inside analytical warehouses. As Databricks and Snowflake became the center of the modern data stack, that narrower focus became more valuable, not less.
The next step is a tighter bundle around those same systems. Once Immuta owns enforcement inside the main warehouse, products like Detect and Discover can layer on top, spotting overprovisioned access, classifying sensitive columns, and feeding fixes back into the policy engine. That turns a deep connector strategy into a broader data security platform over time.