Storage as AI Control Plane

Diving deeper into

Renen Hallak, CEO of VAST Data, on AI agents creating infinite storage demand

Interview
These new AI companies didn't want to build applications in the old way of writing to a data platform and reading from it.
Analyzed 3 sources

This marks a shift from storage as a passive bucket to data infrastructure as the control plane for AI work. In the older pattern, an app wrote data into one system, moved it to another for indexing or analytics, then called a model separately. VAST is arguing that AI labs want one layer where files, embeddings, metadata, and compute all live together, so new data can immediately trigger training, retrieval, fine tuning, or inference without shuffling it across separate tools.

  • That is why VAST moved beyond file and object storage into a database and orchestration layer. The same platform now stores raw files, holds trillion row embedding tables, and handles millions of parallel agent queries, which is the workload shape of retrieval heavy agent systems rather than classic BI dashboards.
  • The practical contrast is with the early LLM stack, where developers often paired a model API with a separate vector database like Pinecone. That stack works for search and RAG, but it still treats storage, retrieval, and application logic as separate components. VAST is trying to collapse those components into one substrate underneath the application.
  • This bundling also explains VAST's expansion path. The company says the share of customers using only storage fell from about 70% two years ago to about 30% today, while the bundled platform rose to about 70%, showing that AI customers are buying more of the stack once their workloads move from simple data access to always on agent workflows.

The next step is that AI infrastructure vendors will compete less on raw storage speed and more on whether they can become the default runtime for data driven applications. If VAST keeps turning storage events, metadata, vector search, and policy controls into one system, it moves closer to replacing parts of the warehouse, ETL, and orchestration stack around AI workloads.