Sakana as Cost-Aware AI Orchestrator
Sakana AI
The important shift is that Sakana can make money from the traffic around model use, not just from selling a model itself. AB-MCTS turns inference into a routing problem, where Sakana decides which model should handle each step of a task, when to try more branches, and when to stop. That makes Sakana valuable even if a customer already pays OpenAI, Google, or DeepSeek directly, because the savings come from choosing the cheapest model that still gets the job done.
-
AB-MCTS is built to mix models at inference time. Sakana describes it as a system that lets frontier models cooperate on one problem, and shows combinations like o4-mini, Gemini-2.5-Pro, and DeepSeek-R1 beating the standalone models on ARC-AGI-2. That is the technical basis for a paid routing layer.
-
The closest business analogue is OpenRouter. It monetizes by sitting between developers and many model providers, taking a roughly 5% markup on inference spend, while also supporting bring-your-own-key so customers keep existing provider relationships and still buy routing, billing, and analytics as a separate layer.
-
Recent partnerships point toward enterprise deployment, not just lab demos. Sakana has announced a multiyear deal with MUFG Bank and a 2026 partnership with Datadog focused on building and operating AI systems at scale for large enterprises in Japan. That fits an orchestration product that needs production monitoring, cost control, and compliance workflows.
This is heading toward AI middleware. If enterprises end up using several models for different jobs, the winning layer will be the one that measures cost, quality, latency, and reliability in real time, then routes traffic automatically. Sakana is building the technical ingredients to own that control point, and usage based orchestration revenue can scale much faster than one off research licensing.