Home  >  Companies  >  ScaleOps
ScaleOps
Tool for automating Kubernetes resource allocation to optimize cloud performance and cost

Funding

$80.00M

2025

View PDF
Details
Headquarters
New York, NY
CEO
Yodar Shafrir
Website
Milestones
FOUNDING YEAR
2022
Listed In

Valuation

ScaleOps raised a $58 million Series B in November 2024 led by Lightspeed Venture Partners, with participation from NFX, Glilot Capital Partners, and Picture Capital. This round brought the company's total funding to $80 million.

The company previously raised $21.5 million across seed and Series A rounds in December 2023, establishing its initial commercial foundation. The Series B came roughly one year after launch, following what the company described as tremendous business growth.

Product

ScaleOps functions as an autopilot system that sits inside Kubernetes clusters and continuously optimizes CPU, memory, replicas, nodes, and GPUs in real-time.

Teams install ScaleOps through a single Helm chart that creates a lightweight control plane in its own namespace. The system samples usage metrics every few seconds and makes all sizing decisions locally within the cluster, ensuring no data leaves the environment.

The platform performs real-time pod rightsizing by vertically adjusting CPU and memory requests and limits using rolling quantile windows and safety buffers. Most users reclaim 50-80% of over-provisioned capacity through this optimization.

ScaleOps also handles replica optimization by predicting demand spikes 10-15 minutes ahead of time and scaling horizontally before traffic hits. This integrates with existing HPA and KEDA systems to prevent SLA breaches during traffic bursts.

The smart pod placement feature bin-packs unevictable or PDB-restricted pods onto appropriate nodes, then consolidates under-utilized nodes or replaces them with cheaper instance types. The spot optimization capability keeps workloads on spot instances when safe and switches to on-demand instances when preemption risk increases.

In November 2025, ScaleOps launched its AI Infrastructure module, adding GPU workload optimization with dynamic GPU rightsizing, MIG partitioning, and LLM-aware memory tuning. Early adopters report 50-70% cuts in GPU spend while reducing inference latency by up to 35%.

The platform offers both self-hosted deployment for air-gapped environments and a cloud option that tunnels anonymized metadata to a multi-tenant SaaS interface for managing multiple clusters.

Business Model

ScaleOps operates as a B2B SaaS company selling Kubernetes optimization software to enterprise platform engineering teams. The company uses a subscription-based pricing model with tiers ranging from $500-$2,000+ per month for standard plans, plus custom enterprise pricing for larger deployments.

The business model centers on delivering immediate cost savings that typically exceed the software licensing fees. Customers often see 50-80% reductions in over-provisioned resources, creating a clear ROI justification for the platform.

ScaleOps differentiates itself through its self-hosted deployment option, which appeals to regulated industries like banking and defense that require air-gapped environments. This approach eliminates data privacy concerns while still providing automated optimization.

The company's go-to-market strategy focuses on platform engineering teams at cloud-native companies that already have significant Kubernetes deployments. The sales process is often driven by immediate cost pressure and the need to optimize cloud spending.

Revenue expansion occurs through both customer growth and increased usage as organizations deploy ScaleOps across more clusters and workloads. The addition of GPU optimization capabilities opens new upsell opportunities with AI-focused customers.

The business model benefits from high switching costs once deployed, as the platform becomes integrated into critical infrastructure operations. Customers typically expand usage rather than churn, given the ongoing cost savings delivered.

Competition

Specialized automation engines

Cast AI represents ScaleOps' closest competitor, operating an external control plane that replaces Cluster Autoscaler with strong capabilities in node recomposition and spot instance arbitrage. Cast AI has raised $108 million in Series D funding and serves 2,100 customers, though its usage-based pricing can erode savings at scale.

StormForge, now part of CloudBolt following a March 2025 acquisition, focuses on ML-driven pod-level rightsizing with per-vCPU SaaS pricing. Densify launched Kubex AI in November 2025, offering an AI chat assistant plus admission-controller automation to democratize rightsizing across infrastructure and GPUs.

Turbonomic, owned by IBM, provides policy-aware automation spanning both VMs and Kubernetes environments, leveraging IBM's broader enterprise software portfolio for distribution.

Cloud-native autoscalers

AWS Karpenter reached version 1.0 in August 2024 and continues absorbing advanced bin-packing features as an open-source solution. Rather than compete directly, ScaleOps now offers a Karpenter Optimizer add-on that enhances instance-type selection and disruption budgets for EKS users.

Google Cloud and Microsoft Azure offer their own native autoscaling solutions that integrate tightly with their respective platforms, creating competitive pressure for third-party optimization tools.

FinOps platforms

CloudZero, Apptio, and Vantage have expanded from cost visibility into optimization primitives, encroaching on the automation space. IBM's Apptio business bundled Kubecost 3.0 with Cloudability governance in November 2025, creating integrated FinOps suites.

These platforms compete by offering broader financial operations capabilities beyond just Kubernetes optimization, appealing to finance teams seeking comprehensive cloud cost management.

TAM Expansion

AI infrastructure optimization

ScaleOps launched its AI Infrastructure module in November 2025, expanding beyond CPU-based containers into GPU scheduling for self-hosted LLMs and GenAI workloads. Early pilots report 50-70% savings on GPU fleets with tens of millions in projected annual savings for large enterprises.

This move broadens ScaleOps' addressable market from the Kubernetes cost-optimization niche into the AI infrastructure segment. With half of all cloud compute expected to run AI workloads by 2029, this represents significant TAM expansion.

The AI module includes model performance optimization that keeps models warm, manages replica counts and weights files, and exposes GPU and LLM metrics in the same interface as traditional Kubernetes optimization.

Enterprise FinOps convergence

Large companies are merging FinOps, IT asset management, and SaaS spend programs, creating demand for tools that surface granular, allocation-ready cost data. ScaleOps' workload-level cost breakdown and GPU utilization analytics address this tooling gap.

This trend enables ScaleOps to sell jointly to platform engineering and finance buyers, expanding its reach within existing customers and opening new customer segments focused on financial operations.

The company's ability to provide detailed cost attribution at the namespace and team level aligns with enterprise needs for chargeback and showback capabilities.

Geographic and regulatory expansion

ScaleOps' self-hosted deployment model provides access to regulated regions including EU Sovereign Cloud, Middle East telecommunications, and Asia-Pacific banking sectors. These markets often require air-gapped or on-premises solutions that many competitors cannot address.

The company's ability to operate without external data transmission makes it suitable for government and defense contractors, as evidenced by its appeal to organizations with strict data sovereignty requirements.

International expansion opportunities exist in markets where data residency and privacy regulations favor self-hosted solutions over cloud-based alternatives.

Risks

Hyperscaler competition: AWS, Google Cloud, and Microsoft Azure continue enhancing their native autoscaling capabilities, potentially reducing demand for third-party optimization tools. As cloud providers integrate more sophisticated cost optimization directly into their platforms, ScaleOps may face pricing pressure and reduced differentiation, particularly among customers heavily committed to single-cloud strategies.

Market consolidation: The Kubernetes optimization space is experiencing rapid consolidation, with larger players like IBM acquiring specialized vendors and FinOps platforms adding optimization features. This trend could squeeze ScaleOps between well-funded competitors with broader product suites and integrated go-to-market capabilities, making it harder to compete for enterprise deals.

AI infrastructure complexity: While GPU optimization represents a major TAM expansion opportunity, the rapidly evolving AI infrastructure landscape introduces technical and competitive risks. ScaleOps must continuously adapt to new GPU architectures, AI frameworks, and deployment patterns while competing against specialized AI infrastructure companies with deeper domain expertise and stronger relationships with AI-native customers.

News

DISCLAIMERS

This report is for information purposes only and is not to be used or considered as an offer or the solicitation of an offer to sell or to buy or subscribe for securities or other financial instruments. Nothing in this report constitutes investment, legal, accounting or tax advice or a representation that any investment or strategy is suitable or appropriate to your individual circumstances or otherwise constitutes a personal trade recommendation to you.

This research report has been prepared solely by Sacra and should not be considered a product of any person or entity that makes such report available, if any.

Information and opinions presented in the sections of the report were obtained or derived from sources Sacra believes are reliable, but Sacra makes no representation as to their accuracy or completeness. Past performance should not be taken as an indication or guarantee of future performance, and no representation or warranty, express or implied, is made regarding future performance. Information, opinions and estimates contained in this report reflect a determination at its original date of publication by Sacra and are subject to change without notice.

Sacra accepts no liability for loss arising from the use of the material presented in this report, except that this exclusion of liability does not apply to the extent that liability arises under specific statutes or regulations applicable to Sacra. Sacra may have issued, and may in the future issue, other reports that are inconsistent with, and reach different conclusions from, the information presented in this report. Those reports reflect different assumptions, views and analytical methods of the analysts who prepared them and Sacra is under no obligation to ensure that such other reports are brought to the attention of any recipient of this report.

All rights reserved. All material presented in this report, unless specifically indicated otherwise is under copyright to Sacra. Sacra reserves any and all intellectual property rights in the report. All trademarks, service marks and logos used in this report are trademarks or service marks or registered trademarks or service marks of Sacra. Any modification, copying, displaying, distributing, transmitting, publishing, licensing, creating derivative works from, or selling any report is strictly prohibited. None of the material, nor its content, nor any copy of it, may be altered in any way, transmitted to, copied or distributed to any other party, without the prior express written permission of Sacra. Any unauthorized duplication, redistribution or disclosure of this report will result in prosecution.