
Valuation
$1.10B
2025
Funding
$111.00M
2025
Valuation
Modal Labs closed an $87 million Series B in September 2025 led by Lux Capital, valuing the company at $1.1 billion post-money.
The company previously raised a $16 million Series A in October 2023 led by Redpoint Ventures, following a $7 million seed round in early 2022 led by Amplify Partners. Other notable investors include Definition Capital and various prominent angel investors who have participated across multiple rounds.
Modal has raised approximately $111 million in total funding since its founding.
Product
Modal Labs provides a serverless computing platform that turns any Python function into an autoscaling cloud workload with GPU acceleration. Developers import the Modal library, wrap functions with decorators like `@app.function()`, and call `.remote()` to execute code in the cloud with sub-second cold starts.
The platform automatically builds containers from local environments, schedules them onto specified CPU or GPU hardware, and streams logs back to developers in real-time. Modal's custom Rust-based container runtime, image builder, and distributed file system enable the fast startup times that differentiate it from traditional cloud platforms.
Key building blocks include Functions for single-process units with autoscaling controls, Sandboxes for isolated container execution of untrusted code, web endpoints that expose functions as HTTPS APIs, and built-in storage through Volumes and Buckets. The platform can scale from zero to thousands of containers or hundreds of GPUs in seconds, with multi-cloud capacity automatically selected from AWS, GCP, and Oracle Cloud Infrastructure.
Modal recently launched browser-based Notebooks with GPU memory snapshots that reduce startup times by 10x, and introduced clustered computing for multi-node, RDMA-connected GPU workloads. The platform supports everything from simple batch jobs to complex AI inference pipelines and agent sandboxing.
Business Model
Modal operates as a developer-first cloud infrastructure platform with a consumption-based SaaS model. The company charges customers by the second for actual compute usage, eliminating the need for long-term commitments or idle resource costs that plague traditional cloud providers.
The go-to-market approach is primarily B2B, targeting machine learning engineers, AI researchers, and development teams at both startups and enterprises. Modal's Python-native SDK creates a low-friction adoption path where developers can start with simple functions and scale to production workloads without architectural changes.
Revenue scales directly with customer compute consumption, creating natural expansion as teams deploy more workloads and larger models. The platform's multi-cloud architecture allows Modal to optimize costs by automatically selecting the most cost-effective GPU capacity across providers, while customers benefit from simplified billing and management.
Modal's partnership with Oracle Cloud Infrastructure provides access to competitively priced GPU resources, while marketplace integrations with major cloud providers allow enterprises to apply existing committed spend. This hybrid approach reduces customer acquisition friction while maintaining margin flexibility across different infrastructure costs.
Competition
Serverless GPU platforms
Direct competitors include Baseten, which raised $150M Series D and focuses on mission-critical inference with dedicated deployments, and Replicate, which serves over 25,000 paying customers through its Series B funding and open-source model library approach. RunPod competes on price with up to 40% GPU cost savings and configurable serverless workers, while Together AI raised $305M Series B at $3.3B valuation to build an AI acceleration cloud on NVIDIA Blackwell GPUs.
Modal differentiates through sub-second cold starts and pure Python integration, while competitors like Baseten rely on REST APIs or YAML configurations. The pricing models vary significantly, with Modal's per-second billing competing against RunPod's per-pod pricing and Baseten's per-GPU-second plus bandwidth charges.
Hyperscaler integration
AWS SageMaker added scale-to-zero capabilities with up to 45% GPU price cuts, directly competing with Modal's pay-per-use promise while bundling into existing enterprise contracts. Google Cloud Run and Azure AI Studio offer similar serverless GPU primitives integrated with their broader cloud ecosystems, creating competitive pressure through existing customer relationships and committed spend programs.
GPU infrastructure specialists
CoreWeave and Lambda Labs focus on providing raw GPU capacity that developers can access programmatically, competing on the infrastructure layer that Modal abstracts away. These companies benefit from direct hardware relationships and can offer lower-level control, while Modal competes on developer experience and operational simplicity.
TAM Expansion
Product suite expansion
Modal has evolved from a functions-as-a-service platform to a comprehensive AI infrastructure suite including Inference, Sandboxes, Batch processing, Training capabilities, and browser-based Notebooks. These modules address the complete machine learning lifecycle from exploration to production, multiplying attachment points within customer organizations and increasing average contract values.
The addition of clustered computing capabilities opens opportunities in high-performance computing workloads like protein folding and weather modeling, expanding beyond traditional AI inference into adjacent compute-intensive markets.
Customer base expansion
Modal serves thousands of ML teams including high-profile customers like Meta's Code World Models team and fast-growing AI startups. The company's startup credit program and academic partnerships create a substantial funnel for future enterprise conversions, while geographic expansion through offices in New York, San Francisco, and Stockholm provides footholds in key markets.
Enterprise adoption accelerates through cloud marketplace integrations that allow customers to apply existing AWS, Azure, and GCP committed spend to Modal services, removing budget allocation friction.
Adjacent market penetration
The platform's ability to handle any embarrassingly parallel or GPU-bound workload opens opportunities in generative media rendering, agent sandboxing, and traditional high-performance computing applications. Modal's multi-cloud architecture and Oracle Cloud Infrastructure partnership provide the foundation for rapid geographic expansion without owning data centers.
Strategic partnerships with IDE vendors and data catalog companies could channel more interactive workloads to Modal Notebooks, increasing daily active developers and platform stickiness across the broader developer ecosystem.
Risks
GPU supply constraints: Modal's growth depends on access to scarce GPU resources across cloud providers, and supply shortages or price increases from NVIDIA and other chip manufacturers could impact margins and availability. The company's multi-cloud strategy provides some protection, but industry-wide capacity constraints could limit scaling ability during peak demand periods.
Hyperscaler competition: AWS, Google Cloud, and Microsoft Azure are rapidly adding serverless GPU capabilities with scale-to-zero billing and tight integration with existing enterprise contracts. These platforms can leverage existing customer relationships and committed spend programs to compete aggressively on price while offering broader service ecosystems that Modal cannot match.
Market commoditization: As serverless GPU computing becomes more standardized, Modal's current advantages in cold start times and developer experience may erode. Competitors are investing heavily in similar capabilities, and the risk of the market becoming commoditized on price rather than features could compress margins and reduce Modal's differentiation over time.
News
DISCLAIMERS
This report is for information purposes only and is not to be used or considered as an offer or the solicitation of an offer to sell or to buy or subscribe for securities or other financial instruments. Nothing in this report constitutes investment, legal, accounting or tax advice or a representation that any investment or strategy is suitable or appropriate to your individual circumstances or otherwise constitutes a personal trade recommendation to you.
This research report has been prepared solely by Sacra and should not be considered a product of any person or entity that makes such report available, if any.
Information and opinions presented in the sections of the report were obtained or derived from sources Sacra believes are reliable, but Sacra makes no representation as to their accuracy or completeness. Past performance should not be taken as an indication or guarantee of future performance, and no representation or warranty, express or implied, is made regarding future performance. Information, opinions and estimates contained in this report reflect a determination at its original date of publication by Sacra and are subject to change without notice.
Sacra accepts no liability for loss arising from the use of the material presented in this report, except that this exclusion of liability does not apply to the extent that liability arises under specific statutes or regulations applicable to Sacra. Sacra may have issued, and may in the future issue, other reports that are inconsistent with, and reach different conclusions from, the information presented in this report. Those reports reflect different assumptions, views and analytical methods of the analysts who prepared them and Sacra is under no obligation to ensure that such other reports are brought to the attention of any recipient of this report.
All rights reserved. All material presented in this report, unless specifically indicated otherwise is under copyright to Sacra. Sacra reserves any and all intellectual property rights in the report. All trademarks, service marks and logos used in this report are trademarks or service marks or registered trademarks or service marks of Sacra. Any modification, copying, displaying, distributing, transmitting, publishing, licensing, creating derivative works from, or selling any report is strictly prohibited. None of the material, nor its content, nor any copy of it, may be altered in any way, transmitted to, copied or distributed to any other party, without the prior express written permission of Sacra. Any unauthorized duplication, redistribution or disclosure of this report will result in prosecution.