
Valuation
$1.60B
2025
Funding
$380.00M
2025
Valuation
Modular closed a $250 million Series C in September 2025 at a $1.6 billion post-money valuation. The round was led by the U.S. Innovative Technology Fund with participation from existing investors including DFJ Growth, GV (Google Ventures), General Catalyst, and Greylock.
The company previously raised a $100 million Series B in August 2023 at approximately a $600 million valuation. Before that, Modular secured a $30 million seed round in June 2022.
Altogether, Modular has raised $380 million in primary equity funding since its founding.
Product
Modular provides a three-layer software stack that lets developers build, deploy, and scale AI applications across any hardware without rewriting code for specific chips.
The foundation is Mojo, a programming language that looks and feels like Python but compiles to produce machine code performance comparable to C or CUDA. Developers can write compute-intensive operations directly in Mojo or gradually optimize existing Python code by adding type annotations. The same source code automatically runs on NVIDIA GPUs, AMD processors, CPUs, or future accelerators because the compiler retargets kernels for each hardware type.
The MAX Engine sits in the middle as a compiler and runtime that takes models from TorchScript, ONNX, or Mojo Graph formats and optimizes them into a single deployable package. The max serve command creates an OpenAI-compatible HTTP endpoint that can run over 500 pre-optimized models including Llama, Mistral, and Whisper variants. Performance benchmarks show roughly 10-50% faster inference than alternatives like vLLM on identical hardware.
At the top, the Mammoth orchestrator acts as a Kubernetes-native scheduler that maintains over 90% GPU cluster utilization by intelligently routing requests and co-locating models. It exposes a batch inference API that's compatible with OpenAI's format, making it easy for developers to switch from other providers.
The entire stack runs in Docker containers with only GPU drivers as a host requirement, enabling the same deployment to work on laptop CPUs, data center GPUs, or cloud instances without modification.
Business Model
Modular operates as a B2B infrastructure software company with a consumption-based revenue model. Customers pay based on compute resources consumed and inference requests processed through the MAX engine, similar to cloud computing pricing but optimized for AI workloads.
The company takes a revenue-sharing approach with cloud partners, earning a percentage of customer spending on platforms that integrate Modular's technology. This creates a scalable distribution model where Modular benefits from increased AI deployment without directly managing customer relationships in every case.
The business model creates strong unit economics because the same software stack serves multiple hardware vendors and cloud providers. Rather than maintaining separate codebases for NVIDIA, AMD, and Intel architectures, Modular's compiler automatically generates optimized kernels for each platform from a single source.
Modular's approach differs from traditional AI infrastructure companies by positioning itself as hardware-neutral rather than optimizing for a single chip vendor. This creates value for customers who want to avoid vendor lock-in and for hardware companies who want software support without building their own complete toolchains.
The company plans to expand into AI training workloads, which could significantly increase average contract values as training typically requires more compute resources and longer-term commitments than inference workloads.
Competition
Chip vendor stacks
NVIDIA dominates with its vertically integrated CUDA ecosystem, TensorRT optimization libraries, and Triton kernel language. The company's recent acquisitions of OctoAI and Run:ai bring formerly independent optimization and orchestration tools under NVIDIA's control, threatening Modular's hardware-neutral positioning.
AMD has accelerated development of ROCm with support for popular model formats and partnerships with Microsoft Azure for MI300X deployments. Intel's oneAPI and OpenVINO platforms provide similar capabilities for Intel hardware, while Arm's KleidiAI integration with ONNX Runtime targets edge deployment scenarios.
These vendors benefit from tight hardware-software integration but create customer lock-in that Modular explicitly aims to avoid.
Independent optimization engines
Companies like Neural Magic, Deci, and the creators of vLLM focus on inference optimization without hardware vendor ties. Apache TVM provides open-source compilation for multiple backends, though with less commercial support than Modular's offering.
ONNX Runtime serves as a widely adopted inference engine with broad hardware support, but lacks the unified development experience that Modular provides through Mojo and MAX.
These competitors often excel in specific areas like quantization or attention mechanisms but require customers to integrate multiple tools rather than providing a complete stack.
Cloud provider platforms
AWS Neuron, Google XLA/JAX, and Azure's AI infrastructure bundle proprietary compilers with cloud-native silicon. These platforms offer tight integration with cloud services but limit portability across providers.
The major cloud vendors have significant advantages in distribution and can bundle AI infrastructure with other enterprise services, but they typically don't support on-premises or multi-cloud deployments as seamlessly as Modular's approach.
TAM Expansion
Training workloads
Modular's latest funding explicitly targets expansion from AI inference into training, which could double the company's addressable market from the current $12-15 billion inference market into the $20+ billion training market expected by 2027.
Training workloads typically involve longer-term contracts and higher compute consumption than inference, potentially increasing average contract values significantly. The same hardware portability benefits apply to training, where customers want to optimize costs across different GPU types and availability.
Enterprise orchestration
The Mammoth orchestrator and batch inference APIs position Modular to capture higher-value enterprise deals comparable to data platform companies. Selling cluster-level scheduling and quality-of-service management unlocks usage-based revenue that scales with customer growth.
Large enterprises with mixed hardware environments represent a particularly valuable segment, as they benefit most from Modular's hardware abstraction and can justify premium pricing for simplified operations.
Edge and specialized hardware
Mojo's ability to generate kernels for CPUs and custom ASICs opens opportunities in automotive, mobile, and industrial IoT markets. The growing $8 billion edge AI silicon market represents a significant expansion beyond data center deployments.
Custom silicon vendors like those in automotive or mobile devices often lack comprehensive software toolchains, making Modular's compiler technology valuable for enabling AI applications on new hardware architectures.
Risks
NVIDIA dominance: NVIDIA's 80%+ market share in AI accelerators gives them significant leverage to bundle software tools and create switching costs for customers. Their recent acquisitions of independent optimization companies directly threaten Modular's positioning as a neutral alternative, and NVIDIA's ability to optimize performance at the hardware level may prove difficult to match through software alone.
Execution complexity: Building a compiler, runtime, and orchestration platform that works reliably across diverse hardware requires deep technical expertise and extensive testing. Any performance regressions or compatibility issues could quickly erode customer confidence, especially when competing against mature, hardware-specific solutions that have been optimized for years.
Market timing: The AI infrastructure market is evolving rapidly, and customers may not yet feel sufficient pain from vendor lock-in to justify switching to a new platform. If hardware standardization occurs naturally or if dominant vendors improve their cross-platform support, the value proposition for hardware-neutral solutions could diminish before Modular reaches scale.
News
DISCLAIMERS
This report is for information purposes only and is not to be used or considered as an offer or the solicitation of an offer to sell or to buy or subscribe for securities or other financial instruments. Nothing in this report constitutes investment, legal, accounting or tax advice or a representation that any investment or strategy is suitable or appropriate to your individual circumstances or otherwise constitutes a personal trade recommendation to you.
This research report has been prepared solely by Sacra and should not be considered a product of any person or entity that makes such report available, if any.
Information and opinions presented in the sections of the report were obtained or derived from sources Sacra believes are reliable, but Sacra makes no representation as to their accuracy or completeness. Past performance should not be taken as an indication or guarantee of future performance, and no representation or warranty, express or implied, is made regarding future performance. Information, opinions and estimates contained in this report reflect a determination at its original date of publication by Sacra and are subject to change without notice.
Sacra accepts no liability for loss arising from the use of the material presented in this report, except that this exclusion of liability does not apply to the extent that liability arises under specific statutes or regulations applicable to Sacra. Sacra may have issued, and may in the future issue, other reports that are inconsistent with, and reach different conclusions from, the information presented in this report. Those reports reflect different assumptions, views and analytical methods of the analysts who prepared them and Sacra is under no obligation to ensure that such other reports are brought to the attention of any recipient of this report.
All rights reserved. All material presented in this report, unless specifically indicated otherwise is under copyright to Sacra. Sacra reserves any and all intellectual property rights in the report. All trademarks, service marks and logos used in this report are trademarks or service marks or registered trademarks or service marks of Sacra. Any modification, copying, displaying, distributing, transmitting, publishing, licensing, creating derivative works from, or selling any report is strictly prohibited. None of the material, nor its content, nor any copy of it, may be altered in any way, transmitted to, copied or distributed to any other party, without the prior express written permission of Sacra. Any unauthorized duplication, redistribution or disclosure of this report will result in prosecution.