One codebase for all accelerators

Diving deeper into

Modular

Company Report
The same source code automatically runs on NVIDIA GPUs, AMD processors, CPUs, or future accelerators because the compiler retargets kernels for each hardware type.
Analyzed 4 sources

This portability is really a control point over the AI hardware stack. If developers can keep one codebase while the compiler emits different low level kernels for NVIDIA, AMD, CPUs, and newer accelerators, Modular can sit above the chip vendors and turn hardware choice into a deployment setting instead of an expensive rewrite project. That matters most for teams juggling scarce GPUs, mixed clusters, or on premises infrastructure.

  • In practice, the hard part is the last mile. NVIDIA has CUDA and TensorRT, AMD has ROCm, Intel has oneAPI and OpenVINO, and ONNX Runtime handles this today through separate execution providers. Modular is packaging that hardware adaptation into one compiler and runtime path, which is simpler for developers and easier to operationalize.
  • This also changes the economics for customers. Instead of maintaining separate optimization work for each target, the same stack can be deployed on laptop CPUs, data center GPUs, or future custom silicon. That is especially valuable for enterprises that buy whatever accelerators are available, cheapest, or already installed.
  • The closest alternatives each leave part of the workflow fragmented. ONNX Runtime is broadly portable across many hardware backends, while vLLM is a popular inference runtime with support across NVIDIA, AMD, Intel GPU, and CPU setups, but neither combines language, compiler, packaged serving, and orchestration in one stack the way Modular does.

This is heading toward a market where AI infrastructure is chosen more like cloud software and less like firmware tied to one chip. If Modular keeps making new accelerators usable on day one, it can become the layer that helps emerging hardware vendors reach developers and helps enterprises treat compute as a fluid pool rather than a fixed vendor bet.