Lambda Labs

View PDF

Revenue

Sacra estimates that Lambda Labs hit a $500M revenue run rate in May 2025, up from $425M in December 2024. The company’s trajectory reflects accelerating adoption of its cloud GPU rental business, which nearly doubled year-over-year in the first half of 2025 and now contributes the bulk of revenue, while legacy hardware sales make up a shrinking portion.

Lambda’s positioning emphasizes competitive pricing—Nvidia H100 PCIe instances at roughly $2.49 per hour compared to $4.25 at CoreWeave—helping drive utilization and expand its customer base of developers and enterprises needing cost-efficient AI compute.

Gross margin year-to-date has been about 50%, or ~61% excluding non-cloud lines, with net losses narrowing as scale improves.

Valuation & Funding

Lambda raised over $1.5B in a Series E led by TWG Global (Thomas Tull and Mark Walter), with participation from US Innovative Technology Fund and other investors in November 2025.

The round brings Lambda’s total funding to $2.3B.

The company was previously valued at $2.5 billion following its $480 million Series D in February 2025, which was co-led by Andra Capital and SGW with participation from Nvidia, ARK Invest, G Squared, and Super Micro.

Notable investors include Nvidia, ARK Invest, Andrej Karpathy, and Bloomberg Beta.

Product

In 2013, Lambda Labs launched its first product—a facial recognition API for Google Glass that let developers build AI-powered apps that could do things like remember faces and track down a specific face in a crowd, creating controversy and driving more than 5 million API calls every month.

In 2017, Lambda Labs pivoted to selling hardware built for AI and deep learning alongside their Face Recognition API, including a dedicated GPU laptop (the TensorBook), and physical workstations with 4 GPUs for more intensive training.

Lambda Labs sold their enterprise workstations and servers to customers like Amazon, Apple, Raytheon, and MIT, each one being pre-configured with TensorFlow, Pytorch, and Caffe to make it quicker to get started. The main customers of Lambda Labs were researchers working on the emerging fields of image recognition and speech generation, alongside early natural language processing.

Today, rebranded as Lambda (lambda.ai), the company is primarily a cloud platform that lets developers get access to on-demand, enterprise-grade cloud GPUs for AI development.

The scarcity of GPU compute has been one of the key themes in the AI boom of 2023. Massive cloud providers like AWS have had trouble meeting demand on their cloud platforms due to shortages in chips upstream at TSMC, creating an opportunity for new cloud providers like CoreWeave and Lambda Labs that have beneficial relationships with Nvidia.

Lambda's cloud product has expanded well beyond single-node instances into a full-spectrum compute portfolio. The company offers H100 SXM instances at $2.99 per GPU-hour and production-scale "1-Click Clusters" on HGX H100 and HGX B200—including multi-node HGX B200 clusters offering on-demand access to 16–512 B200 GPUs connected via NVIDIA Quantum-2 InfiniBand—narrowing the gap with CoreWeave on high-end training workloads while keeping per-GPU pricing competitive for smaller teams. Looking further out, Lambda has announced plans to deploy NVIDIA Vera Rubin NVL72 systems as a core building block for Lambda Superclusters, with production availability targeted for 2H 2026, positioning the company at the leading edge of each successive GPU generation as it comes to market.

Business Model

Lambda Labs, like CoreWeave in the GPUs-on-demand space and like other cloud providers such as AWS, operates on a model where it rents out computing resources (such as GPU power) to businesses and developers over the cloud.

One key differentiation with Lambda Labs is that where CoreWeave is cloud-only, Lambda Labs also has on-premises options. Companies that want to get more compute for less money, ensure data security, and work on big datasets can purchase Lambda workstations without using them as a cloud provider.

The breadth of Lambda's customer base is underscored by the fact that Nvidia itself has agreed to rent back chips from Lambda—reported at roughly $1.5B total, covering approximately $1.3B over four years for 10,000 chips plus a separate $200M deal for 8,000 more. Lambda has also begun shifting from leasing third-party data center space toward owning and building its own facilities, a move that reduces long-term per-unit costs and increases control over infrastructure quality and capacity planning.

Expenses

Lambda Labs incurs a significant upfront cost when purchasing GPUs and setting up data centers. However, these GPUs have a useful life of several years, during which Lambda Labs can continually rent them out. The operational costs include electricity (GPUs are power-hungry), cooling (to prevent overheating), and staffing (for maintenance and customer support).

Improving the efficiency of data center operations (e.g., reducing electricity consumption, negotiating better rates for electricity, or improving cooling systems) can lower operational costs and thus improve margins.

Margin

The cost of a GPU for Lambda Labs includes the purchase price and the operational costs over its lifespan. The revenue from a GPU is the cumulative amount paid by customers to rent the GPU over time. Lambda Labs aims to maximize the utilization of each GPU to ensure that the revenue generated far exceeds the cost.

Margins are generally lowest on Lambda Labs's higher-end GPUs. For example, a high-end H100 PCIe card might cost Lambda Labs roughly $30,000. That GPU is then rented out at an average of $2.49 per hour. Assuming an 80% utilization rate, it would generate roughly $17,268 in revenue per year ($2.49/hour 19 hours/day 365 days/year).

However, cheaper GPUs like the A10 could generate much greater margins. At 80% utilization, an A10—which Lambda Labs could have bought for ~$3,500 a few years ago and is now rented out at $0.75 per hour—could generate $5,201 in revenue every year.

Competition

None

The market for GPU cloud services is highly competitive, with several key players, including major cloud providers like Amazon Web Services, Google Cloud and Azure as well as upstarts like Lambda Labs and Together AI, each offering unique advantages and targeting different segments of the AI and machine learning industry.

Big Cloud

The biggest long-term competition for Lambda Labs is likely to be the major three cloud providers: Google Cloud ($75B in revenue in 2023), Amazon Web Services ($80B in revenue in 2023) and Microsoft Azure ($26B in revenue in 2023).

With far greater revenue scale—vs. Lambda Labs's ~$250M in 2023—the big cloud platforms have the resources to invest both in acquiring GPUs and in developing their own silicon alternatives to Nvidia's GPUs.

However, between Lambda and the big cloud companies are also some "coopetitive" dynamics. In November 2025, Lambda announced a multibillion‑dollar partnership with Microsoft to build AI infrastructure powered by tens of thousands of Nvidia chips, including NVIDIA GB300 NVL72 systems.

CoreWeave

Like Lambda Labs, CoreWeave is a public cloud provider that purchases GPUs from Nvidia and rents them out to AI companies and companies building AI features.

Lambda Labs is generally positioning itself as a better option for smaller companies and developers working on less intensive computational tasks, offering Nvidia H100 PCIe GPUs at a price of roughly $2.49 per hour, compared to CoreWeave at $4.25 per hour. On the other hand, Lambda Labs does not offer access to the more powerful HGX H100—$27.92 per hour for a group of 8 at CoreWeave—which is designed for maximum efficiency in large-scale AI workloads.

Together

Together is fundamentally a GPU reseller that rents GPUs from providers like Lambda Labs and CoreWeave, big cloud platforms like Google Cloud, and from other sources—academic institutions, crypto miners, other companies—and then rents those GPUs out to startups and AI companies, then bundling that in with software for training and fine-tuning open source AI models like Meta's Llama 2, Midjourney's Stable Diffusion, and its own RedPajama.

Sacra estimates that Together hit $10M in annual revenue run rate at the end of 2023, with 90% of that revenue coming from Forge, their bundled compute-and-training product that launched in June 2023. Forge promises A100 and H100 Nvidia server clusters at 20% of the cost of AWS.

TAM Expansion

Looking forward, the key dynamics in understanding Lambda Labs's durable advantage hinges on (1) the long-term state of the GPU industry, (2) the market for on-premises AI hardware, and (3) Lambda's own infrastructure build-out.

GPUs

At the root of the GPU shortage that has benefitted companies like Lambda Labs and CoreWeave to this point is a limitation at TSMC—Taiwan Semiconductor Manufacturing Company. The key shortage there is on chip-on-wafer-on-substrate (CoWoS) packaging capacity, which is used by all GPUs in the manufacturing process. Currently, TSMC expects that the current shortage will last until about March 2026. TSMC recently announced a plan to build a $2.9B packaging facility that will be operational in 2027, further alleviating shortages.

The major cloud providers, as well as companies like Tesla, Meta and OpenAI, wanting to escape the dynamics of this shortage, have all begun or accelerated work on their own AI processors. That said, they're also dependent on TSMC to actually make their chips—and with Nvidia being one of TSMC's biggest and longest-term customers, Nvidia could still have an advantage on manufacturing, at least until shortages are completely alleviated.

On-premises

Lambda Labs, by offering on-premises workstations that can allow companies to train their AI models locally rather than in the cloud, is well-positioned to benefit from the rising tide of these kinds of workloads.

There are a few key reasons why companies would move their training on-premise vs. keeping it in the cloud, particularly as models improve and the need for compute ramps up further: cost, security, and big data.

Cost: Companies with stable, long-term compute needs might find that investing in on-premises infrastructure leads to lower total cost of ownership compared to continually renting cloud resources. This is particularly relevant for organizations that can efficiently manage and utilize their hardware over time.

Security: On-premises solutions can also provide businesses with control over their data, helping them meet stringent compliance requirements and data sovereignty laws by ensuring that sensitive data does not leave the company's premises.

Big data: When training AI models, it's often necessary to use large datasets of images, text, or audio files—so large that transferring them to a cloud-based environment for training can result in significant data transfer costs, also known as egress fees. These fees can add up quickly, especially when transferring large datasets, making it expensive to train AI models in the cloud. By keeping the data on-premises, organizations can avoid the costs associated with transferring data to and from the cloud.

Infrastructure Build-Out

Lambda is aggressively expanding its owned data center footprint, both to increase capacity and to reduce long-term reliance on third-party leased space. This buildout includes a 24MW AI Factory in Kansas City, expected to launch in early 2026, that will house 10,000+ NVIDIA GPUs with the potential to scale to 100MW+. A partnership with EdgeConneX will add 30+ MW of high-density infrastructure across a 23MW single-tenant Chicago site and an Atlanta facility (both targeted for readiness in 2026), adding meaningful owned capacity ahead of anticipated demand growth.

Read more from

Plaud revenue, growth, and valuation

lightningbolt_icon Unlocked Report
Continue Reading

Read more from