- 5 network automation startups to watch
- 4 Security Controls Keeping Up with the Evolution of IT Environments
- ICO Warns of Festive Mobile Phone Privacy Snafu
- La colaboración entre Seguridad y FinOps puede generar beneficios ocultos en la nube
- El papel del CIO en 2024: una retrospectiva del año en clave TI
Lambda raises $320 million for GPU cloud
AI cloud service provider Lambda has scored a $320 million cash infusion to build out its GPU-based services, which provide AI training clusters made up of thousands of Nvidia accelerators.
Lambda is the latest cloud company to offer GPU processing – instead of the standard CPU processing – dedicated to all things AI, particularly inference and training. Vultr, CoreWeave and Voltage Park are all offering similar cloud GPU services.
Lambda is preparing to deploy “tens of thousands” of Nvidia GPUs, including the current top-of-the-line H100 Hopper accelerators as well as Nvidia’s forthcoming G200 GPU accelerators, which are set to double the performance of the H100. Lambda is also looking to deploy Nvidia’s hybrid GH200 CPU/GPU superchips.
Lambda’s stated mission is to build “the #1 AI compute platform in the world,” and to accomplish this, “we’ll need lots of Nvidia GPUs, ultra-fast networking, lots of data center space, and lots of great new software to delight you and your AI engineering team,” it said in a statement announcing the funding.
The $320 million Series C funding is led by a number of venture funds, including B Capital, SK Telecom, T. Rowe Price Associates, Inc., and existing investors Crescent Cove, Mercato Partners, 1517 Fund, Bloomberg Beta, and Gradient Ventures, among others.
“With this new financing, Lambda will accelerate the growth of our GPU cloud, ensuring AI engineering teams have access to thousands of Nvidia GPUs with high-speed Nvidia Quantum-2 InfiniBand networking,” the company said.