
Introducing the Lambda Inference API: Lowest-Cost Inference Anywhere
Today, we’re excited to announce the GA release of the Lambda Inference API, the lowest-cost inference anywhere. For just a fraction of a cent, you can access ...
Ensuring our customers’ success is a core value at Lambda, and MLPerf Inference v5.0 is a part of our commitment to providing the best compute platform for AI innovation. Today, we’re thrilled to share our first public-facing results on NVIDIA HGX B200 and NVIDIA HGX H200 platforms, showcasing how our innovative cloud infrastructure is setting new standards for AI inference performance.
Published on by Amit Kumar
Today, we’re excited to announce the GA release of the Lambda Inference API, the lowest-cost inference anywhere. For just a fraction of a cent, you can access ...
Published on by Nick Harvey
When it comes to large language model (LLM) inference, cost and performance go hand-in-hand. Single GPU instances are practical and economical; however, models ...
Published on by Thomas Bordes
We're excited to announce the launch of the NVIDIA GH200 Grace Hopper Superchip on Lambda On-Demand. Now, with just a few clicks in your Lambda Cloud account, ...
Published on by Nick Harvey
Create a cloud account instantly to spin up GPUs today or contact us to secure a long-term contract for thousands of GPUs