
Benchmarking ZeRO-Inference on the NVIDIA GH200 Grace Hopper Superchip
This blog explores the synergy of DeepSpeed’s ZeRO-Inference, a technology designed to make large AI model inference more accessible and cost-effective, with ...
Ensuring our customers’ success is a core value at Lambda, and MLPerf Inference v5.0 is a part of our commitment to providing the best compute platform for AI innovation. Today, we’re thrilled to share our first public-facing results on NVIDIA HGX B200 and NVIDIA HGX H200 platforms, showcasing how our innovative cloud infrastructure is setting new standards for AI inference performance.
Published on by Amit Kumar
This blog explores the synergy of DeepSpeed’s ZeRO-Inference, a technology designed to make large AI model inference more accessible and cost-effective, with ...
Published on by Chuan Li
Persistent storage for Lambda Cloud recently exited beta and became available in the majority of regions. We are excited to announce that filesystems are now ...
Published on by Kathy Bui
The Lambda Vector One is now available for order. The new single-GPU desktop PC is built to tackle demanding AI/ML tasks, from fine-tuning Stable Diffusion to ...
Published on by Samuel Park
Create a cloud account instantly to spin up GPUs today or contact us to secure a long-term contract for thousands of GPUs