NVIDIA B200s are live on Lambda Cloud! Set up your Demo today! 
meta-logo

Llama 4 Maverick 17B 128E Instruct FP8

The Llama 4 Maverick 17B 128E Instruct FP8 model is an image-text-to-text model designed to process and generate text based on image and text inputs. It is part of the Llama 4 series of models.

This model is available on Lambda Inference.

Price per 1M input tokens
$0.18
Price per 1M output tokens
$0.60
Context
1M

Introduction

The Llama 4 Maverick 17B 128E Instruct FP8 model is a type of Llama 4 model, utilizing a compressed tensor quantization approach. This model is trained to handle image-text-to-text tasks, leveraging the transformers library for efficient processing. Its architecture is designed to facilitate a range of applications, from text generation to multimodal processing. The model's design goals focus on balancing performance and efficiency. The Llama 4 Maverick 17B 128E Instruct FP8 model is licensed under the Llama 4 license, which governs its use and distribution. The model is part of a broader family of models aimed at advancing image-text processing capabilities.

* This content was generated using the llama-4-scout-17b-16e-instruct model via Lambda AI Inference