---
title: Llama 4 Maverick Instruct (17Bx128E) API | LLM | ModelsLab
description: Multimodal 17B active parameter (400B total) MoE LLM with 128 experts, native image support, 1M token context, excels in coding, multilingual tasks, and.
url: https://modelslab.com/models/meta-llama/meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8.md
canonical: https://modelslab.com/models/meta-llama/meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8.md
type: product
component: Playground/LLM/Index
generated_at: 2026-04-09T06:10:03.799918Z
---

Llama 4 Maverick Instruct (17Bx128E)

 [LLMs.txt](https://modelslab.com/models/meta/meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8/llms.txt) [.md](https://modelslab.com/models/meta/meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8.md)

Llama 4 Maverick Instruct (17Bx128E)
---

Choose a prompt below to get started or type your own message

 Explain quantum computing in simple terms

 Write a Python function to sort a list

 Create a marketing email for a SaaS product

 Compare REST vs GraphQL APIs

Send

### Llama 4 Maverick Instruct (17Bx128E)

meta meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8

Copy model ID

PricingInput $0.27 / 1M tokens

Output $0.85 / 1M tokens

API EndpointsOpenAI Compatible

`https://modelslab.com/api/v7/llm/chat/completions`Endpoint

Anthropic Compatible

`https://modelslab.com/api/v7/llm/v1/messages`Messages

`https://modelslab.com/api/v7/llm/v1/messages/count_tokens`Count Tokens

`https://modelslab.com/api/v7/llm/v1/models`Models

Use with Claude Code

cURL Example

ParametersSystem MessageYou are a helpful AI assistant specialized in providing accurate and detailed responses.

Temperature0.7

Max Tokens1000

Top P0.9

Frequency Penalty0

Presence Penalty0

Model Info

Support

---

*This markdown version is optimized for AI agents and LLMs.*

**Links:**
- [Website](https://modelslab.com)
- [API Documentation](https://docs.modelslab.com)
- [Blog](https://modelslab.com/blog)

---
*Generated by ModelsLab - 2026-04-09*