Seedance 2.0 is here - create consistent, multimodal AI videos faster with images, videos, and audio in one prompt.

Try Now
Skip to main content

Llama 4 Maverick Instruct (17Bx128E)

meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8metaClosed Source Model$0.560000 / call

Llama 4 Maverick Instruct (17Bx128E)

Choose a prompt below to get started or type your own message

About Llama 4 Maverick Instruct (17Bx128E)

Multimodal 17B active parameter (400B total) MoE LLM with 128 experts, native image support, 1M token context, excels in coding, multilingual tasks, and enterprise document intelligence.

Technical Specifications

Model ID
meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8
Category
LLM Models
Task
Text Generation
Price
$0.560000 per million tokens
Added
July 22, 2025

Key Features

  • Chat completion and multi-turn conversation API
  • Streaming response with token-by-token output
  • Function calling and tool use support
  • System prompts and role-based messaging
  • JSON mode and structured output

Quick Start

Integrate Llama 4 Maverick Instruct (17Bx128E) into your application with a single API call. Get your API key from the pricing page to get started.

import requests
import json
url = "https://modelslab.com/api/v7/llm/chat/completions"
headers = {
"Content-Type": "application/json"
}
data = {
"model_id": "meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8",
"messages": [
{
"role": "user",
"content": "Hello!"
}
],
"max_tokens": 1000,
"key": "YOUR_API_KEY"
}
try:
response = requests.post(url, headers=headers, json=data)
response.raise_for_status() # Raises an HTTPError for bad responses (4XX or 5XX)
result = response.json()
print("API Response:")
print(json.dumps(result, indent=2))
except requests.exceptions.HTTPError as http_err:
print(f"HTTP error occurred: {http_err} - {response.text}")
except Exception as err:
print(f"Other error occurred: {err}")

View the full API documentation for SDKs, code examples in Python, JavaScript, and more.

Pricing

Llama 4 Maverick Instruct (17Bx128E) API costs $0.560000 per million tokens. Pay only for what you use with no minimum commitments. View pricing plans

Use Cases

  • AI chatbots and virtual assistants
  • Code generation and developer tools
  • Content writing and copywriting automation
  • Data analysis, summarization, and extraction

Llama 4 Maverick Instruct (17Bx128E) FAQ

Multimodal 17B active parameter (400B total) MoE LLM with 128 experts, native image support, 1M token context, excels in coding, multilingual tasks, and enterprise document intelligence.

You can integrate Llama 4 Maverick Instruct (17Bx128E) into your application with a single API call. Sign up on ModelsLab to get your API key, then use the model ID "meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8" in your API requests. We provide SDKs for Python, JavaScript, and cURL examples in the API documentation.

Llama 4 Maverick Instruct (17Bx128E) costs $0.560000 per million tokens. ModelsLab uses pay-per-use pricing with no minimum commitments. A free tier is available to get started.

The model ID for Llama 4 Maverick Instruct (17Bx128E) is "meta-llama-Llama-4-Maverick-17B-128E-Instruct-FP8". Use this ID in your API requests to specify this model.

Yes, ModelsLab offers a free tier that lets you try Llama 4 Maverick Instruct (17Bx128E) and other AI models. Sign up to get free API credits and start building immediately.