The website requires JavaScript to function properly.
Model Library
GPUs
Pricing
Docs
Log In
Get Started
Get Started
Log In
Model Library
GPUs
Pricing
Docs
meta-llama/llama-4-maverick-17b-128e-instruct-fp8
1048576 Context
$0.2 / 1M input tokens
$0.85 / 1M output tokens
Learn about API rate limits
1048576 Context
$0.2/1M input tokens
$0.85/1M output tokens
Demo
API
README
Code
Join Discord
Loading...
Model Configuration
Response format
System Prompt
Be a helpful assistant
max_tokens
temperature
top_p
min_p
top_k
presence_penalty
frequency_penalty
repetition_penalty