The website requires JavaScript to function properly.
Model Library
GPUs
Pricing
Docs
Log In
Get Started
Get Started
Log In
Model Library
GPUs
Pricing
Docs
meta-llama/llama-4-maverick-17b-128e-instruct-fp8
1048576 Context
$0.17 / 1M input tokens
$0.85 / 1M output tokens
Learn about API rate limits
1048576 Context
$0.17/1M input tokens
$0.85/1M output tokens
Demo
API
README
Get up to $500 in LLM API
Code
Join Discord
Loading...
Model Configuration
Response format
System Prompt
Be a helpful assistant
max_tokens
temperature
top_p
min_p
top_k
presence_penalty
frequency_penalty
repetition_penalty