Llama 4 Scout

Meta
meta/llama-4-scout

512K context, vision

Context Window

512K

512,000 tokens

Max Output

33K

32,768 tokens

About this model

Efficient Llama 4 variant with 512K context

This model supports up to 512K tokens of context. It includes native vision understanding for analyzing images and documents. It provides strong code generation and debugging capabilities.

Access it through Chuizi.AI with a single ck- API key β€” no separate Meta account needed.

Highlights

512K context window
33K max output
Native vision support
Strong code generation

Best For

Code generationRefactoringDebuggingDocumentation
2025-04-05

Capabilities

ChatVisionCodetools

Aliases

llama-4-scout

Pricing (per 1M tokens)

Pricing (per 1M tokens)/ 1M tokens
Input / 1M$0.62
Output / 1M$0.92

Final prices shown

Quick Start

main.py
from openai import OpenAI

client = OpenAI(
    base_url="https://api.chuizi.ai/v1",
    api_key="ck-your-key-here",
)

response = client.chat.completions.create(
    model="meta/llama-4-scout",
    messages=[{"role": "user", "content": "Hello!"}],
)
print(response.choices[0].message.content)

FAQ

Related Models

Llama 4 Scout β€” Pricing, Context, Capabilities | Chuizi AI