Llama 3.1 8b

Meta
meta/llama-3.1-8b

128K context

Context Window

128K

128,000 tokens

Max Output

8K

8,192 tokens

About this model

Lightweight and fast, ideal for simple tasks

This model supports up to 128K tokens of context. It provides strong code generation and debugging capabilities.

Access it through Chuizi.AI with a single ck- API key β€” no separate Meta account needed.

Highlights

128K context window
8K max output
Strong code generation

Best For

Code generationRefactoringDebuggingDocumentation
2024-07-23

Capabilities

ChatCodetools

Aliases

llama-3.1-8b

Pricing (per 1M tokens)

Pricing (per 1M tokens)/ 1M tokens
Input / 1M$0.19
Output / 1M$0.19

Final prices shown

Quick Start

main.py
from openai import OpenAI

client = OpenAI(
    base_url="https://api.chuizi.ai/v1",
    api_key="ck-your-key-here",
)

response = client.chat.completions.create(
    model="meta/llama-3.1-8b",
    messages=[{"role": "user", "content": "Hello!"}],
)
print(response.choices[0].message.content)

FAQ

Related Models

Llama 3.1 8b β€” Pricing, Context, Capabilities | Chuizi AI