Llama 3.1 405B

Summary: Llama 3.1 405B is Meta's flagship large language model representing the pinnacle of open-source AI capabilities with exceptional reasoning abilities and comprehensive knowledge coverage. This massive model excels in the most demanding AI applications including advanced research, complex problem-solving, sophisticated content creation, and enterprise-grade AI solutions where maximum intelligence and accuracy are paramount, despite longer inference times inherent to its large-scale architecture.

Intelligence

Speed

Sovereignty

Input

Output

High

Low

Low

Text

Text

Central parameters

Description: Largest open-source model from Meta with 405B parameters, optimized with FP8 quantization for maximum intelligence and knowledge coverage.

Model identifier: meta-llama/Meta-Llama-3.1-405B-Instruct-FP8

IONOS AI Model Hub Lifecycle and Alternatives

IONOS Launch

End of Life

Alternative

Successor

August 1, 2024

N/A

Origin

Provider

Country

License

Flavor

Release

Technology

Context window

Parameters

Quantization

Multilingual

Further details

Modalities

Text

Image

Audio

Input and output

Not supported

Not supported

Endpoints

Chat Completions

Embeddings

Image generation

v1/chat/completions

Not supported

Not supported

Features

Streaming

Reasoning

Tool calling

Supported

Not supported

Supported

Usage example

Chat completions

The following example demonstrates how to use Llama 3.1 405B for complex reasoning tasks.

API Endpoint: POST https://openai.inference.de-txl.ionos.com/v1/chat/completions

Request:

Response:

Troubleshooting

Infinite or repetitive response loops

Llama 3.1 405B can produce repetitive output that does not terminate naturally. When "max_tokens" is set to the context window maximum (128000) or left unconfigured, the model continues generating until it hits that context window ceiling.

Recommended mitigations:

  1. Set max_tokens explicitly. Avoid setting max_tokens (or max_completion_tokens) to the full context window (128000). Instead, limit the value to match your specific use case:

    Use case

    Recommended value

    Conversational use and short responses

    2048

    Detailed analysis and code generation

    8192

    Long-form documents and research summaries

    16384

circle-info

Note: Use values exceeding 16384 only when the task strictly requires them. When doing so, always implement stop sequences (see Step 2) to ensure the model terminates correctly.

  1. Add explicit stop sequences. Add both Llama 3 end-of-turn tokens as stop strings in your request:

  1. Use sampling instead of greedy decoding. Avoid combining temperature: 0 with ambiguous or contradictory prompts, as this often triggers infinite loops. Instead, use the following:

  2. Apply a frequency penalty. Setting frequency_penalty to a small positive value reduces the likelihood of the model repeating the same tokens. A value between 0.1 and 0.3 is effective for most use cases.

Example request with all mitigations applied:

Rate limits

Rate limits ensure fair usage and reliable access to the AI Model Hub. In addition to the contract-wide rate limits, no model-specific limits apply.

Last updated

Was this helpful?