Mistral Nemo 12B

Summary: Mistral Nemo is a compact yet powerful 12-billion parameter language model co-developed with NVIDIA, designed specifically for conversational agents and virtual assistants. This French-engineered model features an advanced Tekken tokenizer that provides superior efficiency for European languages, making it ideal for applications requiring fast response times while maintaining high-quality multilingual performance across German, French, Spanish, Italian, and Portuguese.

Intelligence

Speed

Sovereignty

Input

Output

Low

High

Moderate

Text

Text

Central parameters

Description: Mistral Nemo replaces Mistral 7B as the latest generation small language model, offering enhanced performance with 12.2B parameters and a 128k context window. The model utilizes fp8 quantization for optimized inference efficiency.

Model identifier: mistralai/Mistral-Nemo-Instruct-2407

IONOS AI Model Hub Lifecycle and Alternatives

IONOS Launch

End of Life

Alternative

Successor

May 15, 2025

N/A

Origin

Provider

Country

License

Flavor

Release

France

Instruct

July 18, 2024

Technology

Context window

Parameters

Quantization

Multilingual

128k

12.2B

fp8

Yes

Modalities

Text

Image

Audio

Input and output

Not supported

Not supported

Endpoints

Chat Completions

Embeddings

Image generation

v1/chat/completions

Not supported

Not supported

Features

Streaming

Tool calling

Supported

Supported

Rate limits

Rate limits ensure fair usage and reliable access to the AI Model Hub. In addition to the contract-wide rate limits, no model-specific limits apply.

Last updated

Was this helpful?