Mistral Small 24B

Summary: Medium-sized, high-performance multilingual and multimodal model for conversational agents and virtual assistants, trained by Mistral. Supports both text and image input.

Intelligence

Speed

Sovereignty

Input

Output

Moderate

High

Moderate

Text, Image

Text

Central parameters

Description: Mistral Small 3.1 24B is a medium-sized, high-performance multilingual and multimodal model from Mistral, designed for conversational agents and virtual assistants. It supports a wide range of European languages, is optimized for both speed and quality, and can process both text and image input. Images can be provided as URLs or base64-encoded strings, allowing the model to analyze and respond to visual content in addition to text. This model is particularly effective for applications requiring fast and reliable chat completions, making it suitable for real-time interactions in various domains. The model supports advanced capabilities like streaming and tool calling, enhancing its usability in interactive applications.

Images will add to the context window, which is 128k tokens for this model. Images are counted as input tokens, so including images in your requests will increase the total input token count. Token billing is based on the total number of tokens processed, including both input and output tokens.

Model identifier: mistralai/Mistral-Small-24B-Instruct

IONOS AI Model Hub Lifecycle and Alternatives

IONOS Launch

End of Life

Alternative

Successor

August 1, 2025

N/A

Origin

Provider

Country

License

Flavor

Release

France

Instruct

March 25, 2025

Technology

Context window

Parameters

Quantization

Multilingual

Multimodal

128k

24B

fp8

Yes

Text, Image input

Modalities

Text

Image

Audio

Input and output

Input

Not supported

Endpoints

Chat Completions

Embeddings

Image generation

v1/chat/completions

Not supported

Not supported

Vision Model Example

{
  "model": "mistralai/Mistral-Small-24B-Instruct",
   "messages": [
      {
        "role": "user",
        "content": [
          {
            "type": "text",
            "text": "What do you see in this image?"
          },
          {
            "type": "image_url",
            "image_url": {
              "url": "https://www.ionos.com/favicon.ico?v2",
              "detail": "auto"
            }
          }
        ]
      }
    ],
    "temperature": 0.7,
    "max_tokens": 1000
  }

Features

Streaming

Tool calling

Supported

Supported

Rate limits

Rate limits ensure fair usage and reliable access to the AI Model Hub. In addition to the contract-wide rate limits, no model-specific limits apply.

Last updated

Was this helpful?