Embeddings
AI Model Hub for Free: From December 1, 2024, to June 30, 2025, IONOS is offering all foundation models in the AI Model Hub for free. Create your contract today and kickstart your AI journey!
One seemingly simple task for humans is identifying semantically similar concepts. Take the following three texts, for example:
"AI Model Hub"
"Michael Jackson"
"best-selling music artists"
For a human, it is easy to tell that "Michael Jackson" and "best-selling music artists" are related, while "AI Model Hub" is not.
Embeddings, a core concept in modern foundation models, provide an effective way for machines to perform this kind of semantic comparison. An embedding is a numerical vector representation of a text, where the key property is this: semantically similar texts have similar vectors.
Semantic similarity
To illustrate this, imagine we convert the three example texts into embeddings:
"AI Model Hub": (0.10; 0.10)
"Michael Jackson": (0.95; 0.90)
"best-selling music artists": (0.96; 0.87)
In practice, embedding vectors have dozens or even thousands of dimensions. Here, we simplify them to 2D for clarity. These embeddings can be visualized like this:

As shown, the embeddings for "Michael Jackson" and "best-selling music artists" are close to each other, while "AI Model Hub" is farther away, mirroring our intuitive understanding.
Embedding models are models that convert data—like text, into these embeddings. They exist for different types of content, including text (in one or multiple languages), images, audio, and more. The IONOS AI Model Hub supports embedding models for English text and multilingual models.
To measure how similar two embeddings are, we use a similarity score. One commonly used metric is cosine similarity, which calculates the cosine of the angle between two vectors. The score ranges from -1 to 1:
1: very similar
0: unrelated
-1: opposite meanings (in rare cases)
Explore further
To try out embeddings in practice, see Text Embeddings.
Last updated
Was this helpful?