# Code Llama 13B

**Summary:** Code Llama 13B is a specialized coding model engineered specifically for software development tasks, featuring advanced code infilling capabilities and support for processing up to 100k token sequences. This model excels in code generation, completion, debugging, and refactoring across multiple programming languages, making it an essential tool for developers, software engineers, and automated coding assistants who require intelligent code understanding and generation capabilities.

|                                             **Intelligence**                                            |                                         **Speed**                                         |                   **Sovereignty**                  |                                                                          **Input**                                                                          |                                                                           **Output**                                                                           |
| :-----------------------------------------------------------------------------------------------------: | :---------------------------------------------------------------------------------------: | :------------------------------------------------: | :---------------------------------------------------------------------------------------------------------------------------------------------------------: | :------------------------------------------------------------------------------------------------------------------------------------------------------------: |
| ![Intelligence active](/files/dnDi7yuqXqkBFqwaxdnm) ![Intelligence active](/files/dnDi7yuqXqkBFqwaxdnm) | ![Speed active](/files/evfYW3bq4dTBLlZH3dQf) ![Speed active](/files/evfYW3bq4dTBLlZH3dQf) | ![Sovereignty active](/files/bNpzGRJfez9SidEjNCoy) | ![Text input active](/files/45qlqURbT8c2Ekr8HJfK) ![Image input inactive](/files/0mPVwOtrYhZrpz9clC3D) ![Audio input inactive](/files/PRglWWEC5Zoc5fgynNLM) | ![Text output active](/files/45qlqURbT8c2Ekr8HJfK) ![Image output inactive](/files/0mPVwOtrYhZrpz9clC3D) ![Audio output inactive](/files/PRglWWEC5Zoc5fgynNLM) |
|                                                 *Medium*                                                |                                          *Medium*                                         |                        *Low*                       |                                                                            *Text*                                                                           |                                                                             *Text*                                                                             |

## Central parameters

**Description:** Medium-sized coding model with 13B parameters, specifically trained for Python development and programming tasks.

**Model identifier:** `meta-llama/CodeLlama-13b-Instruct-hf`

## IONOS AI Model Hub Lifecycle and Alternatives

| **IONOS Launch** | **End of Life** |                                                          **Alternative**                                                         | **Successor** |
| :--------------: | :-------------: | :------------------------------------------------------------------------------------------------------------------------------: | :-----------: |
|  *July 1, 2024*  |  *May 21, 2026* | [<mark style="color:blue;">**Qwen3 Coder Next 80B**</mark>](/cloud/ai/ai-model-hub/models/coding-models/qwen3-coder-next-80b.md) |               |

## Origin

|                            **Provider**                            | **Country** |                                     **License**                                    | **Flavor** |     **Release**    |
| :----------------------------------------------------------------: | :---------: | :--------------------------------------------------------------------------------: | :--------: | :----------------: |
| [<mark style="color:blue;">**Meta**</mark>](https://www.meta.com/) |     USA     | [<mark style="color:blue;">**License**</mark>](https://ai.meta.com/llama/license/) |  Instruct  | *January 29, 2024* |

## Technology

| **Context window** | **Parameters** | **Quantization** | **Multilingual** |                                                **Further details**                                               |
| :----------------: | :------------: | :--------------: | :--------------: | :--------------------------------------------------------------------------------------------------------------: |
|        *16k*       |      *13B*     |      *fp16*      |       *Yes*      | [<mark style="color:blue;">**Hugging Face**</mark>](https://huggingface.co/meta-llama/CodeLlama-13b-Instruct-hf) |

## Modalities

|     **Text**     |   **Image**   |   **Audio**   |
| :--------------: | :-----------: | :-----------: |
| Input and output | Not supported | Not supported |

## Endpoints

| **Chat Completions** | **Embeddings** | **Image generation** |
| :------------------: | :------------: | :------------------: |
|  v1/chat/completions |  Not supported |     Not supported    |

## Features

| **Streaming** | **Reasoning** | **Tool calling** |
| :-----------: | :-----------: | :--------------: |
|   Supported   | Not supported |   Not supported  |

## Rate limits

Rate limits ensure fair usage and reliable access to the AI Model Hub. In addition to the [<mark style="color:blue;">contract-wide rate limits</mark>](/cloud/ai/ai-model-hub/how-tos/rate-limits.md), no model-specific limits apply.


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://docs.ionos.com/cloud/ai/ai-model-hub/models/coding-models/meta-code-llama-13b.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
