Models and hardware requirements

  • Tier: Premium, Ultimate
  • Offering: GitLab Self-Managed

You can integrate with industry-leading models from Mistral, Meta, Anthropic, and OpenAI through your preferred serving platform.

You can use:

  • Supported models to match your specific performance needs and use cases.
  • In GitLab 18.3 and later, your own compatible model to experiment with models beyond the officially supported options.
  • GitLab-managed models to connect to AI models without the need to host your own infrastructure. These models are managed entirely by GitLab.

Supported models

GitLab-supported models offer different levels of functionality for GitLab Duo features, depending on the specific model and feature combination.

  • check-circle-filled Full functionality: The model can likely handle the feature without any loss of quality.
  • check-circle-dashed Partial functionality: The model supports the feature, but there might be compromises or limitations.
  • dash-circle Limited functionality: The model is unsuitable for the feature, likely resulting in significant quality loss or performance issues. Models that have limited functionality for a feature will not receive GitLab support for that specific feature.
Model familyModelCode completionCode generationGitLab Duo Non-Agentic ChatGitLab Duo Agent Platform
Claude 4Claude 4 Sonnetcheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionality
Claude 4Claude 4.5 Sonnetcheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionality
Claude 4Claude 4.5 Haikucheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionality
Claude 4Claude 4.5 Opuscheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionality
GPTGPT-4 Turbocheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-dashed Partial functionalitydash-circle Limited functionality
GPTGPT-4ocheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitydash-circle Limited functionality
GPTGPT-4o-minicheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-dashed Partial functionalitydash-circle Limited functionality
GPTGPT-5check-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionality
GPTGPT-5 Minicheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-dashed Partial functionality
GPTGPT-5 Codexcheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionality
GPTGPT-5.1check-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionality
GPTGPT-5.2check-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionality
GPTGPT-oss-120Bcheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitydash-circle Limited functionality
GPTGPT-oss-20Bcheck-circle-dashed Partial functionalitycheck-circle-dashed Partial functionalitycheck-circle-dashed Partial functionalitydash-circle Limited functionality
Mistral CodestralCodestral 22B v0.1check-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-dashed Partial functionalitydash-circle Limited functionality
MistralMistral Small 24B Instruct 2506check-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitydash-circle Limited functionality
LlamaLlama 3 8Bcheck-circle-dashed Partial functionalitycheck-circle-filled Full functionalitydash-circle Limited functionalitydash-circle Limited functionality
LlamaLlama 3.1 8Bcheck-circle-dashed Partial functionalitycheck-circle-filled Full functionalitycheck-circle-dashed Partial functionalitydash-circle Limited functionality
LlamaLlama 3 70Bcheck-circle-dashed Partial functionalitycheck-circle-filled Full functionalitydash-circle Limited functionalitydash-circle Limited functionality
LlamaLlama 3.1 70Bcheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitydash-circle Limited functionality
LlamaLlama 3.3 70Bcheck-circle-filled Full functionalitycheck-circle-filled Full functionalitycheck-circle-filled Full functionalitydash-circle Limited functionality

Compatible models

  • Status: Beta

You can use your own compatible models and platform with GitLab Duo features. For compatible models not included in supported model families, use the general model family. This includes models that you host yourself (for example, served through vLLM or LiteLLM), with the requirement that they are exposed through an OpenAI API–compatible /v1 endpoint.

Compatible models are excluded from the definition of Customer Integrated Models in the AI Functionality Terms. Compatible models and platforms must adhere to the OpenAI API specification. Models and platforms that have previously been marked as experimental or beta are now considered compatible models.

This feature is in beta and is therefore subject to change as we gather feedback and improve the integration:

  • GitLab does not provide technical support for issues specific to your chosen model or platform.
  • Not all GitLab Duo features are guaranteed to work optimally with every compatible model.
  • Response quality, speed, and performance overall might vary significantly based on your model choice.
Model familyModel
GeneralAny model compatible with the OpenAI API specification
CodeGemmaCodeGemma 2b
CodeGemmaCodeGemma 7b-it
CodeGemmaCodeGemma 7b-code
Code LlamaCode-Llama 13b
DeepSeek CoderDeepSeek Coder 33b Instruct
DeepSeek CoderDeepSeek Coder 33b Base

GitLab-managed models

GitLab-managed models integrate with GitLab-hosted AI Gateway infrastructure to provide access to AI models curated and made available by GitLab. Instead of using your own self-hosted models, you can choose to use GitLab-managed models for specific GitLab Duo features.

To choose which features use GitLab-managed models, see select a GitLab-managed model for a feature.

When enabled for a specific feature:

  • All calls to those features configured with a GitLab-managed model use the GitLab-hosted AI Gateway, not the self-hosted AI Gateway.
  • No detailed logs are generated in the GitLab-hosted AI Gateway, even when AI logs are enabled. This prevents unintended leaks of sensitive information.

Hardware requirements

The following hardware specifications are the minimum requirements for running GitLab Duo Self-Hosted on-premise. Requirements vary significantly based on the model size and intended usage:

Base system requirements

  • CPU:
    • Minimum: 8 cores (16 threads)
    • Recommended: 16+ cores for production environments
  • RAM:
    • Minimum: 32 GB
    • Recommended: 64 GB for most models
  • Storage:
    • SSD with sufficient space for model weights and data.

GPU requirements by model size

Model sizeMinimum GPU configurationMinimum VRAM required
7B models
(for example, Mistral 7B)
1x NVIDIA A100 (40 GB)35 GB
22B models
(for example, Codestral 22B)
2x NVIDIA A100 (80 GB)110 GB
Mixtral 8x7B2x NVIDIA A100 (80 GB)220 GB
Mixtral 8x22B8x NVIDIA A100 (80 GB)526 GB

Use Hugging Face’s memory utility to verify memory requirements.

Response time by model size and GPU

Small machine

With a a2-highgpu-2g (2x NVIDIA A100 40 GB - 150 GB vRAM) or equivalent:

Model nameNumber of requestsAverage time per request (sec)Average tokens in responseAverage tokens per second per requestTotal time for requestsTotal TPS
Mistral-7B-Instruct-v0.317.09717.0101.197.09101.17
Mistral-7B-Instruct-v0.3108.41764.290.3513.70557.80
Mistral-7B-Instruct-v0.310013.97693.2349.1720.813331.59

Medium machine

With a a2-ultragpu-4g (4x NVIDIA A100 40 GB - 340 GB vRAM) machine on GCP or equivalent:

Model nameNumber of requestsAverage time per request (sec)Average tokens in responseAverage tokens per second per requestTotal time for requestsTotal TPS
Mistral-7B-Instruct-v0.313.80499.0131.253.80131.23
Mistral-7B-Instruct-v0.3106.00740.6122.858.19904.22
Mistral-7B-Instruct-v0.310011.71695.7159.0615.544477.34
Mixtral-8x7B-Instruct-v0.116.50400.061.556.5061.53
Mixtral-8x7B-Instruct-v0.11016.58768.940.3332.56236.13
Mixtral-8x7B-Instruct-v0.110025.90767.3826.8755.571380.68

Large machine

With a a2-ultragpu-8g (8 x NVIDIA A100 80 GB - 1360 GB vRAM) machine on GCP or equivalent:

Model nameNumber of requestsAverage time per request (sec)Average tokens in responseAverage tokens per second per requestTotal time for requests (sec)Total TPS
Mistral-7B-Instruct-v0.313.23479.0148.413.22148.36
Mistral-7B-Instruct-v0.3104.95678.3135.986.85989.11
Mistral-7B-Instruct-v0.310010.14713.2769.6313.965108.75
Mixtral-8x7B-Instruct-v0.116.08709.0116.696.07116.64
Mixtral-8x7B-Instruct-v0.1109.95645.063.6813.40481.06
Mixtral-8x7B-Instruct-v0.110013.83585.0141.8020.382869.12
Mixtral-8x22B-Instruct-v0.1114.39828.057.5614.3857.55
Mixtral-8x22B-Instruct-v0.11020.57629.730.2428.02224.71
Mixtral-8x22B-Instruct-v0.110027.58592.4921.3436.801609.85

AI Gateway Hardware Requirements

For recommendations on AI Gateway hardware, see the AI Gateway scaling recommendations.