- Prerequisites
- Deploy a self-hosted model
- Self-hosted models compared to the default GitLab AI vendor architecture
Deploy a self-hosted large language model
Tier: For a limited time, Premium and Ultimate. In the future, GitLab Duo Enterprise.
Offering: Self-managed
Status: Experiment
History
-
Introduced in GitLab 17.1 with a flag named
ai_custom_model
. Disabled by default.
The availability of this feature is controlled by a feature flag.
For more information, see the history.
This feature is considered experimental and is not intended for customer usage outside of initial design partners. We expect major changes to this feature.
This page contains information related to upcoming products, features, and functionality.
It is important to note that the information presented is for informational purposes only.
Please do not rely on this information for purchasing or planning purposes.
The development, release, and timing of any products, features, or functionality may be subject to change or delay and remain at the
sole discretion of GitLab Inc.
When you deploy a self-hosted large language model (LLM), you can:
- Manage the end-to-end transmission of requests to enterprise-hosted LLM backends for GitLab Duo features.
- Keep all of these requests within their enterprise network, ensuring no calls to external architecture.
- Isolate the GitLab instance, AI Gateway, and self-hosted AI model within their own environment, ensuring complete privacy and high security for using AI features, with no reliance on public services.
When you use self-hosted models, you:
- Can choose any GitLab-approved LLM model.
- Can keep all data and request/response logs in your own domain.
- Can select specific GitLab Duo features for your users.
- Do not have to rely on the GitLab shared AI Gateway.
You can connect supported models to LLM features. Model-specific prompts and GitLab Duo feature support is provided by the self-hosted models feature. For more information about this offering, see subscriptions and the Blueprint.
Prerequisites
- You must be able to manage your own LLM infrastructure.
- You must have the GitLab Enterprise Edition.
Deploy a self-hosted model
To deploy a self-hosted large language model:
- Set up your self-hosted model deployment infrastructure and connect it to your GitLab instance.
- Configure your GitLab instance to access self-hosted models using instance and group settings.
Self-hosted models compared to the default GitLab AI vendor architecture
%%{init: { "fontFamily": "GitLab Sans" }}%%
sequenceDiagram
actor User
participant GitLab
participant AIGateway as AI Gateway
participant SelfHostedModel as Self Hosted Model
participant CloudConnector as Cloud Connector
participant GitLabAIVendor as GitLab AI Vendor
User ->> GitLab: Send request
GitLab ->> GitLab: Check if self-hosted model is configured
alt Self-hosted model configured
GitLab ->> AIGateway: Create prompt and send request
AIGateway ->> SelfHostedModel: Perform API request to AI model
SelfHostedModel -->> AIGateway: Respond to the prompt
AIGateway -->> GitLab: Forward AI response
else
GitLab ->> CloudConnector: Create prompt and send request
CloudConnector ->> GitLabAIVendor: Perform API request to AI model
GitLabAIVendor -->> CloudConnector: Respond to the prompt
CloudConnector -->> GitLab: Forward AI response
end
GitLab -->> User: Forward AI response