Skip to content

GitLab Duo

GitLab is creating AI-assisted features across our DevSecOps platform. These features aim to help increase velocity and solve key pain points across the software development lifecycle.

Some features are still in development. View details about support for each status (Experiment, Beta, Generally Available).

As features become Generally Available, GitLab is transparent and updates the documentation to clearly state how and where you can access these capabilities.

Goal Feature Tier/Offering/Status
Helps you write code more efficiently by showing code suggestions as you type.

Watch overview
Code Suggestions Free Open Access subject to the Testing Agreement available through February 14, 2024:
- SaaS, All subscription tiers
- Self-managed, Premium and Ultimate tiers

Generally Available starting February 15, 2024:
- SaaS, self-managed
- Premium and Ultimate tiers as part of GitLab Duo Pro
Processes and generates text and code in a conversational manner. Helps you quickly identify useful information in large volumes of text in issues, epics, code, and GitLab documentation. Chat Beta Access subject to the Testing Agreement:
- SaaS, self-managed
- Premium and Ultimate tiers

Status: Beta
Helps you discover or recall Git commands when and where you need them. Git suggestions Tier: Ultimate
Offering: SaaS
Status: Experiment
Assists with quickly getting everyone up to speed on lengthy conversations to help ensure you are all on the same page. Discussion summary Tier: Ultimate
Offering: SaaS
Status: Experiment
Generates issue descriptions. Issue description generation Tier: Ultimate
Offering: SaaS
Status: Experiment
Automates repetitive tasks and helps catch bugs early. Test generation Tier: Ultimate
Offering: SaaS, self-managed
Status: Beta
Generates a description for the merge request based on the contents of the template. Merge request template population Tier: Ultimate
Offering: SaaS
Status: Experiment
Assists in creating faster and higher-quality reviews by automatically suggesting reviewers for your merge request.

Watch overview
Suggested Reviewers Tier: Ultimate
Offering: SaaS
Status: Generally Available
Efficiently communicates the impact of your merge request changes. Merge request summary Tier: Ultimate
Offering: SaaS
Status: Experiment
Helps ease merge request handoff between authors and reviewers and help reviewers efficiently understand suggestions. Code review summary Tier: Ultimate
Offering: SaaS
Status: Experiment
Helps you remediate vulnerabilities more efficiently, boost your skills, and write more secure code.

Watch overview
Vulnerability summary Tier: Ultimate
Offering: SaaS
Status: Beta
Generates a merge request containing the changes required to mitigate a vulnerability. Vulnerability resolution Tier: Ultimate
Offering: SaaS
Status: Experiment
Helps you understand code by explaining it in English language.

Watch overview
Code explanation Tier: Ultimate
Offering: SaaS
Status: Experiment
Assists you in determining the root cause for a pipeline failure and failed CI/CD build. Root cause analysis Tier: Ultimate
Offering: SaaS
Status: Experiment
Assists you with predicting productivity metrics and identifying anomalies across your software development lifecycle. Value stream forecasting Tier: Ultimate
Offering: SaaS, self-managed
Status: Experiment
Processes and responds to your questions about your application's usage data. Product Analytics Tier: Ultimate
Offering: SaaS
Status: Experiment

Enable AI/ML features

Experimental AI features and how to use them

The following subsections describe the experimental AI features in more detail.

Explain code in the Web UI with Code explanation

DETAILS: Tier: Ultimate Offering: SaaS Status: Experiment

  • Introduced in GitLab 15.11 as an Experiment on GitLab.com.

To use this feature:

GitLab can help you get up to speed faster if you:

  • Spend a lot of time trying to understand pieces of code that others have created, or
  • Struggle to understand code written in a language that you are not familiar with.

By using a large language model, GitLab can explain the code in natural language.

To explain your code:

  1. On the left sidebar, select Search or go to and find your project.
  2. Select any file in your project that contains code.
  3. On the file, select the lines that you want to have explained.
  4. On the left side, select the question mark ({question}). You might have to scroll to the first line of your selection to view it. This sends the selected code, together with a prompt, to provide an explanation to the large language model.
  5. A drawer is displayed on the right side of the page. Wait a moment for the explanation to be generated.
  6. Provide feedback about how satisfied you are with the explanation, so we can improve the results.

You can also have code explained in the context of a merge request. To explain code in a merge request:

  1. On the left sidebar, select Search or go to and find your project.

  2. Select Code > Merge requests, then select your merge request.

  3. On the secondary menu, select Changes.

  4. On the file you would like explained, select the three dots ({ellipsis_v}) and select View File @ $SHA.

    A separate browser tab opens and shows the full file with the latest changes.

  5. On the new tab, select the lines that you want to have explained.

  6. On the left side, select the question mark ({question}). You might have to scroll to the first line of your selection to view it. This sends the selected code, together with a prompt, to provide an explanation to the large language model.

  7. A drawer is displayed on the right side of the page. Wait a moment for the explanation to be generated.

  8. Provide feedback about how satisfied you are with the explanation, so we can improve the results.

How to use the Explain Code Experiment

We cannot guarantee that the large language model produces results that are correct. Use the explanation with caution.

Summarize issue discussions with Discussion summary

DETAILS: Tier: Ultimate Offering: SaaS Status: Experiment

To use this feature:

You can generate a summary of discussions on an issue:

  1. In an issue, scroll to the Activity section.
  2. Select View summary.

The comments in the issue are summarized in as many as 10 list items. The summary is displayed only for you.

Provide feedback on this experimental feature in issue 407779.

Data usage: When you use this feature, the text of public comments on the issue are sent to the large language model referenced above.

Forecast deployment frequency with Value stream forecasting

DETAILS: Tier: Ultimate Offering: SaaS, self-managed Status: Experiment

To use this feature:

  • The parent group of the project must:
  • You must be a member of the project with sufficient permissions to view the CI/CD analytics.

In CI/CD Analytics, you can view a forecast of deployment frequency:

  1. On the left sidebar, select Search or go to and find your project.
  2. Select Analyze > CI/CD analytics.
  3. Select the Deployment frequency tab.
  4. Turn on the Show forecast toggle.
  5. On the confirmation dialog, select Accept testing terms.

The forecast is displayed as a dotted line on the chart. Data is forecasted for a duration that is half of the selected date range. For example, if you select a 30-day range, a forecast for the following 15 days is displayed.

Forecast deployment frequency

Provide feedback on this experimental feature in issue 416833.

Root cause analysis

DETAILS: Tier: Ultimate Offering: SaaS Status: Experiment

To use this feature:

When the feature is available, the "Root cause analysis" button will appears on a failed CI/CD job. Selecting this button generates an analysis regarding the reason for the failure.

Summarize an issue with Issue description generation

DETAILS: Tier: Ultimate Offering: SaaS Status: Experiment

To use this feature:

You can generate the description for an issue from a short summary.

  1. Create a new issue.
  2. Above the Description field, select AI actions > Generate issue description.
  3. Write a short description and select Submit.

The issue description is replaced with AI-generated text.

Provide feedback on this experimental feature in issue 409844.

Data usage: When you use this feature, the text you enter is sent to the large language model referenced above.

Language models

Feature Large Language Model
Git suggestions Vertex AI Codey codechat-bison
Discussion summary Vertex AI Codey text-bison
Issue description generation Anthropic Claude-2
Code Suggestions For Code Completion: Vertex AI Codey code-gecko For Code Generation: Anthropic Claude-2
Test generation Anthropic Claude-2
Merge request template population Vertex AI Codey text-bison
Suggested Reviewers GitLab creates a machine learning model for each project, which is used to generate reviewers View the issue
Merge request summary Vertex AI Codey text-bison
Code review summary Vertex AI Codey text-bison
Vulnerability summary Vertex AI Codey text-bison Anthropic Claude-2 if degraded performance
Vulnerability resolution Vertex AI Codey code-bison
Code explanation Vertex AI Codey codechat-bison
GitLab Duo Chat Anthropic Claude-2 Vertex AI Codey textembedding-gecko
Root cause analysis Vertex AI Codey text-bison
Value stream forecasting Statistical forecasting

Data usage

GitLab AI features leverage generative AI to help increase velocity and aim to help make you more productive. Each feature operates independently of other features and is not required for other features to function. GitLab selects the best-in-class large-language models for specific tasks. We use Google Vertex AI Models and Anthropic Claude.

Progressive enhancement

These features are designed as a progressive enhancement to existing GitLab features across our DevSecOps platform. They are designed to fail gracefully and should not prevent the core functionality of the underlying feature. You should note each feature is subject to its expected functionality as defined by the relevant feature support policy.

Stability and performance

These features are in a variety of feature support levels. Due to the nature of these features, there may be high demand for usage which may cause degraded performance or unexpected downtime of the feature. We have built these features to gracefully degrade and have controls in place to allow us to mitigate abuse or misuse. GitLab may disable beta and experimental features for any or all customers at any time at our discretion.

Data privacy

GitLab Duo AI features are powered by a generative AI models. The processing of any personal data is in accordance with our Privacy Statement. You may also visit the Sub-Processors page to see the list of our Sub-Processors that we use to provide these features.

Data retention

The below reflects the current retention periods of GitLab AI model Sub-Processors:

  • Anthropic discards model input and output data immediately after the output is provided. Anthropic currently does not store data for abuse monitoring. Model input and output is not used to train models.
  • Google discards model input and output data immediately after the output is provided. Google currently does not store data for abuse monitoring. Model input and output is not used to train models.

All of these AI providers are under data protection agreements with GitLab that prohibit the use of Customer Content for their own purposes, except to perform their independent legal obligations.

Telemetry

GitLab Duo collects aggregated or de-identified first-party usage data through our Snowplow collector. This usage data includes the following metrics:

  • Number of unique users
  • Number of unique instances
  • Prompt lengths
  • Model used
  • Status code responses
  • API responses times

Training data

GitLab does not train generative AI models based on private (non-public) data. The vendors we work with also do not train models based on private data.

For more information on our AI sub-processors, see:

Model accuracy and quality

Generative AI may produce unexpected results that may be:

  • Low-quality
  • Incoherent
  • Incomplete
  • Produce failed pipelines
  • Insecure code
  • Offensive or insensitive
  • Out of date information

GitLab is actively iterating on all our AI-assisted capabilities to improve the quality of the generated content. We improve the quality through prompt engineering, evaluating new AI/ML models to power these features, and through novel heuristics built into these features directly.