- Types of pipelines
Configure a pipeline
- Ref specs for runners
- View pipelines
- Run a pipeline manually
- Run a pipeline by using a URL query string
- Add manual interaction to your pipeline
- Skip a pipeline
- Delete a pipeline
- Pipeline security on protected branches
- How pipeline duration is calculated
- Visualize pipelines
- Pipelines API
Pipelines are the top-level component of continuous integration, delivery, and deployment.
- Jobs, which define what to do. For example, jobs that compile or test code.
- Stages, which define when to run the jobs. For example, stages that run tests after stages that compile the code.
Jobs are executed by runners. Multiple jobs in the same stage are executed in parallel, if there are enough concurrent runners.
If all jobs in a stage succeed, the pipeline moves on to the next stage.
If any job in a stage fails, the next stage is not (usually) executed and the pipeline ends early.
In general, pipelines are executed automatically and require no intervention once created. However, there are also times when you can manually interact with a pipeline.
A typical pipeline might consist of four stages, executed in the following order:
buildstage, with a job called
teststage, with two jobs called
stagingstage, with a job called
productionstage, with a job called
Pipelines can be configured in many different ways:
- Basic pipelines run everything in each stage concurrently, followed by the next stage.
- Directed Acyclic Graph Pipeline (DAG) pipelines are based on relationships between jobs and can run more quickly than basic pipelines.
- Merge request pipelines run for merge requests only (rather than for every commit).
- Merged results pipelines are merge request pipelines that act as though the changes from the source branch have already been merged into the target branch.
- Merge trains use merged results pipelines to queue merges one after the other.
- Parent-child pipelines break down complex pipelines into one parent pipeline that can trigger multiple child sub-pipelines, which all run in the same project and with the same SHA. This pipeline architecture is commonly used for mono-repos.
- Multi-project pipelines combine pipelines for different projects together.
Parent-child pipelines and multi-project pipelines can sometimes be used for similar purposes, but there are some key differences:
- Run under the same project, ref, and commit SHA as the parent pipeline.
- Affect the overall status of the ref the pipeline runs against. For example,
if a pipeline fails for the main branch, it’s common to say that “main is broken”.
The status of child pipelines don’t directly affect the status of the ref, unless the child
pipeline is triggered with
- Are automatically canceled if the pipeline is configured with
interruptiblewhen a new pipeline is created for the same ref.
- Display only the parent pipelines in the pipeline index page. Child pipelines are visible when visiting their parent pipeline’s page.
- Are limited to 2 levels of nesting. A parent pipeline can trigger multiple child pipelines,
and those child pipeline can trigger multiple child pipelines (
A -> B -> C).
- Are triggered from another pipeline, but the upstream (triggering) pipeline does not have much control over the downstream (triggered) pipeline. However, it can choose the ref of the downstream pipeline, and pass CI/CD variables to it.
- Affect the overall status of the ref of the project it runs in, but does not
affect the status of the triggering pipeline’s ref, unless it was triggered with
- Are not automatically canceled in the downstream project when using
interruptibleif a new pipeline runs for the same ref in the upstream pipeline. They can be automatically canceled if a new pipeline is triggered for the same ref on the downstream project.
- Multi-project pipelines are standalone pipelines because they are normal pipelines that happened to be triggered by an external project. They are all visible on the pipeline index page.
- Are independent, so there are no nesting limits.
Pipelines and their component jobs and stages are defined in the CI/CD pipeline configuration file for each project.
For a list of configuration options in the CI pipeline file, see the GitLab CI/CD Pipeline Configuration Reference.
You can also configure specific aspects of your pipelines through the GitLab UI. For example:
When a runner picks a pipeline job, GitLab provides that job’s metadata. This includes the Git refspecs, which indicate which ref (branch, tag, and so on) and commit (SHA1) are checked out from your project repository.
This table lists the refspecs injected for each pipeline type:
|pipeline for branches||
|pipeline for tags||
|merge request pipeline|
refs/tags/<name> exist in your
project repository. GitLab generates the special ref
refs/pipelines/<id> during a
running pipeline job. This ref can be created even after the associated branch or tag has been
deleted. It’s therefore useful in some features such as automatically stopping an environment,
and merge trains
that might run pipelines after branch deletion.
You can find the current and historical pipeline runs under your project’s CI/CD > Pipelines page. You can also access pipelines for a merge request by navigating to its Pipelines tab.
Select a pipeline to open the Pipeline Details page and show the jobs that were run for that pipeline. From here you can cancel a running pipeline, retry jobs on a failed pipeline, or delete a pipeline.
Starting in GitLab 12.3, a link to the
latest pipeline for the last commit of a given branch is available at
/project/pipelines/latest redirects you to the latest pipeline for the last commit
on the project’s default branch.
Starting in GitLab 13.0, you can filter the pipeline list by:
- Trigger author
- Branch name
- Status (GitLab 13.1 and later)
- Tag (GitLab 13.1 and later)
- Source (GitLab 14.3 and later)
Starting in GitLab 14.2, you can change the pipeline column to display the pipeline ID or the pipeline IID.
Pipelines can be manually executed, with predefined or manually-specified variables.
You might do this if the results of a pipeline (for example, a code build) are required outside the normal operation of the pipeline.
To execute a pipeline manually:
- On the top bar, select Menu > Projects and find your project.
- On the left sidebar, select CI/CD > Pipelines.
- Select Run pipeline.
- In the Run for branch name or tag field, select the branch or tag to run the pipeline for.
- Enter any environment variables required for the pipeline to run. You can set specific variables to have their values prefilled in the form.
- Select Run pipeline.
The pipeline now executes the jobs as configured.
Introduced in GitLab 13.7.
In pipelines triggered manually, the Run pipelines page displays all top-level variables
value defined in the
.gitlab-ci.yml file. The values
can then be modified if needed, which overrides the value for that single pipeline run.
The description is displayed next to the variable. It can be used to explain what the variable is used for, what the acceptable values are, and so on:
variables: DEPLOY_ENVIRONMENT: value: "staging" # Deploy to staging by default description: "The deployment target. Change this variable to 'canary' or 'production' if needed."
You cannot set job-level variables to be pre-filled when you run a pipeline manually.
Introduced in GitLab 12.5.
You can use a query string to pre-populate the Run Pipeline page. For example, the query string
.../pipelines/new?ref=my_branch&var[foo]=bar&file_var[file_foo]=file_bar pre-populates the
Run Pipeline page with:
Run for field:
The format of the
pipelines/new URL is:
The following parameters are supported:
ref: specify the branch to populate the Run for field with.
var: specify a
file_var: specify a
file_var, a key and value are required.
Manual jobs, allow you to require manual interaction before moving forward in the pipeline.
You can do this straight from the pipeline graph. Just select the play button to execute that particular job.
For example, your pipeline can start automatically, but require a manual action to
deploy to production.
In the example below, the
production stage has a job with a manual action:
Introduced in GitLab 11.11.
Multiple manual actions in a single stage can be started at the same time using the “Play all manual” After you select this action, each individual manual action is triggered and refreshed to an updated status.
This functionality is only available:
- For users with at least the Developer role.
- If the stage contains manual actions.
To push a commit without triggering a pipeline, add
[ci skip] or
[skip ci], using any
capitalization, to your commit message.
Alternatively, if you are using Git 2.10 or later, use the
ci.skip Git push option.
ci.skip push option does not skip merge request
Introduced in GitLab 12.7.
Users with the Owner role for a project can delete a pipeline by clicking on the pipeline in the CI/CD > Pipelines to get to the Pipeline Details page, then selecting Delete.
A strict security model is enforced when pipelines are executed on protected branches.
The following actions are allowed on protected branches only if the user is allowed to merge or push on that specific branch:
- Run manual pipelines (using the Web UI or pipelines API).
- Run scheduled pipelines.
- Run pipelines using triggers.
- Run on-demand DAST scan.
- Trigger manual actions on existing pipelines.
- Retry or cancel existing jobs (using the Web UI or pipelines API).
Variables marked as protected are accessible only to jobs that run on protected branches, preventing untrusted users getting unintended access to sensitive information like deployment credentials and tokens.
Runners marked as protected can run jobs only on protected branches, preventing untrusted code from executing on the protected runner and preserving deployment keys and other credentials from being unintentionally accessed. To ensure that jobs intended to be executed on protected runners do not use regular runners, they must be tagged accordingly.
Total running time for a given pipeline excludes retries and pending (queued) time.
Each job is represented as a
Period, which consists of:
Period#first(when the job started).
Period#last(when the job finished).
A simple example is:
- A (1, 3)
- B (2, 4)
- C (6, 7)
In the example:
- A begins at 1 and ends at 3.
- B begins at 2 and ends at 4.
- C begins at 6 and ends at 7.
Visually, it can be viewed as:
0 1 2 3 4 5 6 7 AAAAAAA BBBBBBB CCCC
The union of A, B, and C is (1, 4) and (6, 7). Therefore, the total running time is:
(4 - 1) + (7 - 6) => 4
Pipelines can be complex structures with many sequential and parallel jobs.
To make it easier to understand the flow of a pipeline, GitLab has pipeline graphs for viewing pipelines and their statuses.
Pipeline graphs can be displayed as a large graph or a miniature representation, depending on the page you access the graph from.
GitLab capitalizes the stages’ names in the pipeline graphs.
- Visualization improvements introduced in GitLab 13.11.
The pipeline details page displays the full pipeline graph of all the jobs in the pipeline.
You can group the jobs by:
Stage, which arranges jobs in the same stage together in the same column:
Multi-project pipeline graphs help you visualize the entire pipeline, including all cross-project inter-dependencies.
If a stage contains more than 100 jobs, only the first 100 jobs are listed in the pipeline graph. The remaining jobs still run as normal. To see the jobs:
- Select the pipeline, and the jobs are listed on the right side of the pipeline details page.
- On the left sidebar, select CI/CD > Jobs.
You can arrange jobs in the pipeline graph based on their
Jobs in the leftmost column run first, and jobs that depend on them are grouped in the next columns.
test-job1 depends only on jobs in the first column, so it displays
in the second column from the left.
deploy-job1 depends on jobs in both the first
and second column and displays in the third column:
To add lines that show the
needs relationships between jobs, select the Show dependencies toggle.
These lines are similar to the needs visualization:
To see the full
needs dependency tree for a job, hover over it:
Pipeline mini graphs take less space and can tell you at a quick glance if all jobs passed or something failed. The pipeline mini graph can be found when you go to:
- The pipelines index page.
- A single commit page.
- A merge request page.
- The pipeline editor, in GitLab 14.5 and later.
Pipeline mini graphs allow you to see all related jobs for a single commit and the net result of each stage of your pipeline. This allows you to quickly see what failed and fix it.
Pipeline mini graphs only display jobs by stage.
Stages in pipeline mini graphs are expandable. Hover your mouse over each stage to see the name and status, and select a stage to expand its jobs list.
|Mini graph||Mini graph expanded|
Pipeline analytics are available on the CI/CD Analytics page.
Pipeline status and test coverage report badges are available and configurable for each project. For information on adding pipeline badges to projects, see Pipeline badges.
GitLab provides API endpoints to: