Status | Authors | Coach | DRIs | Owning Stage | Created |
---|---|---|---|---|---|
proposed |
@furkanayhan
|
@ayufan
|
@jreporter
@cheryl.li
| devops verify | 2023-05-15 |
- Summary
- Goals
- Non-Goals
- Motivation
- Technical expectations
- Proposal
- Design and implementation details
Future of CI Pipeline Processing
Summary
GitLab CI is one of the oldest and most complex features in GitLab. Over the years its YAML syntax has considerably grown in size and complexity. In order to keep the syntax highly stable over the years, we have primarily been making additive changes on top of the existing design and patterns. Our user base has grown exponentially over the past years. With that, the need to support their use cases and customization of the workflows.
While delivering huge value over the years, the various additive changes to the syntax have also caused some surprising behaviors in the pipeline processing logic. Some keywords accumulated a number of responsibilities, and some ambiguous overlaps were discovered among keywords and subtle differences in behavior were introduced over time. The current implementation and YAML syntax also make it challenging to implement new features.
In this design document, we will discuss the problems and propose a new architecture for pipeline processing. Most of these problems have been discussed before in the “Restructure CI job when keyword” epic.
Goals
- We want to make the pipeline processing more understandable, predictable and consistent.
- We want to unify the behaviors of DAG and STAGE. STAGE can be written as DAG and vice versa.
- We want to decouple the manual jobs’ blocking behavior from the
allow_failure
keyword. - We want to clarify the responsibilities of the
when
keyword.
Non-Goals
We will not discuss how to avoid breaking changes for now.
Motivation
The list of problems is the main motivation for this design document.
Problem 1: The responsibility of the when
keyword
Right now, the when
keyword has many responsibilities;
on_success
(default): Run the job only when no jobs in earlier stages fail or haveallow_failure: true
.on_failure
: Run the job only when at least one job in an earlier stage fails. A job in an earlier stage withallow_failure: true
is always considered successful.never
: Don’t run the job regardless of the status of jobs in earlier stages. Can only be used in arules
section orworkflow: rules
.always
: Run the job regardless of the status of jobs in earlier stages. Can also be used inworkflow:rules
.manual
: Run the job only when triggered manually.delayed
: Delay the execution of a job for a specified duration.
It answers three questions;
- What’s required to run? =>
on_success
,on_failure
,always
- How to run? =>
manual
,delayed
- Add to the pipeline? =>
never
As a result, for example; we cannot create a manual
job with when: on_failure
.
This can be useful when persona wants to create a job that is only available on failure, but needs to be manually played.
For example; publishing failures to dedicated page or dedicated external service.
Problem 2: Abuse of the allow_failure
keyword
We control the blocker behavior of a manual job by the allow_failure
keyword.
Actually, it has other responsibilities; “determine whether a pipeline should continue running when a job fails”.
Currently, a manual job;
- is not a blocker when it has
allow_failure: true
(by default) - a blocker when it has
allow_failure: false
.
As a result, for example; we cannot create a manual
job that is allow_failure: false
and not a blocker.
job1:
stage: test
when: manual
allow_failure: true # default
job2:
stage: deploy
Currently;
-
job1
is skipped. -
job2
runs becausejob1
is ignored since it hasallow_failure: true
. - When we run/play
job1
;- if it fails, it’s marked as “success with warning”.
allow_failure
with rules
allow_failure
becomes more confusing when using rules
.
From docs:
The default behavior of
allow_failure
changes to true withwhen: manual
. However, if you usewhen: manual
withrules
,allow_failure
defaults tofalse
.
From docs:
The default value for
allow_failure
is:
true
for manual jobs.false
for jobs that usewhen: manual
insiderules
.false
in all other cases.
For example;
job1:
script: ls
when: manual
job2:
script: ls
rules:
- if: $ALWAYS_TRUE
when: manual
job1
and job2
behave differently;
-
job1
is not a blocker because it hasallow_failure: true
by default. -
job2
is a blockerrules: when: manual
does not returnallow_failure: true
by default.
Problem 3: Different behaviors in DAG/needs
The main behavioral difference between DAG and STAGE is about the “skipped” and “ignored” states.
Background information:
- skipped:
- When a job is
when: on_success
and its previous status is failed, it’s skipped. - When a job is
when: on_failure
and its previous status is not “failed”, it’s skipped.
- When a job is
- ignored:
- When a job is
when: manual
withallow_failure: true
, it’s ignored.
- When a job is
Problem:
The skipped
and ignored
states are considered successful in the STAGE processing but not in the DAG processing.
Problem 3.1. Handling of ignored status with manual jobs
Example 1:
build:
stage: build
script: exit 0
when: manual
allow_failure: true # by default
test:
stage: test
script: exit 0
needs: [build]
-
build
is ignored (skipped) because it’swhen: manual
withallow_failure: true
. -
test
is skipped because “ignored” is not a successful state in the DAG processing.
Example 2:
build:
stage: build
script: exit 0
when: manual
allow_failure: true # by default
test:
stage: test
script: exit 0
-
build
is ignored (skipped) because it’swhen: manual
withallow_failure: true
. -
test2
runs and succeeds.
Problem 3.2. Handling of skipped status with when: on_failure
Example 1:
build_job:
stage: build
script: exit 1
test_job:
stage: test
script: exit 0
rollback_job:
stage: deploy
needs: [build_job, test_job]
script: exit 0
when: on_failure
-
build_job
runs and fails. -
test_job
is skipped. - Even though
rollback_job
iswhen: on_failure
and there is a failed job, it is skipped because theneeds
list has a “skipped” job.
Example 2:
build_job:
stage: build
script: exit 1
test_job:
stage: test
script: exit 0
rollback_job:
stage: deploy
script: exit 0
when: on_failure
-
build_job
runs and fails. -
test_job
is skipped. -
rollback_job
runs because there is a failed job before.
Problem 4: The skipped and ignored states
Let’s assume that we solved the problem 3 and the “skipped” and “ignored” states are not different in DAG and STAGE. How should they behave in general? Are they successful or not? Should “skipped” and “ignored” be different? Let’s examine some examples;
Example 4.1. The ignored status with manual jobs
build:
stage: build
script: exit 0
when: manual
allow_failure: true # by default
test:
stage: test
script: exit 0
-
build
is in the “manual” state but considered as “skipped” (ignored) for the pipeline processing. -
test
runs because “skipped” is a successful state.
Alternatively;
build1:
stage: build
script: exit 0
when: manual
allow_failure: true # by default
build2:
stage: build
script: exit 0
test:
stage: test
script: exit 0
-
build1
is in the “manual” state but considered as “skipped” (ignored) for the pipeline processing. -
build2
runs and succeeds. -
test
runs because “success” + “skipped” is a successful state.
Example 4.2. The skipped status with when: on_failure
build:
stage: build
script: exit 0
when: on_failure
test:
stage: test
script: exit 0
-
build
is skipped because it’swhen: on_failure
and its previous status is not “failed”. -
test
runs because “skipped” is a successful state.
Alternatively;
build1:
stage: build
script: exit 0
when: on_failure
build2:
stage: build
script: exit 0
test:
stage: test
script: exit 0
-
build1
is skipped because it’swhen: on_failure
and its previous status is not “failed”. -
build2
runs and succeeds. -
test
runs because “success” + “skipped” is a successful state.
Problem 5: The dependencies
keyword
The dependencies
keyword is used to define a list of jobs to fetch
artifacts from. It is a shared responsibility with the needs
keyword.
Moreover, they can be used together in the same job. We may not need to discuss all possible scenarios but this example
is enough to show the confusion;
test2:
script: exit 0
dependencies: [test1]
needs:
- job: test1
artifacts: false
Information 1: Canceled jobs
Are a canceled job and a failed job the same? They have many differences so we could easily say “no”. However, they have one similarity; they can be “allowed to fail”.
Let’s define their differences first;
- A canceled job;
- It is not a finished job.
- Canceled is a user requested interruption of the job. The intent is to abort the job or stop pipeline processing as soon as possible.
- We don’t know the result, there is no artifacts, etc.
- Since it’s never run, the
after_script
is not run. - Its eventual state is “canceled” so no job can run after it.
- There is no
when: on_canceled
. - Even
when: always
is not run.
- There is no
- A failed job;
- It is a machine response of the CI system to executing the job content. It indicates that execution failed for some reason.
- It is equal answer of the system to success. The fact that something is failed is relative, and might be desired outcome of CI execution, like in when executing tests that some are failing.
- We know the result and there can be artifacts.
-
after_script
is run. - Its eventual state is “failed” so subsequent jobs can run depending on their
when
values.-
when: on_failure
andwhen: always
are run.
-
The one similarity is; they can be “allowed to fail”.
build:
stage: build
script: sleep 10
allow_failure: true
test:
stage: test
script: exit 0
when: on_success
- If
build
runs and getscanceled
, thentest
runs. - If
build
runs and getsfailed
, thentest
runs.
An idea on using canceled
instead of failed
for some cases
There is another aspect. We often drop jobs with a failure_reason
before they get executed,
for example when the namespace ran out of Compute Credits (CI minutes) or when limits are exceeded.
Dropping jobs in the failed
state has been handy because we could communicate to the user the failure_reason
for better feedback. When canceling jobs for various reasons we don’t have a way to indicate that.
We cancel jobs because the user ran out of Compute Credits while the pipeline was running,
or because the pipeline is auto-canceled by another pipeline or other reasons.
If we had a stop_reason
instead of failure_reason
we could use that for both cancelled and failed jobs
and we could also use the canceled
status more appropriately.
Information 2: Empty state
We recently updated the documentation of
the when
keyword for clarification;
on_success
: Run the job only when no jobs in earlier stages fail or haveallow_failure: true
.on_failure
: Run the job only when at least one job in an earlier stage fails.
For example;
test1:
when: on_success
script: exit 0
# needs: [] would lead to the same result
test2:
when: on_failure
script: exit 0
# needs: [] would lead to the same result
-
test1
runs because there is no job failed in the previous stages. -
test2
does not run because there is no job failed in the previous stages.
The on_success
means that “nothing failed”, it does not mean that everything succeeded.
The same goes to on_failure
, it does not mean that everything failed, but does mean that “something failed”.
This semantic goes by a expectation that your pipeline succeeds, and this is happy path.
Not that your pipeline fails, because then it requires user intervention to fix it.
Technical expectations
All proposals or future decisions must follow these goals;
- The
allow_failure
keyword must only responsible for marking failed jobs as “success with warning”.- Why: It should not have another responsibility, such as determining a manual job is a blocker or not.
- How: Another keyword will be introduced to control the blocker behavior of a manual job.
- With
allow_failure
, canceled jobs must not be marked as “success with warning”.- Why: “canceled” is a different state than “failed”.
- How: Canceled with
allow_failure: true
jobs will not be marked as “success with warning”.
- The
when
keyword must only answer the question “What’s required to run?”. And it must be the only source of truth for deciding if a job should run or not. - The
when
keyword must not control if a job is added to the pipeline or not.- Why: It is not its responsibility.
- How: Another keyword will be introduced to control if a job is added to the pipeline or not.
- The “skipped” and “ignored” states must be reconsidered.
- TODO: We need to discuss this more.
- A new keyword structure must be introduced to specify if a job is an “automatic”, “manual”, or “delayed” job.
- Why: It is not the responsibility of the
when
keyword. - How: A new keyword will be introduced to control the behavior of a job.
- Why: It is not the responsibility of the
- The
needs
keyword must only control the order of the jobs. It must not be used to control the behavior of the jobs or to decide if a job should run or not. The DAG and STAGE behaviors must be the same.- Why: It leads to different behaviors and confuses users.
- How: The
needs
keyword will only define previous jobs, like stage does.
- The
needs
anddependencies
keywords must not be used together in the same job.- Why: It is confusing.
- How: The
needs
anddependencies
keywords will be mutually exclusive.
Proposal
N/A
Design and implementation details
N/A