Skip to content

Latest commit

 

History

History
219 lines (181 loc) · 21.5 KB

CI.rst

File metadata and controls

219 lines (181 loc) · 21.5 KB

Continuous Integration is important component of making Apache Airflow robust and stable. We are running a lot of tests for every pull request, for master and v1-10-test branches and regularly as CRON jobs.

Our execution environment for CI is GitHub Actions. GitHub Actions (GA) are very well integrated with GitHub code and Workflow and it has evolved fast in 2019/202 to become a fully-fledged CI environment, easy to use and develop for, so we decided to switch to it. Our previous CI system is Travis CI.

However part of the philosophy we have is that we are not tightly coupled with any of the CI environments we use. Most of our CI jobs are written as bash scripts which are executed as steps in the CI jobs and we are mapping all the CI-specific environment variables to generic "CI" variables. The only two places where CI-specific code might be are:

  • CI-specific declaration file (for example it is .github/workflows/ci.yml for GitHub Actions
  • The get_environment_for_builds_on_ci function in scripts/ci/libraries/_build_images.sh where mapping is performed from the CI-environment specific to generic values. Example for that is CI_EVENT_TYPE variable which determines whether we are running a push. schedule or pull_request kind of CI job. For GitHub Action those are values mapped from GITHUB_EVENT_NAME variable, for Travis several other variables were used to determine it (TRAVIS_PULL_REQUEST, TRAVIS_EVENT_TYPE).

CI architecture of Apache Airflow

The following components are part of the CI infrastructure

  • Apache Airflow Code Repository - our code repository at https://github.com/apache/airflow
  • Apache Airflow Forks - forks of the Apache Airflow Code Repository from which contributors make Pull Requests
  • GitHub Actions - (GA) UI + execution engine for our jobs
  • GA CRON trigger - GitHub Actions CRON triggering our jobs
  • GA Workers - virtual machines running our jobs at GitHub Actions (max 20 in parallel)
  • GitHub Private Image Registry- image registry used as build cache for CI jobs. It is at https://docker.pkg.github.com/apache/airflow/airflow
  • DockerHub Public Image Registry - publicly available image registry at DockerHub. It is at https://hub.docker.com/repository/docker/apache/airflow
  • DockerHub Build Workers - virtual machines running build jibs at DockerHub
  • Official Images (future) - these are official images that are prominently visible in DockerHub. We aim our images to become official images so that you will be able to pull them with docker pull apache-airflow

The following CI Job run types are currently run for Apache Airflow (run by ci.yaml workflow and quarantined.yaml workflows) and each of the run types have different purpose and context.

Those runs are results of PR from the forks made by contributors. Most builds for Apache Airflow fall into this category. They are executed in the context of the "Fork", not main Airflow Code Repository which means that they have only "read" permission to all the GitHub resources (container registry, code repository). This is necessary as the code in those PRs (including CI job definition) might be modified by people who are not committers for the Apache Airflow Code Repository.

The main purpose of those jobs is to check if PR builds cleanly, if the test run properly and if the PR is ready to review and merge. The runs are using cached images from the Private GitHub registry - CI, Production Images as well as base Python images that are also cached in the Private GitHub registry. Also for those builds we only execute Python tests if important files changed (so for example if it is doc-only change, no tests will be executed.

Those runs are results of direct pushes done by the committers or as result of merge of a Pull Request by the committers. Those runs execute in the context of the Apache Airflow Code Repository and have also write permission for GitHub resources (container registry, code repository). The main purpose for the run is to check if the code after merge still holds all the assertions - like whether it still builds, all tests are green.

This is needed because some of the conflicting changes from multiple PRs might cause build and test failures after merge even if they do not fail in isolation. Also those runs are already reviewed and confirmed by the committers so they can be used to do some housekeeping: - pushing most recent image build in the PR to the Github Private Registry (for caching) - upgrading to latest constraints and pushing those constraints if all tests succeed - refresh latest Python base images in case new patch-level is released

The housekeeping is important - Python base images are refreshed with varying frequency (once every few months usually but sometimes several times per week) with the latest security and bug fixes. Those patch level images releases can occasionally break Airflow builds (specifically Docker image builds based on those images) therefore in PRs we only use latest "good" python image that we store in the private GitHub cache. The direct push/master builds are not using registry cache to pull the python images - they are directly pulling the images from DockerHub, therefore they will try the latest images after they are released and in case they are fine, CI Docker image is build and tests are passing - those jobs will push the base images to the private GitHub Registry so that they be used by subsequent PR runs.

Those runs are results of (nightly) triggered job - only for master branch. The main purpose of the job is to check if there was no impact of external dependency changes on the Apache Airflow code (for example transitive dependencies released that fail the build). It also checks if the Docker images can be build from the scratch (again - to see if some dependencies have not changed - for example downloaded package releases etc. Another reason for the nightly build is that the builds tags most recent master with nightly-master tag so that DockerHub build can pick up the moved tag and prepare a nightly public master build in the DockerHub registry. The v1-10-test branch images are build in DockerHub when pushing v1-10-stable manually.

All runs consist of the same jobs, but the jobs behave slightly differently or they are skipped in different run categories. Here is a summary of the run categories with regards of the jobs they are running. Those jobs often have matrix run strategy which runs several different variations of the jobs (with different Backend type / Python version, type of the tests to run for example). The following chapter describes the workflows that execute for each run.

This workflow is a regular workflow that performs the regular checks - none of the jobs should fail. The tests to run do not contain quarantined tests.

Job Description Pull Request Run Direct Push/Merge Run
Scheduled Run
(*) Builds all images from scratch
Cancel previous workflow Cancels the previously running workflow run if there is one running Yes Yes Yes *
Static checks Performs static checks Yes Yes Yes *
Docs Builds documentation Yes Yes Yes *
Prepare Backport packages Prepares Backport Packages for 1.10.* Yes Yes Yes *
Trigger tests Checks if tests should be triggered Yes Yes Yes *
Build Prod Image Builds production image Yes Yes Yes *
Tests Run all the combinations of Pytest tests for Python code Yes (if tests-triggered) Yes Yes *
Tests Kubernetes Run Kubernetes test Yes (if tests-triggered) Yes Yes *
Test OpenAPI client gen Tests if OpenAPIClient continues to generate Yes Yes Yes *
Helm tests Runs tests for the Helm chart Yes Yes Yes *
Constraints Upgrade constraints to latest eagerly pushed ones (only if tests successful)
Yes Yes *
Constraints push Pushes updated constraints (only if tests successful)
Yes Yes *
Push Prod images Pushes production images to GitHub Private Image Registry to cache the build images for following runs
Yes
Push CI images Pushes CI images to GitHub Private Image Registry to cache the build images for following runs
Yes
Tag Repo nightly Tags the repository with nightly tagIt is a lightweight tag that moves nightly
Yes. Triggers DockerHub build for public registry

This workflow runs only quarantined tests. Those tests do not fail the build even if some tests fail (only if the whole pytest execution fails). Instead this workflow updates one of the issues where we keep status of quarantined tests. Once the test succeeds in NUM_RUNS subsequent runs, it is marked as stable and can be removed from quarantine. You can read more about quarantine in TESTING.rst

The issues are only updated if the test is run as direct push or scheduled run and only in the apache/airflow repository - so that the issues are not updated in forks.

The issues that gets updated are different for different branches:

Job Description Pull Request Run Direct Push/Merge Run Scheduled Run
Cancel previous workflow Cancels the previously running workflow run if there is one running Yes Yes Yes *
Trigger tests Checks if tests should be triggered Yes Yes Yes *
Quarantined tests Those are tests that are flaky and we need to fix them Yes (if tests-triggered) Yes (Updates quarantine issue) Yes * (updates quarantine issue)

This workflow is run only on schedule (every 5 minutes) it's only purpose is to cancel other running CI Build workflows if important jobs failed in those runs. This is to save runners for other runs in case we know that the build will not succeed anyway without some basic fixes to static checks or documentation - effectively implementing missing "fail-fast" (on a job level) in Github Actions similar to fail-fast in matrix strategy.

The jobs that are considered as "fail-fast" are:

  • Static checks
  • Docs
  • Prepare Backport packages
  • Helm tests
  • Build Prod Image
  • TTest OpenAPI client gen