# https://circleci.com/docs/2.0/ # We use version 2.1 of CircleCI's configuration format (the docs are still at # the 2.0 link) in order to have access to Windows executors. This means we # can't use dots in job names anymore. They have a new "parameters" feature # that is supposed to remove the need to have version numbers in job names (the # source of our dots), but switching to that is going to be a bigger refactor: # # https://discuss.circleci.com/t/v2-1-job-name-validation/31123 # https://circleci.com/docs/2.0/reusing-config/ # version: 2.1 workflows: ci: jobs: # Start with jobs testing various platforms. - "debian-9": {} - "debian-10": requires: - "debian-9" - "ubuntu-20-04": {} - "ubuntu-18-04": requires: - "ubuntu-20-04" - "ubuntu-16-04": requires: - "ubuntu-20-04" - "fedora-29": {} - "fedora-28": requires: - "fedora-29" - "centos-8": {} - "nixos-19-09": {} - "nixos-21-05": {} # Test against PyPy 2.7 - "pypy27-buster": {} # Just one Python 3.6 configuration while the port is in-progress. - "python36": {} # Other assorted tasks and configurations - "lint": {} - "codechecks3": {} - "pyinstaller": {} - "deprecations": {} - "c-locale": {} # Any locale other than C or UTF-8. - "another-locale": {} - "integration": requires: # If the unit test suite doesn't pass, don't bother running the # integration tests. - "debian-9" - "typechecks": {} - "docs": {} images: # Build the Docker images used by the ci jobs. This makes the ci jobs # faster and takes various spurious failures out of the critical path. triggers: # Build once a day - schedule: cron: "0 0 * * *" filters: branches: only: - "master" jobs: # Every job that pushes a Docker image from Docker Hub needs to provide # credentials. Use this first job to define a yaml anchor that can be # used to supply a CircleCI job context which makes Docker Hub # credentials available in the environment. # # Contexts are managed in the CircleCI web interface: # # https://app.circleci.com/settings/organization/github/tahoe-lafs/contexts - "build-image-debian-10": &DOCKERHUB_CONTEXT context: "dockerhub-auth" - "build-image-debian-9": <<: *DOCKERHUB_CONTEXT - "build-image-ubuntu-16-04": <<: *DOCKERHUB_CONTEXT - "build-image-ubuntu-18-04": <<: *DOCKERHUB_CONTEXT - "build-image-ubuntu-20-04": <<: *DOCKERHUB_CONTEXT - "build-image-fedora-28": <<: *DOCKERHUB_CONTEXT - "build-image-fedora-29": <<: *DOCKERHUB_CONTEXT - "build-image-centos-8": <<: *DOCKERHUB_CONTEXT - "build-image-pypy27-buster": <<: *DOCKERHUB_CONTEXT - "build-image-python36-ubuntu": <<: *DOCKERHUB_CONTEXT jobs: dockerhub-auth-template: # This isn't a real job. It doesn't get scheduled as part of any # workflow. Instead, it's just a place we can hang a yaml anchor to # finish the Docker Hub authentication configuration. Workflow jobs using # the DOCKERHUB_CONTEXT anchor will have access to the environment # variables used here. These variables will allow the Docker Hub image # pull to be authenticated and hopefully avoid hitting and rate limits. docker: &DOCKERHUB_AUTH - image: "null" auth: username: $DOCKERHUB_USERNAME password: $DOCKERHUB_PASSWORD steps: - run: name: "CircleCI YAML schema conformity" command: | # This isn't a real command. We have to have something in this # space, though, or the CircleCI yaml schema validator gets angry. # Since this job is never scheduled this step is never run so the # actual value here is irrelevant. lint: docker: - <<: *DOCKERHUB_AUTH image: "circleci/python:2" steps: - "checkout" - run: name: "Install tox" command: | pip install --user tox - run: name: "Static-ish code checks" command: | ~/.local/bin/tox -e codechecks codechecks3: docker: - <<: *DOCKERHUB_AUTH image: "circleci/python:3" steps: - "checkout" - run: name: "Install tox" command: | pip install --user tox - run: name: "Static-ish code checks" command: | ~/.local/bin/tox -e codechecks3 pyinstaller: docker: - <<: *DOCKERHUB_AUTH image: "circleci/python:2" steps: - "checkout" - run: name: "Install tox" command: | pip install --user tox - run: name: "Make PyInstaller executable" command: | ~/.local/bin/tox -e pyinstaller - run: # To verify that the resultant PyInstaller-generated binary executes # cleanly (i.e., that it terminates with an exit code of 0 and isn't # failing due to import/packaging-related errors, etc.). name: "Test PyInstaller executable" command: | dist/Tahoe-LAFS/tahoe --version debian-9: &DEBIAN docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/debian:9-py2.7" user: "nobody" environment: &UTF_8_ENVIRONMENT # In general, the test suite is not allowed to fail while the job # succeeds. But you can set this to "yes" if you want it to be # otherwise. ALLOWED_FAILURE: "no" # Tell Hypothesis which configuration we want it to use. TAHOE_LAFS_HYPOTHESIS_PROFILE: "ci" # Tell the C runtime things about character encoding (mainly to do with # filenames and argv). LANG: "en_US.UTF-8" # Select a tox environment to run for this job. TAHOE_LAFS_TOX_ENVIRONMENT: "py27" # Additional arguments to pass to tox. TAHOE_LAFS_TOX_ARGS: "" # The path in which test artifacts will be placed. ARTIFACTS_OUTPUT_PATH: "/tmp/artifacts" # Convince all of our pip invocations to look at the cached wheelhouse # we maintain. WHEELHOUSE_PATH: &WHEELHOUSE_PATH "/tmp/wheelhouse" PIP_FIND_LINKS: "file:///tmp/wheelhouse" # Upload the coverage report. UPLOAD_COVERAGE: "" # pip cannot install packages if the working directory is not readable. # We want to run a lot of steps as nobody instead of as root. working_directory: "/tmp/project" steps: - "checkout" - run: &SETUP_VIRTUALENV name: "Setup virtualenv" command: | /tmp/project/.circleci/setup-virtualenv.sh \ "/tmp/venv" \ "/tmp/project" \ "${WHEELHOUSE_PATH}" \ "${TAHOE_LAFS_TOX_ENVIRONMENT}" \ "${TAHOE_LAFS_TOX_ARGS}" - run: &RUN_TESTS name: "Run test suite" command: | /tmp/project/.circleci/run-tests.sh \ "/tmp/venv" \ "/tmp/project" \ "${ALLOWED_FAILURE}" \ "${ARTIFACTS_OUTPUT_PATH}" \ "${TAHOE_LAFS_TOX_ENVIRONMENT}" \ "${TAHOE_LAFS_TOX_ARGS}" # trial output gets directed straight to a log. avoid the circleci # timeout while the test suite runs. no_output_timeout: "20m" - store_test_results: &STORE_TEST_RESULTS path: "/tmp/artifacts/junit" - store_artifacts: &STORE_TEST_LOG # Despite passing --workdir /tmp to tox above, it still runs trial # in the project source checkout. path: "/tmp/project/_trial_temp/test.log" - store_artifacts: &STORE_ELIOT_LOG # Despite passing --workdir /tmp to tox above, it still runs trial # in the project source checkout. path: "/tmp/project/eliot.log" - store_artifacts: &STORE_OTHER_ARTIFACTS # Store any other artifacts, too. This is handy to allow other jobs # sharing most of the definition of this one to be able to # contribute artifacts easily. path: "/tmp/artifacts" - run: &SUBMIT_COVERAGE name: "Submit coverage results" command: | if [ -n "${UPLOAD_COVERAGE}" ]; then /tmp/venv/bin/codecov fi debian-10: <<: *DEBIAN docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/debian:10-py2.7" user: "nobody" pypy27-buster: <<: *DEBIAN docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/pypy:buster-py2" user: "nobody" environment: <<: *UTF_8_ENVIRONMENT # We don't do coverage since it makes PyPy far too slow: TAHOE_LAFS_TOX_ENVIRONMENT: "pypy27" # Since we didn't collect it, don't upload it. UPLOAD_COVERAGE: "" c-locale: <<: *DEBIAN environment: <<: *UTF_8_ENVIRONMENT LANG: "C" another-locale: <<: *DEBIAN environment: <<: *UTF_8_ENVIRONMENT # aka "Latin 1" LANG: "en_US.ISO-8859-1" deprecations: <<: *DEBIAN environment: <<: *UTF_8_ENVIRONMENT # Select the deprecations tox environments. TAHOE_LAFS_TOX_ENVIRONMENT: "deprecations,upcoming-deprecations" # Put the logs somewhere we can report them. TAHOE_LAFS_WARNINGS_LOG: "/tmp/artifacts/deprecation-warnings.log" # The deprecations tox environments don't do coverage measurement. UPLOAD_COVERAGE: "" integration: <<: *DEBIAN environment: <<: *UTF_8_ENVIRONMENT # Select the integration tests tox environments. TAHOE_LAFS_TOX_ENVIRONMENT: "integration" # Disable artifact collection because py.test can't produce any. ARTIFACTS_OUTPUT_PATH: "" steps: - "checkout" # DRY, YAML-style. See the debian-9 steps. - run: *SETUP_VIRTUALENV - run: *RUN_TESTS ubuntu-16-04: <<: *DEBIAN docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/ubuntu:16.04-py2.7" user: "nobody" ubuntu-18-04: &UBUNTU_18_04 <<: *DEBIAN docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/ubuntu:18.04-py2.7" user: "nobody" python36: <<: *UBUNTU_18_04 docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/ubuntu:18.04-py3" user: "nobody" environment: <<: *UTF_8_ENVIRONMENT # The default trial args include --rterrors which is incompatible with # this reporter on Python 3. So drop that and just specify the # reporter. TAHOE_LAFS_TRIAL_ARGS: "--reporter=subunitv2-file" TAHOE_LAFS_TOX_ENVIRONMENT: "py36" ubuntu-20-04: <<: *DEBIAN docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/ubuntu:20.04" user: "nobody" centos-8: &RHEL_DERIV docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/centos:8-py2" user: "nobody" environment: *UTF_8_ENVIRONMENT # pip cannot install packages if the working directory is not readable. # We want to run a lot of steps as nobody instead of as root. working_directory: "/tmp/project" steps: - "checkout" - run: *SETUP_VIRTUALENV - run: *RUN_TESTS - store_test_results: *STORE_TEST_RESULTS - store_artifacts: *STORE_TEST_LOG - store_artifacts: *STORE_ELIOT_LOG - store_artifacts: *STORE_OTHER_ARTIFACTS - run: *SUBMIT_COVERAGE fedora-28: <<: *RHEL_DERIV docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/fedora:28-py" user: "nobody" fedora-29: <<: *RHEL_DERIV docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/fedora:29-py" user: "nobody" nixos-19-09: &NIXOS docker: # Run in a highly Nix-capable environment. - <<: *DOCKERHUB_AUTH image: "nixorg/nix:circleci" environment: NIX_PATH: "nixpkgs=https://github.com/NixOS/nixpkgs-channels/archive/nixos-19.09-small.tar.gz" SOURCE: "nix/" steps: - "checkout" - "run": name: "Build and Test" command: | # CircleCI build environment looks like it has a zillion and a # half cores. Don't let Nix autodetect this high core count # because it blows up memory usage and fails the test run. Pick a # number of cores that suites the build environment we're paying # for (the free one!). # # Also, let it run more than one job at a time because we have to # build a couple simple little dependencies that don't take # advantage of multiple cores and we get a little speedup by doing # them in parallel. nix-build --cores 3 --max-jobs 2 "$SOURCE" nixos-21-05: <<: *NIXOS environment: # Note this doesn't look more similar to the 19.09 NIX_PATH URL because # there was some internal shuffling by the NixOS project about how they # publish stable revisions. NIX_PATH: "nixpkgs=https://github.com/NixOS/nixpkgs/archive/d32b07e6df276d78e3640eb43882b80c9b2b3459.tar.gz" SOURCE: "nix/py3.nix" typechecks: docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/ubuntu:18.04-py3" steps: - "checkout" - run: name: "Validate Types" command: | /tmp/venv/bin/tox -e typechecks docs: docker: - <<: *DOCKERHUB_AUTH image: "tahoelafsci/ubuntu:18.04-py3" steps: - "checkout" - run: name: "Build documentation" command: | /tmp/venv/bin/tox -e docs build-image: &BUILD_IMAGE # This is a template for a job to build a Docker image that has as much of # the setup as we can manage already done and baked in. This cuts down on # the per-job setup time the actual testing jobs have to perform - by # perhaps 10% - 20%. # # https://circleci.com/blog/how-to-build-a-docker-image-on-circleci-2-0/ docker: - <<: *DOCKERHUB_AUTH image: "docker:17.05.0-ce-git" environment: DISTRO: "tahoelafsci/:foo-py2" TAG: "tahoelafsci/distro:-py2" PYTHON_VERSION: "tahoelafsci/distro:tag-py