##################### # Default Variables # ##################### stages: - upload_cache - pre_check - build - assign_test - build_doc - target_test - host_test - test_deploy - deploy - post_deploy - retry_failed_jobs variables: # System environment # Common parameters for the 'make' during CI tests MAKEFLAGS: "-j5 --no-keep-going" # GitLab-CI environment # Thanks to pack-objects cache, clone strategy should behave faster than fetch # so we pick "clone" as default git strategy # Shiny runners by default remove the CI_PROJECT_DIR every time at the beginning of one job # and clone with a --depth=1 # Brew runners will fetch from locally mirror first, and cache the local CI_PROJECT_DIR # In conclusion # - set GIT_STRATEGY: "clone" to shiny runners # - set GIT_STRATEGY: "fetch" to brew runners GIT_STRATEGY: clone # we will download archive for each submodule instead of clone. # we don't do "recursive" when fetch submodule as they're not used in CI now. GIT_SUBMODULE_STRATEGY: none # since we're using merged-result pipelines, the last commit should work for most cases GIT_DEPTH: 1 # --no-recurse-submodules: we use cache for submodules # --prune --prune-tags: in case remote branch or tag is force pushed GIT_FETCH_EXTRA_FLAGS: "--no-recurse-submodules --prune --prune-tags" # we're using .cache folder for caches GIT_CLEAN_FLAGS: -ffdx -e .cache/ LATEST_GIT_TAG: v6.0-dev SUBMODULE_FETCH_TOOL: "tools/ci/ci_fetch_submodule.py" # by default we will fetch all submodules # jobs can overwrite this variable to only fetch submodules they required # set to "none" if don't need to fetch submodules SUBMODULES_TO_FETCH: "all" # tell build system do not check submodule update as we download archive instead of clone IDF_SKIP_CHECK_SUBMODULES: 1 IDF_PATH: "$CI_PROJECT_DIR" V: "0" CHECKOUT_REF_SCRIPT: "$CI_PROJECT_DIR/tools/ci/checkout_project_ref.py" # Docker images ESP_ENV_IMAGE: "${CI_DOCKER_REGISTRY}/esp-env-v6.0:1" ESP_IDF_DOC_ENV_IMAGE: "${CI_DOCKER_REGISTRY}/esp-idf-doc-env-v6.0:1-1" TARGET_TEST_ENV_IMAGE: "${CI_DOCKER_REGISTRY}/target-test-env-v6.0:1" SONARQUBE_SCANNER_IMAGE: "${CI_DOCKER_REGISTRY}/sonarqube-scanner:5" PRE_COMMIT_IMAGE: "${CI_DOCKER_REGISTRY}/esp-idf-pre-commit:1" # cache python dependencies PIP_CACHE_DIR: "$CI_PROJECT_DIR/.cache/pip" # Set this variable to the branch of idf-constraints repo in order to test a custom Python constraint file. The # branch name must be without the remote part ("origin/"). Keep the variable empty in order to use the constraint # file from https://dl.espressif.com/dl/esp-idf. CI_PYTHON_CONSTRAINT_BRANCH: "" # Update the filename for a specific ESP-IDF release. It is used only with CI_PYTHON_CONSTRAINT_BRANCH. CI_PYTHON_CONSTRAINT_FILE: "espidf.constraints.v6.0.txt" # Set this variable to repository name of a Python tool you wish to install and test in the context of ESP-IDF CI. # Keep the variable empty when not used. CI_PYTHON_TOOL_REPO: "" # Set this variable to the branch of a Python tool repo specified in CI_PYTHON_TOOL_REPO. The # branch name must be without the remote part ("origin/"). Keep the variable empty when not used. # This is used only if CI_PYTHON_TOOL_REPO is not empty. CI_PYTHON_TOOL_BRANCH: "" # Set this variable to Clang toolchain distro URL to be used. # NOTE: We have separate toolchains for Xtensa and RISCV, therefore jobs for one arch will fail. # This is OK as far as we use CI_CLANG_DISTRO_URL for pre-release tests purposes only. # Keep the variable empty when not used. CI_CLANG_DISTRO_URL: "" # Set this variable to specify the file name for the known failure cases. KNOWN_FAILURE_CASES_FILE_NAME: "master.txt" IDF_CI_BUILD: 1 # ccache settings # some settings need to set in .gitlab-ci.yml as it takes effect while start-up the job # https://ccache.dev/manual/latest.html#_configuring_ccache # host mapping volume to share ccache between runner concurrent jobs CCACHE_DIR: "/cache/idf_ccache" CCACHE_MAXSIZE: "50G" FF_USE_NEW_BASH_EVAL_STRATEGY: "true" ################################################ # `before_script` and `after_script` Templates # ################################################ .common_before_scripts: &common-before_scripts | source tools/ci/utils.sh is_based_on_commits $REQUIRED_ANCESTOR_COMMITS if [[ -n "$IDF_DONT_USE_MIRRORS" ]]; then export IDF_MIRROR_PREFIX_MAP= fi if echo "$CI_MERGE_REQUEST_LABELS" | egrep "(^|,)include_nightly_run(,|$)"; then export INCLUDE_NIGHTLY_RUN="1" export NIGHTLY_RUN="1" fi # configure cmake related flags source tools/ci/configure_ci_environment.sh # add extra python packages export PYTHONPATH="$IDF_PATH/tools:$IDF_PATH/tools/esp_app_trace:$IDF_PATH/components/partition_table:$IDF_PATH/tools/ci/python_packages:$PYTHONPATH" .setup_tools_and_idf_python_venv: &setup_tools_and_idf_python_venv | # must use after setup_tools_except_target_test # otherwise the export.sh won't work properly # download constraint file for dev if [[ -n "$CI_PYTHON_CONSTRAINT_BRANCH" ]]; then wget -O /tmp/constraint.txt --header="Authorization:Bearer ${ESPCI_TOKEN}" ${GITLAB_HTTP_SERVER}/api/v4/projects/2581/repository/files/${CI_PYTHON_CONSTRAINT_FILE}/raw?ref=${CI_PYTHON_CONSTRAINT_BRANCH} mkdir -p ~/.espressif mv /tmp/constraint.txt ~/.espressif/${CI_PYTHON_CONSTRAINT_FILE} fi # Mirror if [[ -n "$IDF_DONT_USE_MIRRORS" ]]; then export IDF_MIRROR_PREFIX_MAP= fi # Optimize pip install if echo "${CI_RUNNER_TAGS}" | grep "shiny"; then export PIP_INDEX_URL="${PIP_INDEX_URL_SHINY}" fi if [[ "$(uname -m)" == "x86_64" ]] || [[ "$(uname -m)" == "aarch64" ]]; then export IDF_PIP_WHEELS_URL="" fi if [[ "${CI_JOB_STAGE}" != "target_test" ]]; then section_start "running_install_sh" "Running install.sh" if [[ "${CI_JOB_STAGE}" == "build_doc" ]]; then run_cmd bash install.sh --enable-ci --enable-docs elif [[ "${CI_JOB_STAGE}" == "build" ]]; then run_cmd bash install.sh --enable-ci else if ! echo "${CI_JOB_NAME}" | egrep ".*pytest.*"; then run_cmd bash install.sh --enable-ci else run_cmd bash install.sh --enable-ci --enable-pytest --enable-test-specific fi fi section_end "running_install_sh" else section_start "install_python_env" "Install Python environment" run_cmd python tools/idf_tools.py install-python-env --features ci,pytest,test-specific section_end "install_python_env" fi if [[ ! -z "$INSTALL_EXTRA_TOOLS" ]]; then section_start "installing_optional_tools" "Install optional tools ${INSTALL_EXTRA_TOOLS}" $IDF_PATH/tools/idf_tools.py --non-interactive install $INSTALL_EXTRA_TOOLS section_end "installing_optional_tools" fi # Install esp-clang if necessary (esp-clang is separately installed) if [[ "$IDF_TOOLCHAIN" == "clang" && -z "$CI_CLANG_DISTRO_URL" ]]; then $IDF_PATH/tools/idf_tools.py --non-interactive install esp-clang fi if [[ "${CI_JOB_STAGE}" == "target_test" ]]; then section_start "IDF_SKIP_TOOLS_CHECK" "Skip required tools check" export IDF_SKIP_TOOLS_CHECK=1 section_end "IDF_SKIP_TOOLS_CHECK" fi section_start "source_export" "Source export.sh" source ./export.sh section_end "source_export" # Custom clang toolchain if [[ "$IDF_TOOLCHAIN" == "clang" && ! -z "$CI_CLANG_DISTRO_URL" ]]; then echo "Using custom clang from ${CI_CLANG_DISTRO_URL}" wget $CI_CLANG_DISTRO_URL ARCH_NAME=$(basename $CI_CLANG_DISTRO_URL) tar -x -f $ARCH_NAME export PATH=$PWD/esp-clang/bin:$PATH fi # Custom OpenOCD if [[ "$CI_JOB_STAGE" == "target_test" ]]; then machine="$(uname -m)" if [[ "$machine" == "armv7l" ]] ; then OOCD_DISTRO_URL="$OOCD_DISTRO_URL_ARMHF" elif [[ "$machine" == "aarch64" ]] ; then OOCD_DISTRO_URL="$OOCD_DISTRO_URL_ARM64" fi if [[ ! -z "$OOCD_DISTRO_URL" ]]; then echo "Using custom OpenOCD from ${OOCD_DISTRO_URL}" wget $OOCD_DISTRO_URL ARCH_NAME=$(basename $OOCD_DISTRO_URL) tar -x -f $ARCH_NAME export OPENOCD_SCRIPTS=$PWD/openocd-esp32/share/openocd/scripts export PATH=$PWD/openocd-esp32/bin:$PATH fi fi if [[ -n "$CI_PYTHON_TOOL_REPO" ]]; then git clone --quiet --depth=1 -b ${CI_PYTHON_TOOL_BRANCH} https://gitlab-ci-token:${ESPCI_TOKEN}@${GITLAB_HTTPS_HOST}/espressif/${CI_PYTHON_TOOL_REPO}.git pip install ./${CI_PYTHON_TOOL_REPO} rm -rf ${CI_PYTHON_TOOL_REPO} fi info "setup tools and python venv done" .show_ccache_statistics: &show_ccache_statistics | # Show ccache statistics if enabled globally test "$CI_CCACHE_STATS" == 1 && test -n "$(which ccache)" && ccache --show-stats -vv || true .upload_failed_job_log_artifacts: &upload_failed_job_log_artifacts | if [ $CI_JOB_STATUS = "failed" ]; then python tools/ci/artifacts_handler.py upload --type logs fi .before_script:minimal: before_script: - *common-before_scripts .before_script:build:macos: before_script: # macos is running shell executor, which means it would use # the system installed /usr/local/bin/python3 by default. # Ensure pyenv and PYENV_VERSION installed - eval "$(pyenv init -)" - *common-before_scripts # remove idf-env.json, since it may contains enabled "features" - rm -f $IDF_TOOLS_PATH/idf-env.json # This adds tools (compilers) and the version-specific Python environment to PATH - *setup_tools_and_idf_python_venv - fetch_submodules variables: INSTALL_EXTRA_TOOLS: cmake ninja .after_script:build:macos:upload-failed-job-logs:ccache-show-stats: after_script: # macos is running shell executor, which means it would use # the system installed /usr/local/bin/python3 by default. # Ensure pyenv and PYENV_VERSION installed - eval "$(pyenv init -)" - *upload_failed_job_log_artifacts - *show_ccache_statistics .before_script:build: before_script: - *common-before_scripts - *setup_tools_and_idf_python_venv - add_gitlab_ssh_keys - fetch_submodules - export EXTRA_CFLAGS=${PEDANTIC_CFLAGS} - export EXTRA_CXXFLAGS=${PEDANTIC_CXXFLAGS} .after_script:build:ccache-show-stats: after_script: - *show_ccache_statistics .after_script:build:ccache-show-stats:upload-failed-job-logs: after_script: - *show_ccache_statistics - *upload_failed_job_log_artifacts ############################## # Git Strategy Job Templates # ############################## .git_init: &git_init | mkdir -p "${CI_PROJECT_DIR}" cd "${CI_PROJECT_DIR}" git init .git_fetch_from_mirror_url_if_exists: &git_fetch_from_mirror_url_if_exists | # check if set mirror if [ -n "${LOCAL_GITLAB_HTTPS_HOST:-}" ] && [ -n "${ESPCI_TOKEN:-}" ]; then MIRROR_REPO_URL="https://bot:${ESPCI_TOKEN}@${LOCAL_GITLAB_HTTPS_HOST}/${CI_PROJECT_PATH}" elif [ -n "${LOCAL_GIT_MIRROR:-}" ]; then MIRROR_REPO_URL="${LOCAL_GIT_MIRROR}/${CI_PROJECT_PATH}" fi # fetch from mirror first if set if [ -n "${MIRROR_REPO_URL:-}" ]; then if git remote -v | grep origin; then git remote set-url origin "${MIRROR_REPO_URL}" else git remote add origin "${MIRROR_REPO_URL}" fi # mirror url may fail with authentication issue git fetch origin --no-recurse-submodules || true fi # set remote url to CI_REPOSITORY_URL if git remote -v | grep origin; then git remote set-url origin "${CI_REPOSITORY_URL}" else git remote add origin "${CI_REPOSITORY_URL}" fi .git_checkout_ci_commit_sha: &git_checkout_ci_commit_sha | git checkout $CI_COMMIT_SHA git clean ${GIT_CLEAN_FLAGS} # git diff requires two commits, with different CI env var # # By default, we use git strategy "clone" with depth 1 to speed up the clone process. # But for jobs requires running `git diff`, we need to fetch more commits to get the correct diffs. # # Since there's no way to get the correct git_depth before the job starts, # we can't set `GIT_DEPTH` in the job definition. # # Set git strategy to "none" and fetch manually instead. .before_script:fetch:git_diff: variables: GIT_STRATEGY: none before_script: - *git_init - *git_fetch_from_mirror_url_if_exists - | # Store the diff output in a temporary file TEMP_FILE=$(mktemp) # merged results pipelines, by default if [[ -n $CI_MERGE_REQUEST_SOURCE_BRANCH_SHA ]]; then git fetch origin $CI_MERGE_REQUEST_DIFF_BASE_SHA --depth=1 ${GIT_FETCH_EXTRA_FLAGS} git fetch origin $CI_MERGE_REQUEST_SOURCE_BRANCH_SHA --depth=1 ${GIT_FETCH_EXTRA_FLAGS} git diff --name-only $CI_MERGE_REQUEST_DIFF_BASE_SHA $CI_MERGE_REQUEST_SOURCE_BRANCH_SHA > "$TEMP_FILE" GIT_DIFF_OUTPUT=$(cat "$TEMP_FILE") git fetch origin $CI_COMMIT_SHA --depth=1 ${GIT_FETCH_EXTRA_FLAGS} # merge request pipelines, when the mr got conflicts elif [[ -n $CI_MERGE_REQUEST_DIFF_BASE_SHA ]]; then git fetch origin $CI_MERGE_REQUEST_DIFF_BASE_SHA --depth=1 ${GIT_FETCH_EXTRA_FLAGS} git fetch origin $CI_COMMIT_SHA --depth=1 ${GIT_FETCH_EXTRA_FLAGS} git diff --name-only $CI_MERGE_REQUEST_DIFF_BASE_SHA $CI_COMMIT_SHA > "$TEMP_FILE" GIT_DIFF_OUTPUT=$(cat "$TEMP_FILE") # other pipelines, like the protected branches pipelines elif [[ "$CI_COMMIT_BEFORE_SHA" != "0000000000000000000000000000000000000000" ]]; then git fetch origin $CI_COMMIT_BEFORE_SHA --depth=1 ${GIT_FETCH_EXTRA_FLAGS} git fetch origin $CI_COMMIT_SHA --depth=1 ${GIT_FETCH_EXTRA_FLAGS} git diff --name-only $CI_COMMIT_BEFORE_SHA $CI_COMMIT_SHA > "$TEMP_FILE" GIT_DIFF_OUTPUT=$(cat "$TEMP_FILE") else # pipeline source could be web, scheduler, etc. git fetch origin $CI_COMMIT_SHA --depth=2 ${GIT_FETCH_EXTRA_FLAGS} git diff --name-only $CI_COMMIT_SHA~1 $CI_COMMIT_SHA > "$TEMP_FILE" GIT_DIFF_OUTPUT=$(cat "$TEMP_FILE") fi - *git_checkout_ci_commit_sha - *common-before_scripts - *setup_tools_and_idf_python_venv - add_gitlab_ssh_keys # target test runners may locate in different places # for runners set git mirror, we fetch from the mirror first, then fetch the HEAD commit .before_script:fetch:target_test: variables: GIT_STRATEGY: none before_script: - *git_init - *git_fetch_from_mirror_url_if_exists - git fetch origin "${CI_COMMIT_SHA}" --depth=1 ${GIT_FETCH_EXTRA_FLAGS} - *git_checkout_ci_commit_sha - *common-before_scripts - *setup_tools_and_idf_python_venv - add_gitlab_ssh_keys # no submodules ############# # `default` # ############# default: cache: # pull only for most of the use cases since it's cache dir. # Only set "push" policy for "upload_cache" stage jobs - key: pip-cache-${LATEST_GIT_TAG} fallback_keys: - pip-cache paths: - .cache/pip policy: pull - key: submodule-cache-${LATEST_GIT_TAG} fallback_keys: - submodule-cache paths: - .cache/submodule_archives policy: pull before_script: - *common-before_scripts - *setup_tools_and_idf_python_venv - add_gitlab_ssh_keys - fetch_submodules # gitlab bug, setting them here doesn't work # - expire_in: https://gitlab.com/gitlab-org/gitlab/-/issues/404563 # - when: https://gitlab.com/gitlab-org/gitlab/-/issues/440672 # artifacts: # expire_in: 1 week # when: always retry: max: 2 when: # In case of a runner failure we could hop to another one, or a network error could go away. - runner_system_failure # Job execution timeout may be caused by a network issue. - job_execution_timeout