# SPDX-FileCopyrightText: 2021-2024 Espressif Systems (Shanghai) CO LTD # SPDX-License-Identifier: Apache-2.0 # pylint: disable=W0621 # redefined-outer-name # # IDF is using [pytest](https://github.com/pytest-dev/pytest) and # [pytest-embedded plugin](https://github.com/espressif/pytest-embedded) as its test framework. # # if you found any bug or have any question, # please report to https://github.com/espressif/pytest-embedded/issues # or discuss at https://github.com/espressif/pytest-embedded/discussions import os import sys if os.path.join(os.path.dirname(__file__), 'tools', 'ci') not in sys.path: sys.path.append(os.path.join(os.path.dirname(__file__), 'tools', 'ci')) if os.path.join(os.path.dirname(__file__), 'tools', 'ci', 'python_packages') not in sys.path: sys.path.append(os.path.join(os.path.dirname(__file__), 'tools', 'ci', 'python_packages')) import glob import io import logging import os import re import typing as t import zipfile from copy import deepcopy from datetime import datetime import common_test_methods # noqa: F401 import gitlab_api import pytest import requests import yaml from _pytest.config import Config from _pytest.fixtures import FixtureRequest from artifacts_handler import ArtifactType from dynamic_pipelines.constants import TEST_RELATED_APPS_DOWNLOAD_URLS_FILENAME from idf_ci.app import import_apps_from_txt from idf_ci.uploader import AppDownloader, AppUploader from idf_ci_utils import IDF_PATH, idf_relpath from idf_pytest.constants import DEFAULT_SDKCONFIG, ENV_MARKERS, SPECIAL_MARKERS, TARGET_MARKERS, PytestCase from idf_pytest.plugin import IDF_PYTEST_EMBEDDED_KEY, ITEM_PYTEST_CASE_KEY, IdfPytestEmbedded from idf_pytest.utils import format_case_id from pytest_embedded.plugin import multi_dut_argument, multi_dut_fixture from pytest_embedded_idf.dut import IdfDut from pytest_embedded_idf.unity_tester import CaseTester ############ # Fixtures # ############ @pytest.fixture(scope='session') def idf_path() -> str: return os.path.dirname(__file__) @pytest.fixture(scope='session', autouse=True) def session_tempdir() -> str: _tmpdir = os.path.join( os.path.dirname(__file__), 'pytest_embedded_log', datetime.now().strftime('%Y-%m-%d_%H-%M-%S'), ) os.makedirs(_tmpdir, exist_ok=True) return _tmpdir @pytest.fixture def case_tester(unity_tester: CaseTester) -> CaseTester: return unity_tester @pytest.fixture @multi_dut_argument def config(request: FixtureRequest) -> str: return getattr(request, 'param', None) or DEFAULT_SDKCONFIG # type: ignore @pytest.fixture @multi_dut_fixture def target(request: FixtureRequest, dut_total: int, dut_index: int) -> str: plugin = request.config.stash[IDF_PYTEST_EMBEDDED_KEY] if dut_total == 1: return plugin.target[0] # type: ignore return plugin.target[dut_index] # type: ignore @pytest.fixture def test_func_name(request: FixtureRequest) -> str: return request.node.function.__name__ # type: ignore @pytest.fixture def test_case_name(request: FixtureRequest, target: str, config: str) -> str: is_qemu = request._pyfuncitem.get_closest_marker('qemu') is not None if hasattr(request._pyfuncitem, 'callspec'): params = deepcopy(request._pyfuncitem.callspec.params) # type: ignore else: params = {} filtered_params = {} for k, v in params.items(): if k not in request.session._fixturemanager._arg2fixturedefs: # type: ignore filtered_params[k] = v # not fixture ones return format_case_id(target, config, request.node.originalname, is_qemu=is_qemu, params=filtered_params) # type: ignore @pytest.fixture(scope='session') def pipeline_id(request: FixtureRequest) -> t.Optional[str]: return request.config.getoption('pipeline_id', None) or os.getenv('PARENT_PIPELINE_ID', None) # type: ignore class BuildReportDownloader(AppDownloader): def __init__(self, presigned_url_yaml: str) -> None: self.app_presigned_urls_dict: t.Dict[str, t.Dict[str, str]] = yaml.safe_load(presigned_url_yaml) def _download_app(self, app_build_path: str, artifact_type: ArtifactType) -> None: url = self.app_presigned_urls_dict[app_build_path][artifact_type.value] logging.info('Downloading app from %s', url) with io.BytesIO() as f: for chunk in requests.get(url).iter_content(chunk_size=1024 * 1024): if chunk: f.write(chunk) f.seek(0) with zipfile.ZipFile(f) as zip_ref: zip_ref.extractall(IDF_PATH) def download_app(self, app_build_path: str, artifact_type: t.Optional[ArtifactType] = None) -> None: if app_build_path not in self.app_presigned_urls_dict: raise ValueError( f'No presigned url found for {app_build_path}. ' f'Usually this should not happen, please re-trigger a pipeline.' f'If this happens again, please report this bug to the CI channel.' ) super().download_app(app_build_path, artifact_type) @pytest.fixture(scope='session') def app_downloader(pipeline_id: t.Optional[str]) -> t.Optional[AppDownloader]: if not pipeline_id: return None if ( 'IDF_S3_BUCKET' in os.environ and 'IDF_S3_ACCESS_KEY' in os.environ and 'IDF_S3_SECRET_KEY' in os.environ and 'IDF_S3_SERVER' in os.environ and 'IDF_S3_BUCKET' in os.environ ): return AppUploader(pipeline_id) logging.info('Downloading build report from the build pipeline %s', pipeline_id) test_app_presigned_urls_file = None gl = gitlab_api.Gitlab(os.getenv('CI_PROJECT_ID', 'espressif/esp-idf')) for child_pipeline in gl.project.pipelines.get(pipeline_id, lazy=True).bridges.list(iterator=True): if child_pipeline.name == 'build_child_pipeline': for job in gl.project.pipelines.get(child_pipeline.downstream_pipeline['id'], lazy=True).jobs.list( iterator=True ): if job.name == 'generate_pytest_build_report': test_app_presigned_urls_file = gl.download_artifact( job.id, [TEST_RELATED_APPS_DOWNLOAD_URLS_FILENAME] )[0] break if test_app_presigned_urls_file: return BuildReportDownloader(test_app_presigned_urls_file) return None @pytest.fixture @multi_dut_fixture def build_dir( request: FixtureRequest, app_path: str, target: t.Optional[str], config: t.Optional[str], app_downloader: t.Optional[AppDownloader], ) -> str: """ Check local build dir with the following priority: 1. build__ 2. build_ 3. build_ 4. build Returns: valid build directory """ # download from minio on CI case: PytestCase = request._pyfuncitem.stash[ITEM_PYTEST_CASE_KEY] if app_downloader: # somehow hardcoded... app_build_path = os.path.join(idf_relpath(app_path), f'build_{target}_{config}') if case.requires_elf_or_map: app_downloader.download_app(app_build_path) else: app_downloader.download_app(app_build_path, ArtifactType.BUILD_DIR_WITHOUT_MAP_AND_ELF_FILES) check_dirs = [f'build_{target}_{config}'] else: check_dirs = [] if target is not None and config is not None: check_dirs.append(f'build_{target}_{config}') if target is not None: check_dirs.append(f'build_{target}') if config is not None: check_dirs.append(f'build_{config}') check_dirs.append('build') for check_dir in check_dirs: binary_path = os.path.join(app_path, check_dir) if os.path.isdir(binary_path): logging.info(f'found valid binary path: {binary_path}') return check_dir logging.warning('checking binary path: %s... missing... try another place', binary_path) raise ValueError( f'no build dir valid. Please build the binary via "idf.py -B {check_dirs[0]} build" and run pytest again' ) @pytest.fixture(autouse=True) @multi_dut_fixture def junit_properties(test_case_name: str, record_xml_attribute: t.Callable[[str, object], None]) -> None: """ This fixture is autoused and will modify the junit report test case name to .. """ record_xml_attribute('name', test_case_name) @pytest.fixture(autouse=True) @multi_dut_fixture def ci_job_url(record_xml_attribute: t.Callable[[str, object], None]) -> None: if ci_job_url := os.getenv('CI_JOB_URL'): record_xml_attribute('ci_job_url', ci_job_url) @pytest.fixture(autouse=True) def set_test_case_name(request: FixtureRequest, test_case_name: str) -> None: request.node.funcargs['test_case_name'] = test_case_name ###################### # Log Util Functions # ###################### @pytest.fixture def log_performance(record_property: t.Callable[[str, object], None]) -> t.Callable[[str, str], None]: """ log performance item with pre-defined format to the console and record it under the ``properties`` tag in the junit report if available. """ def real_func(item: str, value: str) -> None: """ :param item: performance item name :param value: performance value """ logging.info('[Performance][%s]: %s', item, value) record_property(item, value) return real_func @pytest.fixture def check_performance(idf_path: str) -> t.Callable[[str, float, str], None]: """ check if the given performance item meets the passing standard or not """ def real_func(item: str, value: float, target: str) -> None: """ :param item: performance item name :param value: performance item value :param target: target chip :raise: AssertionError: if check fails """ def _find_perf_item(operator: str, path: str) -> float: with open(path) as f: data = f.read() match = re.search(fr'#define\s+IDF_PERFORMANCE_{operator}_{item.upper()}\s+([\d.]+)', data) return float(match.group(1)) # type: ignore def _check_perf(operator: str, standard_value: float) -> None: if operator == 'MAX': ret = value <= standard_value else: ret = value >= standard_value if not ret: raise AssertionError( f"[Performance] {item} value is {value}, doesn't meet pass standard {standard_value}" ) path_prefix = os.path.join(idf_path, 'components', 'idf_test', 'include') performance_files = ( os.path.join(path_prefix, target, 'idf_performance_target.h'), os.path.join(path_prefix, 'idf_performance.h'), ) found_item = False for op in ['MIN', 'MAX']: for performance_file in performance_files: try: standard = _find_perf_item(op, performance_file) except (OSError, AttributeError): # performance file doesn't exist or match is not found in it continue _check_perf(op, standard) found_item = True break if not found_item: raise AssertionError(f'Failed to get performance standard for {item}') return real_func @pytest.fixture def log_minimum_free_heap_size(dut: IdfDut, config: str) -> t.Callable[..., None]: def real_func() -> None: res = dut.expect(r'Minimum free heap size: (\d+) bytes') logging.info( '\n------ heap size info ------\n' '[app_name] {}\n' '[config_name] {}\n' '[target] {}\n' '[minimum_free_heap_size] {} Bytes\n' '------ heap size end ------'.format( os.path.basename(dut.app.app_path), config, dut.target, res.group(1).decode('utf8'), ) ) return real_func @pytest.fixture(scope='session') def dev_password(request: FixtureRequest) -> str: return request.config.getoption('dev_passwd') or '' @pytest.fixture(scope='session') def dev_user(request: FixtureRequest) -> str: return request.config.getoption('dev_user') or '' ################## # Hook functions # ################## def pytest_addoption(parser: pytest.Parser) -> None: idf_group = parser.getgroup('idf') idf_group.addoption( '--sdkconfig', help='sdkconfig postfix, like sdkconfig.ci.. (Default: None, which would build all found apps)', ) idf_group.addoption( '--dev-user', help='user name associated with some specific device/service used during the test execution', ) idf_group.addoption( '--dev-passwd', help='password associated with some specific device/service used during the test execution', ) idf_group.addoption( '--app-info-filepattern', help='glob pattern to specify the files that include built app info generated by ' '`idf-build-apps --collect-app-info ...`. will not raise ValueError when binary ' 'paths not exist in local file system if not listed recorded in the app info.', ) idf_group.addoption( '--pipeline-id', help='main pipeline id, not the child pipeline id. Specify this option to download the artifacts ' 'from the minio server for debugging purpose.', ) def pytest_configure(config: Config) -> None: # cli option "--target" target = [_t.strip().lower() for _t in (config.getoption('target', '') or '').split(',') if _t.strip()] # add markers based on idf_pytest/constants.py for name, description in { **TARGET_MARKERS, **ENV_MARKERS, **SPECIAL_MARKERS, }.items(): config.addinivalue_line('markers', f'{name}: {description}') help_commands = ['--help', '--fixtures', '--markers', '--version'] for cmd in help_commands: if cmd in config.invocation_params.args: target = ['unneeded'] break markexpr = config.getoption('markexpr') or '' # check marker expr set via "pytest -m" if not target and markexpr: # we use `-m "esp32 and generic"` in our CI to filter the test cases # this doesn't cover all use cases, but fit what we do in CI. for marker in markexpr.split('and'): marker = marker.strip() if marker in TARGET_MARKERS: target.append(marker) # "--target" must be set if not target: raise SystemExit( """Pass `--target TARGET[,TARGET...]` to specify all targets the test cases are using. - for single DUT, we run with `pytest --target esp32` - for multi DUT, we run with `pytest --target esp32,esp32,esp32s2` to indicate all DUTs """ ) apps = None app_info_filepattern = config.getoption('app_info_filepattern') if app_info_filepattern: apps = [] for f in glob.glob(os.path.join(IDF_PATH, app_info_filepattern)): apps.extend(import_apps_from_txt(f)) if '--collect-only' not in config.invocation_params.args: config.stash[IDF_PYTEST_EMBEDDED_KEY] = IdfPytestEmbedded( config_name=config.getoption('sdkconfig'), target=target, apps=apps, ) config.pluginmanager.register(config.stash[IDF_PYTEST_EMBEDDED_KEY]) def pytest_unconfigure(config: Config) -> None: _pytest_embedded = config.stash.get(IDF_PYTEST_EMBEDDED_KEY, None) if _pytest_embedded: del config.stash[IDF_PYTEST_EMBEDDED_KEY] config.pluginmanager.unregister(_pytest_embedded) dut_artifacts_url = [] @pytest.hookimpl(hookwrapper=True) def pytest_runtest_makereport(item, call): # type: ignore outcome = yield report = outcome.get_result() report.sections = [] if report.failed: _dut = item.funcargs.get('dut') if not _dut: return job_id = os.getenv('CI_JOB_ID', 0) url = os.getenv('CI_PAGES_URL', '').replace('esp-idf', '-/esp-idf') template = f'{url}/-/jobs/{job_id}/artifacts/pytest_embedded_log/{{}}' logs_files = [] def get_path(x: str) -> str: return x.split('pytest_embedded_log/', 1)[1] if isinstance(_dut, list): logs_files.extend([template.format(get_path(d.logfile)) for d in _dut]) dut_artifacts_url.append('{}:'.format(_dut[0].test_case_name)) else: logs_files.append(template.format(get_path(_dut.logfile))) dut_artifacts_url.append('{}:'.format(_dut.test_case_name)) for file in logs_files: dut_artifacts_url.append(' - {}'.format(file)) def pytest_terminal_summary(terminalreporter, exitstatus, config): # type: ignore if dut_artifacts_url: terminalreporter.ensure_newline() terminalreporter.section('Failed Test Artifacts URL', sep='-', red=True, bold=True) terminalreporter.line('\n'.join(dut_artifacts_url))