| #!/usr/bin/env python3 |
| |
| # Copyright (c) 2018,2020 Intel Corporation |
| # SPDX-License-Identifier: Apache-2.0 |
| |
| import collections |
| import sys |
| import subprocess |
| import re |
| import os |
| from email.utils import parseaddr |
| import logging |
| import argparse |
| from junitparser import TestCase, TestSuite, JUnitXml, Skipped, Error, Failure, Attr |
| import tempfile |
| import traceback |
| import magic |
| import shlex |
| from pathlib import Path |
| |
| # '*' makes it italic |
| EDIT_TIP = "\n\n*Tip: The bot edits this comment instead of posting a new " \ |
| "one, so you can check the comment's history to see earlier " \ |
| "messages.*" |
| |
| logger = None |
| |
| # This ends up as None when we're not running in a Zephyr tree |
| ZEPHYR_BASE = os.environ.get('ZEPHYR_BASE') |
| |
| |
| def git(*args, cwd=None): |
| # Helper for running a Git command. Returns the rstrip()ed stdout output. |
| # Called like git("diff"). Exits with SystemError (raised by sys.exit()) on |
| # errors. 'cwd' is the working directory to use (default: current |
| # directory). |
| |
| git_cmd = ("git",) + args |
| try: |
| git_process = subprocess.Popen( |
| git_cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, cwd=cwd) |
| except OSError as e: |
| err(f"failed to run '{cmd2str(git_cmd)}': {e}") |
| |
| stdout, stderr = git_process.communicate() |
| stdout = stdout.decode("utf-8") |
| stderr = stderr.decode("utf-8") |
| if git_process.returncode or stderr: |
| err(f"""\ |
| '{cmd2str(git_cmd)}' exited with status {git_process.returncode} and/or wrote |
| to stderr. |
| |
| ==stdout== |
| {stdout} |
| ==stderr== |
| {stderr}""") |
| |
| return stdout.rstrip() |
| |
| |
| def get_shas(refspec): |
| """ |
| Returns the list of Git SHAs for 'refspec'. |
| |
| :param refspec: |
| :return: |
| """ |
| return git('rev-list', |
| '--max-count={}'.format(-1 if "." in refspec else 1), |
| refspec).split() |
| |
| |
| class MyCase(TestCase): |
| """ |
| Custom junitparser.TestCase for our tests that adds some extra <testcase> |
| XML attributes. These will be preserved when tests are saved and loaded. |
| """ |
| classname = Attr() |
| # Remembers informational messages. These can appear on successful tests |
| # too, where TestCase.result isn't set. |
| info_msg = Attr() |
| |
| |
| class ComplianceTest: |
| """ |
| Base class for tests. Inheriting classes should have a run() method and set |
| these class variables: |
| |
| name: |
| Test name |
| |
| doc: |
| Link to documentation related to what's being tested |
| |
| path_hint: |
| The path the test runs itself in. This is just informative and used in |
| the message that gets printed when running the test. |
| |
| The magic string "<git-top>" refers to the top-level repository |
| directory. This avoids running 'git' to find the top-level directory |
| before main() runs (class variable assignments run when the 'class ...' |
| statement runs). That avoids swallowing errors, because main() reports |
| them to GitHub. |
| """ |
| def __init__(self): |
| self.case = MyCase(self.name) |
| self.case.classname = "Guidelines" |
| |
| def error(self, msg): |
| """ |
| Signals a problem with running the test, with message 'msg'. |
| |
| Raises an exception internally, so you do not need to put a 'return' |
| after error(). |
| |
| Any failures generated prior to the error() are included automatically |
| in the message. Usually, any failures would indicate problems with the |
| test code. |
| """ |
| if self.case.result: |
| msg += "\n\nFailures before error: " + self.case.result._elem.text |
| |
| self.case.result = Error(msg, "error") |
| |
| raise EndTest |
| |
| def skip(self, msg): |
| """ |
| Signals that the test should be skipped, with message 'msg'. |
| |
| Raises an exception internally, so you do not need to put a 'return' |
| after skip(). |
| |
| Any failures generated prior to the skip() are included automatically |
| in the message. Usually, any failures would indicate problems with the |
| test code. |
| """ |
| if self.case.result: |
| msg += "\n\nFailures before skip: " + self.case.result._elem.text |
| |
| self.case.result = Skipped(msg, "skipped") |
| |
| raise EndTest |
| |
| def add_failure(self, msg): |
| """ |
| Signals that the test failed, with message 'msg'. Can be called many |
| times within the same test to report multiple failures. |
| """ |
| if not self.case.result: |
| # First reported failure |
| self.case.result = Failure(self.name + " issues", "failure") |
| self.case.result._elem.text = msg.rstrip() |
| else: |
| # If there are multiple Failures, concatenate their messages |
| self.case.result._elem.text += "\n\n" + msg.rstrip() |
| |
| def add_info(self, msg): |
| """ |
| Adds an informational message without failing the test. The message is |
| shown on GitHub, and is shown regardless of whether the test passes or |
| fails. If the test fails, then both the informational message and the |
| failure message are shown. |
| |
| Can be called many times within the same test to add multiple messages. |
| """ |
| def escape(s): |
| # Hack to preserve e.g. newlines and tabs in the attribute when |
| # tests are saved to .xml and reloaded. junitparser doesn't seem to |
| # handle it correctly, though it does escape stuff like quotes. |
| # unicode-escape replaces newlines with \n (two characters), etc. |
| return s.encode("unicode-escape").decode("utf-8") |
| |
| if not self.case.info_msg: |
| self.case.info_msg = escape(msg) |
| else: |
| self.case.info_msg += r"\n\n" + escape(msg) |
| |
| |
| class EndTest(Exception): |
| """ |
| Raised by ComplianceTest.error()/skip() to end the test. |
| |
| Tests can raise EndTest themselves to immediately end the test, e.g. from |
| within a nested function call. |
| """ |
| |
| |
| class CheckPatch(ComplianceTest): |
| """ |
| Runs checkpatch and reports found issues |
| |
| """ |
| name = "checkpatch" |
| doc = "See https://docs.zephyrproject.org/latest/contribute/#coding-style for more details." |
| path_hint = "<git-top>" |
| |
| def run(self): |
| # Default to Zephyr's checkpatch if ZEPHYR_BASE is set |
| checkpatch = os.path.join(ZEPHYR_BASE or GIT_TOP, 'scripts', |
| 'checkpatch.pl') |
| if not os.path.exists(checkpatch): |
| self.skip(checkpatch + " not found") |
| |
| # git diff's output doesn't depend on the current (sub)directory |
| diff = subprocess.Popen(('git', 'diff', COMMIT_RANGE), |
| stdout=subprocess.PIPE) |
| try: |
| subprocess.check_output((checkpatch, '--mailback', '--no-tree', '-'), |
| stdin=diff.stdout, |
| stderr=subprocess.STDOUT, |
| shell=True, cwd=GIT_TOP) |
| |
| except subprocess.CalledProcessError as ex: |
| output = ex.output.decode("utf-8") |
| if re.search("[1-9][0-9]* errors,", output): |
| self.add_failure(output) |
| else: |
| # No errors found, but warnings. Show them. |
| self.add_info(output) |
| |
| |
| class KconfigCheck(ComplianceTest): |
| """ |
| Checks is we are introducing any new warnings/errors with Kconfig, |
| for example using undefiend Kconfig variables. |
| """ |
| name = "Kconfig" |
| doc = "See https://docs.zephyrproject.org/latest/guides/kconfig/index.html for more details." |
| path_hint = ZEPHYR_BASE |
| |
| def run(self): |
| kconf = self.parse_kconfig() |
| |
| self.check_top_menu_not_too_long(kconf) |
| self.check_no_pointless_menuconfigs(kconf) |
| self.check_no_undef_within_kconfig(kconf) |
| self.check_no_undef_outside_kconfig(kconf) |
| |
| def get_modules(self, modules_file): |
| """ |
| Get a list of modules and put them in a file that is parsed by |
| Kconfig |
| |
| This is needed to complete Kconfig sanity tests. |
| |
| """ |
| # Invoke the script directly using the Python executable since this is |
| # not a module nor a pip-installed Python utility |
| zephyr_module_path = os.path.join(ZEPHYR_BASE, "scripts", |
| "zephyr_module.py") |
| cmd = [sys.executable, zephyr_module_path, |
| '--kconfig-out', modules_file] |
| try: |
| _ = subprocess.check_output(cmd, stderr=subprocess.STDOUT) |
| except subprocess.CalledProcessError as ex: |
| self.error(ex.output) |
| |
| def write_kconfig_soc(self): |
| """ |
| Write KConfig soc files to be sourced during Kconfig parsing |
| |
| """ |
| |
| soc_defconfig_file = os.path.join(tempfile.gettempdir(), "Kconfig.soc.defconfig") |
| soc_file = os.path.join(tempfile.gettempdir(), "Kconfig.soc") |
| soc_arch_file = os.path.join(tempfile.gettempdir(), "Kconfig.soc.arch") |
| try: |
| with open(soc_defconfig_file, 'w', encoding="utf-8") as fp: |
| fp.write(f'osource "{ZEPHYR_BASE}/soc/$(ARCH)/*/Kconfig.defconfig"\n') |
| |
| with open(soc_file, 'w', encoding="utf-8") as fp: |
| fp.write(f'osource "{ZEPHYR_BASE}/soc/$(ARCH)/*/Kconfig.soc"\n') |
| |
| with open(soc_arch_file, 'w', encoding="utf-8") as fp: |
| fp.write(f'osource "{ZEPHYR_BASE}/soc/$(ARCH)/Kconfig"\n\ |
| osource "{ZEPHYR_BASE}/soc/$(ARCH)/*/Kconfig"\n') |
| except IOError as ex: |
| self.error(ex.output) |
| |
| def parse_kconfig(self): |
| """ |
| Returns a kconfiglib.Kconfig object for the Kconfig files. We reuse |
| this object for all tests to avoid having to reparse for each test. |
| """ |
| if not ZEPHYR_BASE: |
| self.skip("Not a Zephyr tree (ZEPHYR_BASE unset)") |
| |
| # Put the Kconfiglib path first to make sure no local Kconfiglib version is |
| # used |
| kconfig_path = os.path.join(ZEPHYR_BASE, "scripts", "kconfig") |
| if not os.path.exists(kconfig_path): |
| self.error(kconfig_path + " not found") |
| |
| sys.path.insert(0, kconfig_path) |
| # Import globally so that e.g. kconfiglib.Symbol can be referenced in |
| # tests |
| global kconfiglib |
| import kconfiglib |
| |
| # Look up Kconfig files relative to ZEPHYR_BASE |
| os.environ["srctree"] = ZEPHYR_BASE |
| |
| # Parse the entire Kconfig tree, to make sure we see all symbols |
| os.environ["SOC_DIR"] = "soc/" |
| os.environ["ARCH_DIR"] = "arch/" |
| os.environ["BOARD_DIR"] = "boards/*/*" |
| os.environ["ARCH"] = "*" |
| os.environ["KCONFIG_BINARY_DIR"] = tempfile.gettempdir() |
| os.environ['DEVICETREE_CONF'] = "dummy" |
| |
| # Older name for DEVICETREE_CONF, for compatibility with older Zephyr |
| # versions that don't have the renaming |
| os.environ["GENERATED_DTS_BOARD_CONF"] = "dummy" |
| |
| # For multi repo support |
| self.get_modules(os.path.join(tempfile.gettempdir(), "Kconfig.modules")) |
| |
| # For list of SOC_ROOT support |
| self.write_kconfig_soc() |
| |
| # Tells Kconfiglib to generate warnings for all references to undefined |
| # symbols within Kconfig files |
| os.environ["KCONFIG_WARN_UNDEF"] = "y" |
| |
| try: |
| # Note this will both print warnings to stderr _and_ return |
| # them: so some warnings might get printed |
| # twice. "warn_to_stderr=False" could unfortunately cause |
| # some (other) warnings to never be printed. |
| return kconfiglib.Kconfig() |
| except kconfiglib.KconfigError as e: |
| self.add_failure(str(e)) |
| raise EndTest |
| |
| def check_top_menu_not_too_long(self, kconf): |
| """ |
| Checks that there aren't too many items in the top-level menu (which |
| might be a sign that stuff accidentally got added there) |
| """ |
| max_top_items = 50 |
| |
| n_top_items = 0 |
| node = kconf.top_node.list |
| while node: |
| # Only count items with prompts. Other items will never be |
| # shown in the menuconfig (outside show-all mode). |
| if node.prompt: |
| n_top_items += 1 |
| node = node.next |
| |
| if n_top_items > max_top_items: |
| self.add_failure(""" |
| Expected no more than {} potentially visible items (items with prompts) in the |
| top-level Kconfig menu, found {} items. If you're deliberately adding new |
| entries, then bump the 'max_top_items' variable in {}. |
| """.format(max_top_items, n_top_items, __file__)) |
| |
| def check_no_pointless_menuconfigs(self, kconf): |
| # Checks that there are no pointless 'menuconfig' symbols without |
| # children in the Kconfig files |
| |
| bad_mconfs = [] |
| for node in kconf.node_iter(): |
| # 'kconfiglib' is global |
| # pylint: disable=undefined-variable |
| |
| # Avoid flagging empty regular menus and choices, in case people do |
| # something with 'osource' (could happen for 'menuconfig' symbols |
| # too, though it's less likely) |
| if node.is_menuconfig and not node.list and \ |
| isinstance(node.item, kconfiglib.Symbol): |
| |
| bad_mconfs.append(node) |
| |
| if bad_mconfs: |
| self.add_failure("""\ |
| Found pointless 'menuconfig' symbols without children. Use regular 'config' |
| symbols instead. See |
| https://docs.zephyrproject.org/latest/guides/kconfig/tips.html#menuconfig-symbols. |
| |
| """ + "\n".join(f"{node.item.name:35} {node.filename}:{node.linenr}" |
| for node in bad_mconfs)) |
| |
| def check_no_undef_within_kconfig(self, kconf): |
| """ |
| Checks that there are no references to undefined Kconfig symbols within |
| the Kconfig files |
| """ |
| undef_ref_warnings = "\n\n\n".join(warning for warning in kconf.warnings |
| if "undefined symbol" in warning) |
| |
| if undef_ref_warnings: |
| self.add_failure("Undefined Kconfig symbols:\n\n" |
| + undef_ref_warnings) |
| |
| def check_no_undef_outside_kconfig(self, kconf): |
| """ |
| Checks that there are no references to undefined Kconfig symbols |
| outside Kconfig files (any CONFIG_FOO where no FOO symbol exists) |
| """ |
| # Grep for symbol references. |
| # |
| # Example output line for a reference to CONFIG_FOO at line 17 of |
| # foo/bar.c: |
| # |
| # foo/bar.c<null>17<null>#ifdef CONFIG_FOO |
| # |
| # 'git grep --only-matching' would get rid of the surrounding context |
| # ('#ifdef '), but it was added fairly recently (second half of 2018), |
| # so we extract the references from each line ourselves instead. |
| # |
| # The regex uses word boundaries (\b) to isolate the reference, and |
| # negative lookahead to automatically whitelist the following: |
| # |
| # - ##, for token pasting (CONFIG_FOO_##X) |
| # |
| # - $, e.g. for CMake variable expansion (CONFIG_FOO_${VAR}) |
| # |
| # - @, e.g. for CMakes's configure_file() (CONFIG_FOO_@VAR@) |
| # |
| # - {, e.g. for Python scripts ("CONFIG_FOO_{}_BAR".format(...)") |
| # |
| # - *, meant for comments like '#endif /* CONFIG_FOO_* */ |
| |
| defined_syms = get_defined_syms(kconf) |
| |
| # Maps each undefined symbol to a list <filename>:<linenr> strings |
| undef_to_locs = collections.defaultdict(list) |
| |
| # Warning: Needs to work with both --perl-regexp and the 're' module |
| regex = r"\bCONFIG_[A-Z0-9_]+\b(?!\s*##|[$@{*])" |
| |
| # Skip doc/releases, which often references removed symbols |
| grep_stdout = git("grep", "--line-number", "-I", "--null", |
| "--perl-regexp", regex, "--", ":!/doc/releases", |
| cwd=ZEPHYR_BASE) |
| |
| # splitlines() supports various line terminators |
| for grep_line in grep_stdout.splitlines(): |
| path, lineno, line = grep_line.split("\0") |
| |
| # Extract symbol references (might be more than one) within the |
| # line |
| for sym_name in re.findall(regex, line): |
| sym_name = sym_name[7:] # Strip CONFIG_ |
| if sym_name not in defined_syms and \ |
| sym_name not in UNDEF_KCONFIG_WHITELIST: |
| |
| undef_to_locs[sym_name].append("{}:{}".format(path, lineno)) |
| |
| if not undef_to_locs: |
| return |
| |
| # String that describes all referenced but undefined Kconfig symbols, |
| # in alphabetical order, along with the locations where they're |
| # referenced. Example: |
| # |
| # CONFIG_ALSO_MISSING arch/xtensa/core/fatal.c:273 |
| # CONFIG_MISSING arch/xtensa/core/fatal.c:264, subsys/fb/cfb.c:20 |
| undef_desc = "\n".join( |
| "CONFIG_{:35} {}".format(sym_name, ", ".join(locs)) |
| for sym_name, locs in sorted(undef_to_locs.items())) |
| |
| self.add_failure(""" |
| Found references to undefined Kconfig symbols. If any of these are false |
| positives, then add them to UNDEF_KCONFIG_WHITELIST in {} in the ci-tools repo. |
| |
| If the reference is for a comment like /* CONFIG_FOO_* */ (or |
| /* CONFIG_FOO_*_... */), then please use exactly that form (with the '*'). The |
| CI check knows not to flag it. |
| |
| More generally, a reference followed by $, @, {{, *, or ## will never be |
| flagged. |
| |
| {}""".format(os.path.basename(__file__), undef_desc)) |
| |
| |
| def get_defined_syms(kconf): |
| # Returns a set() with the names of all defined Kconfig symbols (with no |
| # 'CONFIG_' prefix). This is complicated by samples and tests defining |
| # their own Kconfig trees. For those, just grep for 'config FOO' to find |
| # definitions. Doing it "properly" with Kconfiglib is still useful for the |
| # main tree, because some symbols are defined using preprocessor macros. |
| |
| # Warning: Needs to work with both --perl-regexp and the 're' module. |
| # (?:...) is a non-capturing group. |
| regex = r"^\s*(?:menu)?config\s*([A-Z0-9_]+)\s*(?:#|$)" |
| |
| # Grep samples/ and tests/ for symbol definitions |
| grep_stdout = git("grep", "-I", "-h", "--perl-regexp", regex, "--", |
| ":samples", ":tests", cwd=ZEPHYR_BASE) |
| |
| # Symbols from the main Kconfig tree + grepped definitions from samples and |
| # tests |
| return set([sym.name for sym in kconf.unique_defined_syms] |
| + re.findall(regex, grep_stdout, re.MULTILINE)) |
| |
| |
| # Many of these are symbols used as examples. Note that the list is sorted |
| # alphabetically, and skips the CONFIG_ prefix. |
| UNDEF_KCONFIG_WHITELIST = { |
| "ALSO_MISSING", |
| "APP_LINK_WITH_", |
| "CDC_ACM_PORT_NAME_", |
| "CLOCK_STM32_SYSCLK_SRC_", |
| "CMU", |
| "BT_6LOWPAN", # Defined in Linux, mentioned in docs |
| "COUNTER_RTC_STM32_CLOCK_SRC", |
| "CRC", # Used in TI CC13x2 / CC26x2 SDK comment |
| "DEEP_SLEEP", # #defined by RV32M1 in ext/ |
| "DESCRIPTION", |
| "ERR", |
| "ESP_DIF_LIBRARY", # Referenced in CMake comment |
| "EXPERIMENTAL", |
| "FFT", # Used as an example in cmake/extensions.cmake |
| "FLAG", # Used as an example |
| "FOO", |
| "FOO_LOG_LEVEL", |
| "FOO_SETTING_1", |
| "FOO_SETTING_2", |
| "LIS2DW12_INT_PIN", |
| "LSM6DSO_INT_PIN", |
| "MISSING", |
| "MODULES", |
| "MYFEATURE", |
| "MY_DRIVER_0", |
| "NORMAL_SLEEP", # #defined by RV32M1 in ext/ |
| "OPT", |
| "OPT_0", |
| "PEDO_THS_MIN", |
| "REG1", |
| "REG2", |
| "SAMPLE_MODULE_LOG_LEVEL", # Used as an example in samples/subsys/logging |
| "SEL", |
| "SHIFT", |
| "SOC_WATCH", # Issue 13749 |
| "SOME_BOOL", |
| "SOME_INT", |
| "SOME_OTHER_BOOL", |
| "SOME_STRING", |
| "SRAM2", # Referenced in a comment in samples/application_development |
| "STACK_SIZE", # Used as an example in the Kconfig docs |
| "STD_CPP", # Referenced in CMake comment |
| "TEST1", |
| "TYPE_BOOLEAN", |
| "USB_CONSOLE", |
| "USE_STDC_", |
| "WHATEVER", |
| } |
| |
| |
| class DeviceTreeCheck(ComplianceTest): |
| """ |
| Runs the dtlib and edtlib test suites in scripts/dts/. |
| """ |
| name = "Devicetree" |
| doc = "See https://docs.zephyrproject.org/latest/guides/dts/index.html for more details" |
| path_hint = ZEPHYR_BASE |
| |
| def run(self): |
| if not ZEPHYR_BASE: |
| self.skip("Not a Zephyr tree (ZEPHYR_BASE unset)") |
| |
| scripts_path = os.path.join(ZEPHYR_BASE, "scripts", "dts") |
| |
| sys.path.insert(0, scripts_path) |
| import testdtlib |
| import testedtlib |
| |
| # Hack: The test suites expect to be run from the scripts/dts |
| # directory, because they compare repr() output that contains relative |
| # paths against an expected string. Temporarily change the working |
| # directory to scripts/dts/. |
| # |
| # Warning: This is not thread-safe, though the test suites run in a |
| # fraction of a second. |
| old_dir = os.getcwd() |
| os.chdir(scripts_path) |
| try: |
| logger.info("cd %s && ./testdtlib.py", scripts_path) |
| testdtlib.run() |
| logger.info("cd %s && ./testedtlib.py", scripts_path) |
| testedtlib.run() |
| except SystemExit as e: |
| # The dtlib and edtlib test suites call sys.exit() on failure, |
| # which raises SystemExit. Let any errors in the test scripts |
| # themselves trickle through and turn into an internal CI error. |
| self.add_failure(str(e)) |
| except Exception as e: |
| # Report other exceptions as an internal test failure |
| self.error(str(e)) |
| finally: |
| # Restore working directory |
| os.chdir(old_dir) |
| |
| |
| class Codeowners(ComplianceTest): |
| """ |
| Check if added files have an owner. |
| """ |
| name = "Codeowners" |
| doc = "See https://help.github.com/articles/about-code-owners/ for more details." |
| path_hint = "<git-top>" |
| |
| def ls_owned_files(self, codeowners): |
| """Returns an OrderedDict mapping git patterns from the CODEOWNERS file |
| to the corresponding list of files found on the filesystem. It |
| unfortunately does not seem possible to invoke git and re-use |
| how 'git ignore' and/or 'git attributes' already implement this, |
| we must re-invent it. |
| """ |
| |
| # TODO: filter out files not in "git ls-files" (e.g., |
| # sanity-out) _if_ the overhead isn't too high for a clean tree. |
| # |
| # pathlib.match() doesn't support **, so it looks like we can't |
| # recursively glob the output of ls-files directly, only real |
| # files :-( |
| |
| pattern2files = collections.OrderedDict() |
| top_path = Path(GIT_TOP) |
| |
| with open(codeowners, "r") as codeo: |
| for lineno, line in enumerate(codeo, start=1): |
| |
| if line.startswith("#") or not line.strip(): |
| continue |
| |
| match = re.match(r"^([^\s,]+)\s+[^\s]+", line) |
| if not match: |
| self.add_failure( |
| "Invalid CODEOWNERS line %d\n\t%s" % |
| (lineno, line)) |
| continue |
| |
| git_patrn = match.group(1) |
| glob = self.git_pattern_to_glob(git_patrn) |
| files = [] |
| for abs_path in top_path.glob(glob): |
| # comparing strings is much faster later |
| files.append(str(abs_path.relative_to(top_path))) |
| |
| if not files: |
| self.add_failure("Path '{}' not found in the tree but is listed in " |
| "CODEOWNERS".format(git_patrn)) |
| |
| pattern2files[git_patrn] = files |
| |
| return pattern2files |
| |
| def git_pattern_to_glob(self, git_pattern): |
| """Appends and prepends '**[/*]' when needed. Result has neither a |
| leading nor a trailing slash. |
| """ |
| |
| if git_pattern.startswith("/"): |
| ret = git_pattern[1:] |
| else: |
| ret = "**/" + git_pattern |
| |
| if git_pattern.endswith("/"): |
| ret = ret + "**/*" |
| elif os.path.isdir(os.path.join(GIT_TOP, ret)): |
| self.add_failure("Expected '/' after directory '{}' " |
| "in CODEOWNERS".format(ret)) |
| |
| return ret |
| |
| def run(self): |
| # TODO: testing an old self.commit range that doesn't end |
| # with HEAD is most likely a mistake. Should warn, see |
| # https://github.com/zephyrproject-rtos/ci-tools/pull/24 |
| codeowners = os.path.join(GIT_TOP, "CODEOWNERS") |
| if not os.path.exists(codeowners): |
| self.skip("CODEOWNERS not available in this repo") |
| |
| name_changes = git("diff", "--name-only", "--diff-filter=ARCD", |
| COMMIT_RANGE) |
| |
| owners_changes = git("diff", "--name-only", COMMIT_RANGE, |
| "--", codeowners) |
| |
| if not name_changes and not owners_changes: |
| # TODO: 1. decouple basic and cheap CODEOWNERS syntax |
| # validation from the expensive ls_owned_files() scanning of |
| # the entire tree. 2. run the former always. |
| return |
| |
| logging.info("If this takes too long then cleanup and try again") |
| patrn2files = self.ls_owned_files(codeowners) |
| |
| # The way git finds Renames and Copies is not "exact science", |
| # however if one is missed then it will always be reported as an |
| # Addition instead. |
| new_files = git("diff", "--name-only", "--diff-filter=ARC", |
| COMMIT_RANGE).splitlines() |
| logging.debug("New files %s", new_files) |
| |
| # Convert to pathlib.Path string representation (e.g., |
| # backslashes 'dir1\dir2\' on Windows) to be consistent |
| # with self.ls_owned_files() |
| new_files = [str(Path(f)) for f in new_files] |
| |
| new_not_owned = [] |
| for newf in new_files: |
| f_is_owned = False |
| |
| for git_pat, owned in patrn2files.items(): |
| logging.debug("Scanning %s for %s", git_pat, newf) |
| |
| if newf in owned: |
| logging.info("%s matches new file %s", git_pat, newf) |
| f_is_owned = True |
| # Unlike github, we don't care about finding any |
| # more specific owner. |
| break |
| |
| if not f_is_owned: |
| new_not_owned.append(newf) |
| |
| if new_not_owned: |
| self.add_failure("New files added that are not covered in " |
| "CODEOWNERS:\n\n" + "\n".join(new_not_owned) + |
| "\n\nPlease add one or more entries in the " |
| "CODEOWNERS file to cover those files") |
| |
| |
| class Nits(ComplianceTest): |
| """ |
| Checks various nits in added/modified files. Doesn't check stuff that's |
| already covered by e.g. checkpatch.pl and pylint. |
| """ |
| name = "Nits" |
| doc = "See https://docs.zephyrproject.org/latest/contribute/#coding-style for more details." |
| path_hint = "<git-top>" |
| |
| def run(self): |
| # Loop through added/modified files |
| for fname in git("diff", "--name-only", "--diff-filter=d", |
| COMMIT_RANGE).splitlines(): |
| if "Kconfig" in fname: |
| self.check_kconfig_header(fname) |
| self.check_redundant_zephyr_source(fname) |
| |
| if fname.startswith("dts/bindings/"): |
| self.check_redundant_document_separator(fname) |
| |
| if fname.endswith((".c", ".conf", ".cpp", ".dts", ".overlay", |
| ".h", ".ld", ".py", ".rst", ".txt", ".yaml", |
| ".yml")) or \ |
| "Kconfig" in fname or \ |
| "defconfig" in fname or \ |
| fname == "README": |
| |
| self.check_source_file(fname) |
| |
| def check_kconfig_header(self, fname): |
| # Checks for a spammy copy-pasted header format |
| |
| with open(os.path.join(GIT_TOP, fname), encoding="utf-8") as f: |
| contents = f.read() |
| |
| # 'Kconfig - yada yada' has a copy-pasted redundant filename at the |
| # top. This probably means all of the header was copy-pasted. |
| if re.match(r"\s*#\s*(K|k)config[\w.-]*\s*-", contents): |
| self.add_failure(""" |
| Please use this format for the header in '{}' (see |
| https://docs.zephyrproject.org/latest/guides/kconfig/index.html#header-comments-and-other-nits): |
| |
| # <Overview of symbols defined in the file, preferably in plain English> |
| (Blank line) |
| # Copyright (c) 2019 ... |
| # SPDX-License-Identifier: <License> |
| (Blank line) |
| (Kconfig definitions) |
| |
| Skip the "Kconfig - " part of the first line, since it's clear that the comment |
| is about Kconfig from context. The "# Kconfig - " is what triggers this |
| failure. |
| """.format(fname)) |
| |
| def check_redundant_zephyr_source(self, fname): |
| # Checks for 'source "$(ZEPHYR_BASE)/Kconfig[.zephyr]"', which can be |
| # be simplified to 'source "Kconfig[.zephyr]"' |
| |
| with open(os.path.join(GIT_TOP, fname), encoding="utf-8") as f: |
| # Look for e.g. rsource as well, for completeness |
| match = re.search( |
| r'^\s*(?:o|r|or)?source\s*"\$\(?ZEPHYR_BASE\)?/(Kconfig(?:\.zephyr)?)"', |
| f.read(), re.MULTILINE) |
| |
| if match: |
| self.add_failure(""" |
| Redundant 'source "$(ZEPHYR_BASE)/{0}" in '{1}'. Just do 'source "{0}"' |
| instead. The $srctree environment variable already points to the Zephyr root, |
| and all 'source's are relative to it.""".format(match.group(1), fname)) |
| |
| def check_redundant_document_separator(self, fname): |
| # Looks for redundant '...' document separators in bindings |
| |
| with open(os.path.join(GIT_TOP, fname), encoding="utf-8") as f: |
| if re.search(r"^\.\.\.", f.read(), re.MULTILINE): |
| self.add_failure(f"""\ |
| Redundant '...' document separator in {fname}. Binding YAML files are never |
| concatenated together, so no document separators are needed.""") |
| |
| def check_source_file(self, fname): |
| # Generic nits related to various source files |
| |
| with open(os.path.join(GIT_TOP, fname), encoding="utf-8") as f: |
| contents = f.read() |
| |
| if not contents.endswith("\n"): |
| self.add_failure("Missing newline at end of '{}'. Check your text " |
| "editor settings.".format(fname)) |
| |
| if contents.startswith("\n"): |
| self.add_failure("Please remove blank lines at start of '{}'" |
| .format(fname)) |
| |
| if contents.endswith("\n\n"): |
| self.add_failure("Please remove blank lines at end of '{}'" |
| .format(fname)) |
| |
| |
| class GitLint(ComplianceTest): |
| """ |
| Runs gitlint on the commits and finds issues with style and syntax |
| |
| """ |
| name = "Gitlint" |
| doc = "See https://docs.zephyrproject.org/latest/contribute/#commit-guidelines for more details" |
| path_hint = "<git-top>" |
| |
| def run(self): |
| # By default gitlint looks for .gitlint configuration only in |
| # the current directory |
| proc = subprocess.Popen('gitlint --commits ' + COMMIT_RANGE, |
| stdout=subprocess.PIPE, stderr=subprocess.STDOUT, |
| shell=True, cwd=GIT_TOP) |
| |
| msg = "" |
| if proc.wait() != 0: |
| msg = proc.stdout.read() |
| |
| if msg != "": |
| self.add_failure(msg.decode("utf-8")) |
| |
| |
| class PyLint(ComplianceTest): |
| """ |
| Runs pylint on all .py files, with a limited set of checks enabled. The |
| configuration is in the pylintrc file. |
| """ |
| name = "pylint" |
| doc = "See https://www.pylint.org/ for more details" |
| path_hint = "<git-top>" |
| |
| def run(self): |
| # Path to pylint configuration file |
| pylintrc = os.path.abspath(os.path.join(os.path.dirname(__file__), |
| "pylintrc")) |
| |
| # List of files added/modified by the commit(s). |
| files = git( |
| "diff", "--name-only", "--diff-filter=d", COMMIT_RANGE, "--", |
| # Skip to work around crash in pylint 2.2.2: |
| # https://github.com/PyCQA/pylint/issues/2906 |
| ":!boards/xtensa/intel_s1000_crb/support/create_board_img.py") \ |
| .splitlines() |
| |
| # Filter out everything but Python files. Keep filenames |
| # relative (to GIT_TOP) to stay farther from any command line |
| # limit. |
| py_files = filter_py(GIT_TOP, files) |
| if not py_files: |
| return |
| |
| pylintcmd = ["pylint", "--rcfile=" + pylintrc] + py_files |
| logger.info(cmd2str(pylintcmd)) |
| try: |
| # Run pylint on added/modified Python files |
| process = subprocess.Popen( |
| pylintcmd, |
| stdout=subprocess.PIPE, |
| stderr=subprocess.PIPE, |
| cwd=GIT_TOP) |
| except OSError as e: |
| self.error(f"Failed to run {cmd2str(pylintcmd)}: {e}") |
| |
| stdout, stderr = process.communicate() |
| if process.returncode or stderr: |
| # Issues found, or a problem with pylint itself |
| self.add_failure(stdout.decode("utf-8") + stderr.decode("utf-8")) |
| |
| |
| def filter_py(root, fnames): |
| # PyLint check helper. Returns all Python script filenames among the |
| # filenames in 'fnames', relative to directory 'root'. Uses the |
| # python-magic library, so that we can detect Python files that |
| # don't end in .py as well. python-magic is a frontend to libmagic, |
| # which is also used by 'file'. |
| |
| return [fname for fname in fnames |
| if fname.endswith(".py") or |
| magic.from_file(os.path.join(root, fname), |
| mime=True) == "text/x-python"] |
| |
| |
| class Identity(ComplianceTest): |
| """ |
| Checks if Emails of author and signed-off messages are consistent. |
| """ |
| name = "Identity" |
| doc = "See https://docs.zephyrproject.org/latest/contribute/#commit-guidelines for more details" |
| # git rev-list and git log don't depend on the current (sub)directory |
| # unless explicited |
| path_hint = "<git-top>" |
| |
| def run(self): |
| for shaidx in get_shas(COMMIT_RANGE): |
| commit = git("log", "--decorate=short", "-n 1", shaidx) |
| signed = [] |
| author = "" |
| sha = "" |
| parsed_addr = None |
| for line in commit.split("\n"): |
| match = re.search(r"^commit\s([^\s]*)", line) |
| if match: |
| sha = match.group(1) |
| match = re.search(r"^Author:\s(.*)", line) |
| if match: |
| author = match.group(1) |
| parsed_addr = parseaddr(author) |
| match = re.search(r"signed-off-by:\s(.*)", line, re.IGNORECASE) |
| if match: |
| signed.append(match.group(1)) |
| |
| error1 = "%s: author email (%s) needs to match one of the signed-off-by entries." % ( |
| sha, author) |
| error2 = "%s: author email (%s) does not follow the syntax: First Last <email>." % ( |
| sha, author) |
| error3 = "%s: author email (%s) must be a real email and cannot end in @users.noreply.github.com" % ( |
| sha, author) |
| failure = None |
| if author not in signed: |
| failure = error1 |
| |
| if not parsed_addr or len(parsed_addr[0].split(" ")) < 2: |
| if not failure: |
| |
| failure = error2 |
| else: |
| failure = failure + "\n" + error2 |
| elif parsed_addr[1].endswith("@users.noreply.github.com"): |
| failure = error3 |
| |
| if failure: |
| self.add_failure(failure) |
| |
| |
| def init_logs(cli_arg): |
| # Initializes logging |
| |
| # TODO: there may be a shorter version thanks to: |
| # logging.basicConfig(...) |
| |
| global logger |
| |
| level = os.environ.get('LOG_LEVEL', "WARN") |
| |
| console = logging.StreamHandler() |
| console.setFormatter(logging.Formatter('%(levelname)-8s: %(message)s')) |
| |
| logger = logging.getLogger('') |
| logger.addHandler(console) |
| logger.setLevel(cli_arg if cli_arg else level) |
| |
| logging.info("Log init completed, level=%s", |
| logging.getLevelName(logger.getEffectiveLevel())) |
| |
| |
| |
| def parse_args(): |
| parser = argparse.ArgumentParser( |
| description="Check for coding style and documentation warnings.") |
| parser.add_argument('-c', '--commits', default="HEAD~1..", |
| help='''Commit range in the form: a..[b], default is |
| HEAD~1..HEAD''') |
| parser.add_argument('-r', '--repo', default=None, |
| help="GitHub repository") |
| parser.add_argument('-p', '--pull-request', default=0, type=int, |
| help="Pull request number") |
| parser.add_argument('-S', '--sha', default=None, help="Commit SHA") |
| parser.add_argument('-o', '--output', default="compliance.xml", |
| help='''Name of outfile in JUnit format, |
| default is ./compliance.xml''') |
| |
| parser.add_argument('-l', '--list', action="store_true", |
| help="List all checks and exit") |
| |
| parser.add_argument("-v", "--loglevel", help="python logging level") |
| |
| parser.add_argument('-m', '--module', action="append", default=[], |
| help="Checks to run. All checks by default.") |
| |
| parser.add_argument('-e', '--exclude-module', action="append", default=[], |
| help="Do not run the specified checks") |
| |
| parser.add_argument('-j', '--previous-run', default=None, |
| help='''Pre-load JUnit results in XML format |
| from a previous run and combine with new results.''') |
| |
| return parser.parse_args() |
| |
| def _main(args): |
| # The "real" main(), which is wrapped to catch exceptions and report them |
| # to GitHub. Returns the number of test failures. |
| |
| # The absolute path of the top-level git directory. Initialize it here so |
| # that issues running Git can be reported to GitHub. |
| global GIT_TOP |
| GIT_TOP = git("rev-parse", "--show-toplevel") |
| |
| # The commit range passed in --commit, e.g. "HEAD~3" |
| global COMMIT_RANGE |
| COMMIT_RANGE = args.commits |
| |
| init_logs(args.loglevel) |
| |
| if args.list: |
| for testcase in ComplianceTest.__subclasses__(): |
| print(testcase.name) |
| return 0 |
| |
| # Load saved test results from an earlier run, if requested |
| if args.previous_run: |
| if not os.path.exists(args.previous_run): |
| # This probably means that an earlier pass had an internal error |
| # (the script is currently run multiple times by the ci-pipelines |
| # repo). Since that earlier pass might've posted an error to |
| # GitHub, avoid generating a GitHub comment here, by avoiding |
| # sys.exit() (which gets caught in main()). |
| print("error: '{}' not found".format(args.previous_run), |
| file=sys.stderr) |
| return 1 |
| |
| logging.info("Loading previous results from " + args.previous_run) |
| for loaded_suite in JUnitXml.fromfile(args.previous_run): |
| suite = loaded_suite |
| break |
| else: |
| suite = TestSuite("Compliance") |
| |
| for testcase in ComplianceTest.__subclasses__(): |
| # "Modules" and "testcases" are the same thing. Better flags would have |
| # been --tests and --exclude-tests or the like, but it's awkward to |
| # change now. |
| |
| if args.module and testcase.name not in args.module: |
| continue |
| |
| if testcase.name in args.exclude_module: |
| print("Skipping " + testcase.name) |
| continue |
| |
| test = testcase() |
| try: |
| print(f"Running {test.name:16} tests in " |
| f"{GIT_TOP if test.path_hint == '<git-top>' else test.path_hint} ...") |
| test.run() |
| except EndTest: |
| pass |
| |
| suite.add_testcase(test.case) |
| |
| xml = JUnitXml() |
| xml.add_testsuite(suite) |
| xml.update_statistics() |
| xml.write(args.output, pretty=True) |
| |
| failed_cases = [] |
| name2doc = {testcase.name: testcase.doc |
| for testcase in ComplianceTest.__subclasses__()} |
| |
| for case in suite: |
| if case.result: |
| if case.result.type == 'skipped': |
| logging.warning("Skipped %s, %s", case.name, case.result.message) |
| else: |
| failed_cases.append(case) |
| else: |
| # Some checks like codeowners can produce no .result |
| logging.info("No JUnit result for %s", case.name) |
| |
| n_fails = len(failed_cases) |
| |
| if n_fails: |
| print("{} checks failed".format(n_fails)) |
| for case in failed_cases: |
| # not clear why junitxml doesn't clearly expose the most |
| # important part of its underlying etree.Element |
| errmsg = case.result._elem.text |
| logging.error("Test %s failed: %s", case.name, |
| errmsg.strip() if errmsg else case.result.message) |
| |
| with open(f"{case.name}.txt", "w") as f: |
| docs = name2doc.get(case.name) |
| f.write(f"{docs}\n\n") |
| f.write(errmsg.strip() if errmsg else case.result.message) |
| |
| print("\nComplete results in " + args.output) |
| return n_fails |
| |
| |
| def main(): |
| args = parse_args() |
| |
| try: |
| n_fails = _main(args) |
| except BaseException: |
| # Catch BaseException instead of Exception to include stuff like |
| # SystemExit (raised by sys.exit()) |
| print(format(__file__, traceback.format_exc())) |
| |
| raise |
| |
| sys.exit(n_fails) |
| |
| |
| def cmd2str(cmd): |
| # Formats the command-line arguments in the iterable 'cmd' into a string, |
| # for error messages and the like |
| |
| return " ".join(shlex.quote(word) for word in cmd) |
| |
| |
| def err(msg): |
| cmd = sys.argv[0] # Empty if missing |
| if cmd: |
| cmd += ": " |
| sys.exit(cmd + "error: " + msg) |
| |
| |
| if __name__ == "__main__": |
| main() |