| #!/usr/bin/env python3 |
| # |
| # Copyright (c) 2021 Raspberry Pi (Trading) Ltd. |
| # |
| # SPDX-License-Identifier: BSD-3-Clause |
| # |
| # |
| # Script to scan the Raspberry Pi Pico SDK tree searching for configuration items |
| # Outputs a tab separated file of the configuration item: |
| # name location platform chip description type advanced default depends enumvalues group max min |
| # |
| # Usage: |
| # |
| # tools/extract_configs.py <root of repo> [output file] |
| # |
| # If not specified, output file will be `pico_configs.tsv` |
| |
| |
| import os |
| import sys |
| import re |
| import csv |
| import logging |
| |
| from collections import defaultdict |
| |
| logger = logging.getLogger(__name__) |
| logging.basicConfig(level=logging.INFO) |
| |
| scandir = sys.argv[1] |
| outfile = sys.argv[2] if len(sys.argv) > 2 else 'pico_configs.tsv' |
| |
| BASE_CONFIG_NAME = 'PICO_CONFIG' |
| BASE_CONFIG_RE = re.compile(r'\b{}\b'.format(BASE_CONFIG_NAME)) |
| BASE_CMAKE_CONFIG_NAME = 'PICO_CMAKE_CONFIG' |
| BASE_CMAKE_CONFIG_RE = re.compile(r'\b{}\b'.format(BASE_CMAKE_CONFIG_NAME)) |
| BASE_BUILD_DEFINE_NAME = 'PICO_BUILD_DEFINE' |
| BASE_BUILD_DEFINE_RE = re.compile(r'\b{}\b'.format(BASE_BUILD_DEFINE_NAME)) |
| |
| CONFIG_RE = re.compile(r'//\s+{}:\s+(\w+),\s+([^,]+)(?:,\s+(.*))?$'.format(BASE_CONFIG_NAME)) |
| DEFINE_RE = re.compile(r'#define\s+(\w+)\s+(.+?)(\s*///.*)?$') |
| |
| ALLOWED_CONFIG_PROPERTIES = set(['type', 'default', 'min', 'max', 'enumvalues', 'group', 'advanced', 'depends']) |
| |
| CHIP_NAMES = ["rp2040", "rp2350"] |
| |
| chips_all_configs = defaultdict(dict) |
| all_attrs = set() |
| chips_all_descriptions = defaultdict(dict) |
| chips_all_defines = defaultdict(dict) |
| |
| |
| |
| def ValidateAttrs(config_name, config_attrs, file_path, linenum): |
| _type = config_attrs.get('type', 'int') |
| |
| # Validate attrs |
| for key in config_attrs.keys(): |
| if key not in ALLOWED_CONFIG_PROPERTIES: |
| raise Exception('{} at {}:{} has unexpected property "{}"'.format(config_name, file_path, linenum, key)) |
| |
| if _type == 'int': |
| assert 'enumvalues' not in config_attrs |
| _min = _max = _default = None |
| if config_attrs.get('min', None) is not None: |
| try: |
| value = config_attrs['min'] |
| m = re.match(r'^(\d+)e(\d+)$', value.lower()) |
| if m: |
| _min = int(m.group(1)) * 10**int(m.group(2)) |
| else: |
| _min = int(value, 0) |
| except ValueError: |
| logger.info('{} at {}:{} has non-integer min value "{}"'.format(config_name, file_path, linenum, config_attrs['min'])) |
| if config_attrs.get('max', None) is not None: |
| try: |
| value = config_attrs['max'] |
| m = re.match(r'^(\d+)e(\d+)$', value.lower()) |
| if m: |
| _max = int(m.group(1)) * 10**int(m.group(2)) |
| else: |
| _max = int(value, 0) |
| except ValueError: |
| logger.info('{} at {}:{} has non-integer max value "{}"'.format(config_name, file_path, linenum, config_attrs['max'])) |
| if config_attrs.get('default', None) is not None: |
| if '/' not in config_attrs['default']: |
| try: |
| value = config_attrs['default'] |
| m = re.match(r'^(\d+)e(\d+)$', value.lower()) |
| if m: |
| _default = int(m.group(1)) * 10**int(m.group(2)) |
| else: |
| _default = int(value, 0) |
| except ValueError: |
| logger.info('{} at {}:{} has non-integer default value "{}"'.format(config_name, file_path, linenum, config_attrs['default'])) |
| if _min is not None and _max is not None: |
| if _min > _max: |
| raise Exception('{} at {}:{} has min {} > max {}'.format(config_name, file_path, linenum, config_attrs['min'], config_attrs['max'])) |
| if _min is not None and _default is not None: |
| if _min > _default: |
| raise Exception('{} at {}:{} has min {} > default {}'.format(config_name, file_path, linenum, config_attrs['min'], config_attrs['default'])) |
| if _default is not None and _max is not None: |
| if _default > _max: |
| raise Exception('{} at {}:{} has default {} > max {}'.format(config_name, file_path, linenum, config_attrs['default'], config_attrs['max'])) |
| elif _type == 'bool': |
| |
| assert 'min' not in config_attrs |
| assert 'max' not in config_attrs |
| assert 'enumvalues' not in config_attrs |
| |
| _default = config_attrs.get('default', None) |
| if _default is not None: |
| if '/' not in _default: |
| if (_default not in ('0', '1')) and (_default not in all_config_names): |
| logger.info('{} at {}:{} has non-integer default value "{}"'.format(config_name, file_path, linenum, config_attrs['default'])) |
| |
| elif _type == 'enum': |
| |
| assert 'min' not in config_attrs |
| assert 'max' not in config_attrs |
| assert 'enumvalues' in config_attrs |
| |
| _enumvalues = tuple(config_attrs['enumvalues'].split('|')) |
| _default = None |
| if config_attrs.get('default', None) is not None: |
| _default = config_attrs['default'] |
| if _default is not None: |
| if _default not in _enumvalues: |
| raise Exception('{} at {}:{} has default value {} which isn\'t in list of enumvalues {}'.format(config_name, file_path, linenum, config_attrs['default'], config_attrs['enumvalues'])) |
| else: |
| raise Exception("Found unknown {} type {} at {}:{}".format(BASE_CONFIG_NAME, _type, file_path, linenum)) |
| |
| |
| |
| |
| # Scan all .c and .h files in the specific path, recursively. |
| |
| for dirpath, dirnames, filenames in os.walk(scandir): |
| for filename in filenames: |
| file_ext = os.path.splitext(filename)[1] |
| if file_ext in ('.c', '.h'): |
| file_path = os.path.join(dirpath, filename) |
| applicable = "all" |
| for chip in (*CHIP_NAMES, "host"): |
| if "/{}/".format(chip) in dirpath: |
| applicable = chip |
| break |
| |
| with open(file_path, encoding="ISO-8859-1") as fh: |
| linenum = 0 |
| for line in fh.readlines(): |
| linenum += 1 |
| line = line.strip() |
| if BASE_CMAKE_CONFIG_RE.search(line): |
| raise Exception("Found {} at {}:{} ({}) which isn't expected in {} files".format(BASE_CMAKE_CONFIG_NAME, file_path, linenum, line, file_ext)) |
| elif BASE_BUILD_DEFINE_RE.search(line): |
| raise Exception("Found {} at {}:{} ({}) which isn't expected in {} files".format(BASE_BUILD_DEFINE_NAME, file_path, linenum, line, file_ext)) |
| elif BASE_CONFIG_RE.search(line): |
| m = CONFIG_RE.match(line) |
| if not m: |
| if line.startswith("//// "): |
| logger.info("Possible misformatted {} at {}:{} ({})".format(BASE_CONFIG_NAME, file_path, linenum, line)) |
| else: |
| raise Exception("Found misformatted {} at {}:{} ({})".format(BASE_CONFIG_NAME, file_path, linenum, line)) |
| else: |
| config_name = m.group(1) |
| config_description = m.group(2) |
| _attrs = m.group(3) |
| # allow commas to appear inside brackets by converting them to and from NULL chars |
| _attrs = re.sub(r'(\(.+\))', lambda m: m.group(1).replace(',', '\0'), _attrs) |
| |
| if '=' in config_description: |
| raise Exception("For {} at {}:{} the description was set to '{}' - has the description field been omitted?".format(config_name, file_path, linenum, config_description)) |
| all_descriptions = chips_all_descriptions[applicable] |
| if config_description in all_descriptions: |
| raise Exception("Found description {} at {}:{} but it was already used at {}:{}".format(config_description, file_path, linenum, os.path.join(scandir, all_descriptions[config_description]['filename']), all_descriptions[config_description]['line_number'])) |
| else: |
| all_descriptions[config_description] = {'config_name': config_name, 'filename': os.path.relpath(file_path, scandir), 'line_number': linenum} |
| |
| config_attrs = {} |
| prev = None |
| # Handle case where attr value contains a comma |
| for item in _attrs.split(','): |
| if "=" not in item: |
| assert(prev) |
| item = prev + "," + item |
| try: |
| k, v = (i.strip() for i in item.split('=')) |
| except ValueError: |
| raise Exception('{} at {}:{} has malformed value {}'.format(config_name, file_path, linenum, item)) |
| config_attrs[k] = v.replace('\0', ',') |
| all_attrs.add(k) |
| prev = item |
| #print(file_path, config_name, config_attrs) |
| |
| if 'group' not in config_attrs: |
| raise Exception('{} at {}:{} has no group attribute'.format(config_name, file_path, linenum)) |
| |
| #print(file_path, config_name, config_attrs) |
| all_configs = chips_all_configs[applicable] |
| if config_name in all_configs: |
| raise Exception("Found {} at {}:{} but it was already declared at {}:{}".format(config_name, file_path, linenum, os.path.join(scandir, all_configs[config_name]['filename']), all_configs[config_name]['line_number'])) |
| else: |
| all_configs[config_name] = {'attrs': config_attrs, 'filename': os.path.relpath(file_path, scandir), 'line_number': linenum, 'description': config_description} |
| else: |
| m = DEFINE_RE.match(line) |
| if m: |
| name = m.group(1) |
| value = m.group(2) |
| # discard any 'u' qualifier |
| m = re.match(r'^((0x)?\d+)u$', value.lower()) |
| if m: |
| value = m.group(1) |
| else: |
| # discard any '_u(X)' macro |
| m = re.match(r'^_u\(((0x)?\d+)\)$', value.lower()) |
| if m: |
| value = m.group(1) |
| all_defines = chips_all_defines[applicable] |
| if name not in all_defines: |
| all_defines[name] = dict() |
| if value not in all_defines[name]: |
| all_defines[name][value] = set() |
| all_defines[name][value] = (file_path, linenum) |
| |
| all_config_names = set() |
| for all_configs in chips_all_configs.values(): |
| all_config_names.update(all_configs.keys()) |
| |
| # Check for defines with missing PICO_CONFIG entries |
| chips_resolved_defines = defaultdict(dict) |
| for applicable, all_defines in chips_all_defines.items(): |
| for d in all_defines: |
| if d not in all_config_names and d.startswith("PICO_"): |
| logger.warning("Potential unmarked PICO define {}".format(d)) |
| resolved_defines = chips_resolved_defines[applicable] |
| # resolve "nested defines" - this allows e.g. USB_DPRAM_MAX to resolve to USB_DPRAM_SIZE which is set to 4096 (which then matches the relevant PICO_CONFIG entry) |
| for val in all_defines[d]: |
| if val in all_defines: |
| resolved_defines[d] = all_defines[val] |
| |
| for applicable, all_configs in chips_all_configs.items(): |
| all_defines = chips_all_defines[applicable] |
| for config_name, config_obj in all_configs.items(): |
| file_path = os.path.join(scandir, config_obj['filename']) |
| linenum = config_obj['line_number'] |
| |
| ValidateAttrs(config_name, config_obj['attrs'], file_path, linenum) |
| |
| # Check that default values match up |
| if 'default' in config_obj['attrs']: |
| config_default = config_obj['attrs']['default'] |
| if config_name in all_defines: |
| defines_obj = all_defines[config_name] |
| if config_default not in defines_obj and (config_name not in resolved_defines or config_default not in resolved_defines[config_name]): |
| if '/' in config_default or ' ' in config_default: |
| continue |
| # There _may_ be multiple matching defines, but arbitrarily display just one in the error message |
| first_define_value = list(defines_obj.keys())[0] |
| first_define_file_path, first_define_linenum = defines_obj[first_define_value] |
| raise Exception('Found {} at {}:{} with a default of {}, but #define says {} (at {}:{})'.format(config_name, file_path, linenum, config_default, first_define_value, first_define_file_path, first_define_linenum)) |
| else: |
| raise Exception('Found {} at {}:{} with a default of {}, but no matching #define found'.format(config_name, file_path, linenum, config_default)) |
| |
| # All settings in "host" should also be in "all" |
| for config_name, config_obj in chips_all_configs["host"].items(): |
| if config_name not in chips_all_configs["all"]: |
| file_path = os.path.join(scandir, config_obj['filename']) |
| linenum = config_obj['line_number'] |
| raise Exception("Found 'host' config {} at {}:{}, but no matching non-host config found".format(config_name, file_path, linenum)) |
| |
| # Any chip-specific settings should not be in "all" |
| for chip in CHIP_NAMES: |
| for config_name, chip_config_obj in chips_all_configs[chip].items(): |
| if config_name in chips_all_configs["all"]: |
| all_config_obj = chips_all_configs["all"][config_name] |
| chip_file_path = os.path.join(scandir, chip_config_obj['filename']) |
| chip_linenum = chip_config_obj['line_number'] |
| all_file_path = os.path.join(scandir, all_config_obj['filename']) |
| all_linenum = all_config_obj['line_number'] |
| raise Exception("'{}' config {} at {}:{} also found at {}:{}".format(chip, config_name, chip_file_path, chip_linenum, all_file_path, all_linenum)) |
| |
| def build_mismatch_exception_message(name, thing, config_obj1, value1, config_obj2, value2): |
| obj1_filepath = os.path.join(scandir, config_obj1['filename']) |
| obj2_filepath = os.path.join(scandir, config_obj2['filename']) |
| return "'{}' {} mismatch at {}:{} ({}) and {}:{} ({})".format(name, thing, obj1_filepath, config_obj1['line_number'], value1, obj2_filepath, config_obj2['line_number'], value2) |
| |
| # Check that any identically-named setttings have appropriate matching attributes |
| for applicable in chips_all_configs: |
| for other in chips_all_configs: |
| if other == applicable: |
| continue |
| for config_name, applicable_config_obj in chips_all_configs[applicable].items(): |
| if config_name in chips_all_configs[other]: |
| other_config_obj = chips_all_configs[other][config_name] |
| # Check that fields match |
| for field in ['description']: |
| applicable_value = applicable_config_obj[field] |
| other_value = other_config_obj[field] |
| if applicable_value != other_value: |
| raise Exception(build_mismatch_exception_message(config_name, field, applicable_config_obj, applicable_value, other_config_obj, other_value)) |
| # Check that attributes match |
| for attr in applicable_config_obj['attrs']: |
| if attr != 'default': # totally fine for defaults to vary per-platform |
| applicable_value = applicable_config_obj['attrs'][attr] |
| other_value = other_config_obj['attrs'][attr] |
| if applicable_value != other_value: |
| raise Exception(build_mismatch_exception_message(config_name, "attribute '{}'".format(attr), applicable_config_obj, applicable_value, other_config_obj, other_value)) |
| |
| # Sort the output alphabetically by name and then by chip |
| output_rows = set() |
| for chip in (*CHIP_NAMES, "host", "all"): |
| if chip in chips_all_configs: |
| all_configs = chips_all_configs[chip] |
| for config_name in all_configs: |
| output_rows.add((config_name, chip)) |
| |
| with open(outfile, 'w', newline='') as csvfile: |
| fieldnames = ('name', 'location', 'platform', 'chip', 'description', 'type') + tuple(sorted(all_attrs - set(['type']))) |
| writer = csv.DictWriter(csvfile, fieldnames=fieldnames, extrasaction='ignore', dialect='excel-tab') |
| |
| writer.writeheader() |
| for config_name, chip in sorted(output_rows): |
| config_obj = chips_all_configs[chip][config_name] |
| writer.writerow({'name': config_name, 'location': '/{}:{}'.format(config_obj['filename'], config_obj['line_number']), 'platform': "host" if chip == "host" else "rp2", 'chip': chip if chip in CHIP_NAMES else "all", 'description': config_obj['description'], **config_obj['attrs']}) |