parent
ca2259d986
commit
b8f6536b29
@ -0,0 +1,495 @@ |
|||||||
|
"""Functions that help us generate and use info.json files. |
||||||
|
""" |
||||||
|
from functools import lru_cache |
||||||
|
from glob import glob |
||||||
|
from pathlib import Path |
||||||
|
|
||||||
|
import jsonschema |
||||||
|
from dotty_dict import dotty |
||||||
|
from milc import cli |
||||||
|
|
||||||
|
from qmk.constants import CHIBIOS_PROCESSORS, LUFA_PROCESSORS, VUSB_PROCESSORS |
||||||
|
from qmk.c_parse import find_layouts |
||||||
|
from qmk.json_schema import deep_update, json_load, keyboard_validate |
||||||
|
from qmk.keyboard import config_h, rules_mk |
||||||
|
from qmk.makefile import parse_rules_mk_file |
||||||
|
from qmk.math import compute |
||||||
|
|
||||||
|
true_values = ['1', 'on', 'yes', 'true'] |
||||||
|
false_values = ['0', 'off', 'no', 'false'] |
||||||
|
|
||||||
|
|
||||||
|
@lru_cache(maxsize=None) |
||||||
|
def basic_info_json(keyboard): |
||||||
|
"""Generate a subset of info.json for a specific keyboard. |
||||||
|
|
||||||
|
This does no validation, and should only be used as needed to avoid loops or when performance is critical. |
||||||
|
""" |
||||||
|
cur_dir = Path('keyboards') |
||||||
|
rules = parse_rules_mk_file(cur_dir / keyboard / 'rules.mk') |
||||||
|
|
||||||
|
if 'DEFAULT_FOLDER' in rules: |
||||||
|
keyboard = rules['DEFAULT_FOLDER'] |
||||||
|
rules = parse_rules_mk_file(cur_dir / keyboard / 'rules.mk', rules) |
||||||
|
|
||||||
|
info_data = { |
||||||
|
'keyboard_name': str(keyboard), |
||||||
|
'keyboard_folder': str(keyboard), |
||||||
|
'keymaps': {}, |
||||||
|
'layouts': {}, |
||||||
|
'parse_errors': [], |
||||||
|
'parse_warnings': [], |
||||||
|
'maintainer': 'qmk', |
||||||
|
} |
||||||
|
|
||||||
|
# Populate layout data |
||||||
|
layouts, aliases = _find_all_layouts(info_data, keyboard) |
||||||
|
|
||||||
|
if aliases: |
||||||
|
info_data['layout_aliases'] = aliases |
||||||
|
|
||||||
|
for layout_name, layout_json in layouts.items(): |
||||||
|
if not layout_name.startswith('LAYOUT_kc'): |
||||||
|
layout_json['c_macro'] = True |
||||||
|
info_data['layouts'][layout_name] = layout_json |
||||||
|
|
||||||
|
# Merge in the data from info.json, config.h, and rules.mk |
||||||
|
info_data = merge_info_jsons(keyboard, info_data) |
||||||
|
info_data = _extract_config_h(info_data) |
||||||
|
info_data = _extract_rules_mk(info_data) |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def _extract_features(info_data, rules): |
||||||
|
"""Find all the features enabled in rules.mk. |
||||||
|
""" |
||||||
|
# Special handling for bootmagic which also supports a "lite" mode. |
||||||
|
if rules.get('BOOTMAGIC_ENABLE') == 'lite': |
||||||
|
rules['BOOTMAGIC_LITE_ENABLE'] = 'on' |
||||||
|
del rules['BOOTMAGIC_ENABLE'] |
||||||
|
if rules.get('BOOTMAGIC_ENABLE') == 'full': |
||||||
|
rules['BOOTMAGIC_ENABLE'] = 'on' |
||||||
|
|
||||||
|
# Skip non-boolean features we haven't implemented special handling for |
||||||
|
for feature in 'HAPTIC_ENABLE', 'QWIIC_ENABLE': |
||||||
|
if rules.get(feature): |
||||||
|
del rules[feature] |
||||||
|
|
||||||
|
# Process the rest of the rules as booleans |
||||||
|
for key, value in rules.items(): |
||||||
|
if key.endswith('_ENABLE'): |
||||||
|
key = '_'.join(key.split('_')[:-1]).lower() |
||||||
|
value = True if value.lower() in true_values else False if value.lower() in false_values else value |
||||||
|
|
||||||
|
if 'config_h_features' not in info_data: |
||||||
|
info_data['config_h_features'] = {} |
||||||
|
|
||||||
|
if 'features' not in info_data: |
||||||
|
info_data['features'] = {} |
||||||
|
|
||||||
|
if key in info_data['features']: |
||||||
|
info_log_warning(info_data, 'Feature %s is specified in both info.json and rules.mk, the rules.mk value wins.' % (key,)) |
||||||
|
|
||||||
|
info_data['features'][key] = value |
||||||
|
info_data['config_h_features'][key] = value |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def _pin_name(pin): |
||||||
|
"""Returns the proper representation for a pin. |
||||||
|
""" |
||||||
|
pin = pin.strip() |
||||||
|
|
||||||
|
if not pin: |
||||||
|
return None |
||||||
|
|
||||||
|
elif pin.isdigit(): |
||||||
|
return int(pin) |
||||||
|
|
||||||
|
elif pin == 'NO_PIN': |
||||||
|
return None |
||||||
|
|
||||||
|
elif pin[0] in 'ABCDEFGHIJK' and pin[1].isdigit(): |
||||||
|
return pin |
||||||
|
|
||||||
|
raise ValueError(f'Invalid pin: {pin}') |
||||||
|
|
||||||
|
|
||||||
|
def _extract_pins(pins): |
||||||
|
"""Returns a list of pins from a comma separated string of pins. |
||||||
|
""" |
||||||
|
return [_pin_name(pin) for pin in pins.split(',')] |
||||||
|
|
||||||
|
|
||||||
|
def _extract_direct_matrix(info_data, direct_pins): |
||||||
|
""" |
||||||
|
""" |
||||||
|
info_data['matrix_pins'] = {} |
||||||
|
direct_pin_array = [] |
||||||
|
|
||||||
|
while direct_pins[-1] != '}': |
||||||
|
direct_pins = direct_pins[:-1] |
||||||
|
|
||||||
|
for row in direct_pins.split('},{'): |
||||||
|
if row.startswith('{'): |
||||||
|
row = row[1:] |
||||||
|
|
||||||
|
if row.endswith('}'): |
||||||
|
row = row[:-1] |
||||||
|
|
||||||
|
direct_pin_array.append([]) |
||||||
|
|
||||||
|
for pin in row.split(','): |
||||||
|
if pin == 'NO_PIN': |
||||||
|
pin = None |
||||||
|
|
||||||
|
direct_pin_array[-1].append(pin) |
||||||
|
|
||||||
|
return direct_pin_array |
||||||
|
|
||||||
|
|
||||||
|
def _extract_matrix_info(info_data, config_c): |
||||||
|
"""Populate the matrix information. |
||||||
|
""" |
||||||
|
row_pins = config_c.get('MATRIX_ROW_PINS', '').replace('{', '').replace('}', '').strip() |
||||||
|
col_pins = config_c.get('MATRIX_COL_PINS', '').replace('{', '').replace('}', '').strip() |
||||||
|
direct_pins = config_c.get('DIRECT_PINS', '').replace(' ', '')[1:-1] |
||||||
|
|
||||||
|
if 'MATRIX_ROWS' in config_c and 'MATRIX_COLS' in config_c: |
||||||
|
if 'matrix_size' in info_data: |
||||||
|
info_log_warning(info_data, 'Matrix size is specified in both info.json and config.h, the config.h values win.') |
||||||
|
|
||||||
|
info_data['matrix_size'] = { |
||||||
|
'cols': compute(config_c.get('MATRIX_COLS', '0')), |
||||||
|
'rows': compute(config_c.get('MATRIX_ROWS', '0')), |
||||||
|
} |
||||||
|
|
||||||
|
if row_pins and col_pins: |
||||||
|
if 'matrix_pins' in info_data: |
||||||
|
info_log_warning(info_data, 'Matrix pins are specified in both info.json and config.h, the config.h values win.') |
||||||
|
|
||||||
|
info_data['matrix_pins'] = { |
||||||
|
'cols': _extract_pins(col_pins), |
||||||
|
'rows': _extract_pins(row_pins), |
||||||
|
} |
||||||
|
|
||||||
|
if direct_pins: |
||||||
|
if 'matrix_pins' in info_data: |
||||||
|
info_log_warning(info_data, 'Direct pins are specified in both info.json and config.h, the config.h values win.') |
||||||
|
|
||||||
|
info_data['matrix_pins']['direct'] = _extract_direct_matrix(info_data, direct_pins) |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def _extract_config_h(info_data): |
||||||
|
"""Pull some keyboard information from existing config.h files |
||||||
|
""" |
||||||
|
config_c = config_h(info_data['keyboard_folder']) |
||||||
|
|
||||||
|
# Pull in data from the json map |
||||||
|
dotty_info = dotty(info_data) |
||||||
|
info_config_map = json_load(Path('data/mappings/info_config.json')) |
||||||
|
|
||||||
|
for config_key, info_dict in info_config_map.items(): |
||||||
|
info_key = info_dict['info_key'] |
||||||
|
key_type = info_dict.get('value_type', 'str') |
||||||
|
|
||||||
|
try: |
||||||
|
if config_key in config_c and info_dict.get('to_json', True): |
||||||
|
if dotty_info.get(info_key) and info_dict.get('warn_duplicate', True): |
||||||
|
info_log_warning(info_data, '%s in config.h is overwriting %s in info.json' % (config_key, info_key)) |
||||||
|
|
||||||
|
if key_type.startswith('array'): |
||||||
|
if '.' in key_type: |
||||||
|
key_type, array_type = key_type.split('.', 1) |
||||||
|
else: |
||||||
|
array_type = None |
||||||
|
|
||||||
|
config_value = config_c[config_key].replace('{', '').replace('}', '').strip() |
||||||
|
|
||||||
|
if array_type == 'int': |
||||||
|
dotty_info[info_key] = list(map(int, config_value.split(','))) |
||||||
|
else: |
||||||
|
dotty_info[info_key] = config_value.split(',') |
||||||
|
|
||||||
|
elif key_type == 'bool': |
||||||
|
dotty_info[info_key] = config_c[config_key] in true_values |
||||||
|
|
||||||
|
elif key_type == 'hex': |
||||||
|
dotty_info[info_key] = '0x' + config_c[config_key][2:].upper() |
||||||
|
|
||||||
|
elif key_type == 'list': |
||||||
|
dotty_info[info_key] = config_c[config_key].split() |
||||||
|
|
||||||
|
elif key_type == 'int': |
||||||
|
dotty_info[info_key] = int(config_c[config_key]) |
||||||
|
|
||||||
|
else: |
||||||
|
dotty_info[info_key] = config_c[config_key] |
||||||
|
|
||||||
|
except Exception as e: |
||||||
|
info_log_warning(info_data, f'{config_key}->{info_key}: {e}') |
||||||
|
|
||||||
|
info_data.update(dotty_info) |
||||||
|
|
||||||
|
# Pull data that easily can't be mapped in json |
||||||
|
_extract_matrix_info(info_data, config_c) |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def _extract_rules_mk(info_data): |
||||||
|
"""Pull some keyboard information from existing rules.mk files |
||||||
|
""" |
||||||
|
rules = rules_mk(info_data['keyboard_folder']) |
||||||
|
info_data['processor'] = rules.get('MCU', info_data.get('processor', 'atmega32u4')) |
||||||
|
|
||||||
|
if info_data['processor'] in CHIBIOS_PROCESSORS: |
||||||
|
arm_processor_rules(info_data, rules) |
||||||
|
|
||||||
|
elif info_data['processor'] in LUFA_PROCESSORS + VUSB_PROCESSORS: |
||||||
|
avr_processor_rules(info_data, rules) |
||||||
|
|
||||||
|
else: |
||||||
|
cli.log.warning("%s: Unknown MCU: %s" % (info_data['keyboard_folder'], info_data['processor'])) |
||||||
|
unknown_processor_rules(info_data, rules) |
||||||
|
|
||||||
|
# Pull in data from the json map |
||||||
|
dotty_info = dotty(info_data) |
||||||
|
info_rules_map = json_load(Path('data/mappings/info_rules.json')) |
||||||
|
|
||||||
|
for rules_key, info_dict in info_rules_map.items(): |
||||||
|
info_key = info_dict['info_key'] |
||||||
|
key_type = info_dict.get('value_type', 'str') |
||||||
|
|
||||||
|
try: |
||||||
|
if rules_key in rules and info_dict.get('to_json', True): |
||||||
|
if dotty_info.get(info_key) and info_dict.get('warn_duplicate', True): |
||||||
|
info_log_warning(info_data, '%s in rules.mk is overwriting %s in info.json' % (rules_key, info_key)) |
||||||
|
|
||||||
|
if key_type.startswith('array'): |
||||||
|
if '.' in key_type: |
||||||
|
key_type, array_type = key_type.split('.', 1) |
||||||
|
else: |
||||||
|
array_type = None |
||||||
|
|
||||||
|
rules_value = rules[rules_key].replace('{', '').replace('}', '').strip() |
||||||
|
|
||||||
|
if array_type == 'int': |
||||||
|
dotty_info[info_key] = list(map(int, rules_value.split(','))) |
||||||
|
else: |
||||||
|
dotty_info[info_key] = rules_value.split(',') |
||||||
|
|
||||||
|
elif key_type == 'list': |
||||||
|
dotty_info[info_key] = rules[rules_key].split() |
||||||
|
|
||||||
|
elif key_type == 'bool': |
||||||
|
dotty_info[info_key] = rules[rules_key] in true_values |
||||||
|
|
||||||
|
elif key_type == 'hex': |
||||||
|
dotty_info[info_key] = '0x' + rules[rules_key][2:].upper() |
||||||
|
|
||||||
|
elif key_type == 'int': |
||||||
|
dotty_info[info_key] = int(rules[rules_key]) |
||||||
|
|
||||||
|
else: |
||||||
|
dotty_info[info_key] = rules[rules_key] |
||||||
|
|
||||||
|
except Exception as e: |
||||||
|
info_log_warning(info_data, f'{rules_key}->{info_key}: {e}') |
||||||
|
|
||||||
|
info_data.update(dotty_info) |
||||||
|
|
||||||
|
# Merge in config values that can't be easily mapped |
||||||
|
_extract_features(info_data, rules) |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def _search_keyboard_h(path): |
||||||
|
current_path = Path('keyboards/') |
||||||
|
aliases = {} |
||||||
|
layouts = {} |
||||||
|
|
||||||
|
for directory in path.parts: |
||||||
|
current_path = current_path / directory |
||||||
|
keyboard_h = '%s.h' % (directory,) |
||||||
|
keyboard_h_path = current_path / keyboard_h |
||||||
|
if keyboard_h_path.exists(): |
||||||
|
new_layouts, new_aliases = find_layouts(keyboard_h_path) |
||||||
|
layouts.update(new_layouts) |
||||||
|
|
||||||
|
for alias, alias_text in new_aliases.items(): |
||||||
|
if alias_text in layouts: |
||||||
|
aliases[alias] = alias_text |
||||||
|
|
||||||
|
return layouts, aliases |
||||||
|
|
||||||
|
|
||||||
|
def _find_all_layouts(info_data, keyboard): |
||||||
|
"""Looks for layout macros associated with this keyboard. |
||||||
|
""" |
||||||
|
found_layouts = False |
||||||
|
layouts, aliases = _search_keyboard_h(Path(keyboard)) |
||||||
|
|
||||||
|
if layouts: |
||||||
|
found_layouts = True |
||||||
|
else: |
||||||
|
for layout in info_data['layouts'].values(): |
||||||
|
if 'matrix' in layout['layout']: |
||||||
|
found_layouts = True |
||||||
|
break |
||||||
|
|
||||||
|
if not found_layouts: |
||||||
|
# If we don't find any layouts from info.json or keyboard.h we widen our search. This is error prone which is why we want to encourage people to follow the standard. |
||||||
|
info_data['parse_warnings'].append('%s: Falling back to searching for KEYMAP/LAYOUT macros.' % (keyboard)) |
||||||
|
|
||||||
|
for file in glob('keyboards/%s/*.h' % keyboard): |
||||||
|
if file.endswith('.h'): |
||||||
|
these_layouts, these_aliases = find_layouts(file) |
||||||
|
|
||||||
|
if these_layouts: |
||||||
|
layouts.update(these_layouts) |
||||||
|
|
||||||
|
for alias, alias_text in these_aliases.items(): |
||||||
|
if alias_text in layouts: |
||||||
|
aliases[alias] = alias_text |
||||||
|
|
||||||
|
return layouts, aliases |
||||||
|
|
||||||
|
|
||||||
|
def info_log_error(info_data, message): |
||||||
|
"""Send an error message to both JSON and the log. |
||||||
|
""" |
||||||
|
info_data['parse_errors'].append(message) |
||||||
|
cli.log.error('%s: %s', info_data.get('keyboard_folder', 'Unknown Keyboard!'), message) |
||||||
|
|
||||||
|
|
||||||
|
def info_log_warning(info_data, message): |
||||||
|
"""Send a warning message to both JSON and the log. |
||||||
|
""" |
||||||
|
info_data['parse_warnings'].append(message) |
||||||
|
cli.log.warning('%s: %s', info_data.get('keyboard_folder', 'Unknown Keyboard!'), message) |
||||||
|
|
||||||
|
|
||||||
|
def arm_processor_rules(info_data, rules): |
||||||
|
"""Setup the default info for an ARM board. |
||||||
|
""" |
||||||
|
info_data['processor_type'] = 'arm' |
||||||
|
info_data['protocol'] = 'ChibiOS' |
||||||
|
|
||||||
|
if 'bootloader' not in info_data: |
||||||
|
if 'STM32' in info_data['processor']: |
||||||
|
info_data['bootloader'] = 'stm32-dfu' |
||||||
|
else: |
||||||
|
info_data['bootloader'] = 'unknown' |
||||||
|
|
||||||
|
if 'STM32' in info_data['processor']: |
||||||
|
info_data['platform'] = 'STM32' |
||||||
|
elif 'MCU_SERIES' in rules: |
||||||
|
info_data['platform'] = rules['MCU_SERIES'] |
||||||
|
elif 'ARM_ATSAM' in rules: |
||||||
|
info_data['platform'] = 'ARM_ATSAM' |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def avr_processor_rules(info_data, rules): |
||||||
|
"""Setup the default info for an AVR board. |
||||||
|
""" |
||||||
|
info_data['processor_type'] = 'avr' |
||||||
|
info_data['platform'] = rules['ARCH'] if 'ARCH' in rules else 'unknown' |
||||||
|
info_data['protocol'] = 'V-USB' if rules.get('MCU') in VUSB_PROCESSORS else 'LUFA' |
||||||
|
|
||||||
|
if 'bootloader' not in info_data: |
||||||
|
info_data['bootloader'] = 'atmel-dfu' |
||||||
|
|
||||||
|
# FIXME(fauxpark/anyone): Eventually we should detect the protocol by looking at PROTOCOL inherited from mcu_selection.mk: |
||||||
|
# info_data['protocol'] = 'V-USB' if rules.get('PROTOCOL') == 'VUSB' else 'LUFA' |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def unknown_processor_rules(info_data, rules): |
||||||
|
"""Setup the default keyboard info for unknown boards. |
||||||
|
""" |
||||||
|
info_data['bootloader'] = 'unknown' |
||||||
|
info_data['platform'] = 'unknown' |
||||||
|
info_data['processor'] = 'unknown' |
||||||
|
info_data['processor_type'] = 'unknown' |
||||||
|
info_data['protocol'] = 'unknown' |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def merge_info_jsons(keyboard, info_data): |
||||||
|
"""Return a merged copy of all the info.json files for a keyboard. |
||||||
|
""" |
||||||
|
for info_file in find_info_json(keyboard): |
||||||
|
# Load and validate the JSON data |
||||||
|
new_info_data = json_load(info_file) |
||||||
|
|
||||||
|
if not isinstance(new_info_data, dict): |
||||||
|
info_log_error(info_data, "Invalid file %s, root object should be a dictionary." % (str(info_file),)) |
||||||
|
continue |
||||||
|
|
||||||
|
try: |
||||||
|
keyboard_validate(new_info_data) |
||||||
|
except jsonschema.ValidationError as e: |
||||||
|
json_path = '.'.join([str(p) for p in e.absolute_path]) |
||||||
|
cli.log.error('Not including data from file: %s', info_file) |
||||||
|
cli.log.error('\t%s: %s', json_path, e.message) |
||||||
|
continue |
||||||
|
|
||||||
|
# Merge layout data in |
||||||
|
if 'layout_aliases' in new_info_data: |
||||||
|
info_data['layout_aliases'] = {**info_data.get('layout_aliases', {}), **new_info_data['layout_aliases']} |
||||||
|
del new_info_data['layout_aliases'] |
||||||
|
|
||||||
|
for layout_name, layout in new_info_data.get('layouts', {}).items(): |
||||||
|
if layout_name in info_data.get('layout_aliases', {}): |
||||||
|
info_log_warning(info_data, f"info.json uses alias name {layout_name} instead of {info_data['layout_aliases'][layout_name]}") |
||||||
|
layout_name = info_data['layout_aliases'][layout_name] |
||||||
|
|
||||||
|
if layout_name in info_data['layouts']: |
||||||
|
for new_key, existing_key in zip(layout['layout'], info_data['layouts'][layout_name]['layout']): |
||||||
|
existing_key.update(new_key) |
||||||
|
else: |
||||||
|
layout['c_macro'] = False |
||||||
|
info_data['layouts'][layout_name] = layout |
||||||
|
|
||||||
|
# Update info_data with the new data |
||||||
|
if 'layouts' in new_info_data: |
||||||
|
del new_info_data['layouts'] |
||||||
|
|
||||||
|
deep_update(info_data, new_info_data) |
||||||
|
|
||||||
|
return info_data |
||||||
|
|
||||||
|
|
||||||
|
def find_info_json(keyboard): |
||||||
|
"""Finds all the info.json files associated with a keyboard. |
||||||
|
""" |
||||||
|
# Find the most specific first |
||||||
|
base_path = Path('keyboards') |
||||||
|
keyboard_path = base_path / keyboard |
||||||
|
keyboard_parent = keyboard_path.parent |
||||||
|
info_jsons = [keyboard_path / 'info.json'] |
||||||
|
|
||||||
|
# Add DEFAULT_FOLDER before parents, if present |
||||||
|
rules = rules_mk(keyboard) |
||||||
|
if 'DEFAULT_FOLDER' in rules: |
||||||
|
info_jsons.append(Path(rules['DEFAULT_FOLDER']) / 'info.json') |
||||||
|
|
||||||
|
# Add in parent folders for least specific |
||||||
|
for _ in range(5): |
||||||
|
info_jsons.append(keyboard_parent / 'info.json') |
||||||
|
if keyboard_parent.parent == base_path: |
||||||
|
break |
||||||
|
keyboard_parent = keyboard_parent.parent |
||||||
|
|
||||||
|
# Return a list of the info.json files that actually exist |
||||||
|
return [info_json for info_json in info_jsons if info_json.exists()] |
Loading…
Reference in new issue