#! /usr/bin/env python3 import random import re import os import getopt import sys from os import path ENCODING = "utf-8" SOURCE_FILE_PATTERN = r"\b\d+_error\b" def read_file(file_name): content = None _, tail = path.split(file_name) is_latin = tail == "invalid_utf8_sequence.sol" try: with open(file_name, "r", encoding="latin-1" if is_latin else ENCODING) as f: content = f.read() finally: if content == None: print(f"Error reading: {file_name}") return content def write_file(file_name, content): with open(file_name, "w", encoding=ENCODING) as f: f.write(content) def in_comment(source, pos): slash_slash_pos = source.rfind("//", 0, pos) lf_pos = source.rfind("\n", 0, pos) if slash_slash_pos > lf_pos: return True slash_star_pos = source.rfind("/*", 0, pos) star_slash_pos = source.rfind("*/", 0, pos) return slash_star_pos > star_slash_pos def find_ids_in_source_file(file_name, id_to_file_names): source = read_file(file_name) for m in re.finditer(SOURCE_FILE_PATTERN, source): if in_comment(source, m.start()): continue underscore_pos = m.group(0).index("_") id = m.group(0)[0:underscore_pos] if id in id_to_file_names: id_to_file_names[id].append(file_name) else: id_to_file_names[id] = [file_name] def find_ids_in_source_files(file_names): """Returns a dictionary with list of source files for every appearance of every id""" id_to_file_names = {} for file_name in file_names: find_ids_in_source_file(file_name, id_to_file_names) return id_to_file_names def get_next_id(available_ids): assert len(available_ids) > 0, "Out of IDs" next_id = random.choice(list(available_ids)) available_ids.remove(next_id) return next_id def fix_ids_in_source_file(file_name, id_to_count, available_ids): source = read_file(file_name) k = 0 destination = [] for m in re.finditer(SOURCE_FILE_PATTERN, source): destination.extend(source[k:m.start()]) underscore_pos = m.group(0).index("_") id = m.group(0)[0:underscore_pos] # incorrect id or id has a duplicate somewhere if not in_comment(source, m.start()) and (len(id) != 4 or id[0] == "0" or id_to_count[id] > 1): assert id in id_to_count new_id = get_next_id(available_ids) assert new_id not in id_to_count id_to_count[id] -= 1 else: new_id = id destination.extend(new_id + "_error") k = m.end() destination.extend(source[k:]) destination = ''.join(destination) if source != destination: write_file(file_name, destination) print(f"Fixed file: {file_name}") def fix_ids_in_source_files(file_names, id_to_count): """ Fixes ids in given source files; id_to_count contains number of appearances of every id in sources """ available_ids = {str(id) for id in range(1000, 10000)} - id_to_count.keys() for file_name in file_names: fix_ids_in_source_file(file_name, id_to_count, available_ids) def find_files(top_dir, sub_dirs, extensions): """Builds a list of files with given extensions in specified subdirectories""" source_file_names = [] for dir in sub_dirs: for root, _, file_names in os.walk(os.path.join(top_dir, dir), onerror=lambda e: exit(f"Walk error: {e}")): for file_name in file_names: _, ext = path.splitext(file_name) if ext in extensions: source_file_names.append(path.join(root, file_name)) return source_file_names def find_ids_in_test_file(file_name): source = read_file(file_name) pattern = r"^// (.*Error|Warning) \d\d\d\d:" return {m.group(0)[-5:-1] for m in re.finditer(pattern, source, flags=re.MULTILINE)} def find_ids_in_test_files(file_names): """Returns a set containing all ids in tests""" ids = set() for file_name in file_names: ids |= find_ids_in_test_file(file_name) return ids def find_ids_in_cmdline_test_err(file_name): source = read_file(file_name) pattern = r' \(\d\d\d\d\):' return {m.group(0)[-6:-2] for m in re.finditer(pattern, source, flags=re.MULTILINE)} def print_ids(ids): for k, id in enumerate(sorted(ids)): if k % 10 > 0: print(" ", end="") elif k > 0: print() print(id, end="") def print_ids_per_file(ids, id_to_file_names, top_dir): file_name_to_ids = {} for id in ids: for file_name in id_to_file_names[id]: relpath = path.relpath(file_name, top_dir) if relpath not in file_name_to_ids: file_name_to_ids[relpath] = [] file_name_to_ids[relpath].append(id) for file_name in sorted(file_name_to_ids): print(file_name) for id in sorted(file_name_to_ids[file_name]): print(f" {id}", end="") print() def examine_id_coverage(top_dir, source_id_to_file_names, new_ids_only=False): test_sub_dirs = [ path.join("test", "libsolidity", "errorRecoveryTests"), path.join("test", "libsolidity", "smtCheckerTests"), path.join("test", "libsolidity", "syntaxTests"), path.join("test", "libyul", "yulSyntaxTests") ] test_file_names = find_files( top_dir, test_sub_dirs, [".sol", ".yul"] ) source_ids = source_id_to_file_names.keys() test_ids = find_ids_in_test_files(test_file_names) # special case, we are interested in warnings which are ignored by regular tests: # Warning (1878): SPDX license identifier not provided in source file. .... # Warning (3420): Source file does not specify required compiler version! test_ids |= find_ids_in_cmdline_test_err(path.join(top_dir, "test", "cmdlineTests", "error_codes", "err")) test_ids |= find_ids_in_cmdline_test_err(path.join(top_dir, "test", "cmdlineTests", "yul_unimplemented", "err")) # white list of ids which are not covered by tests white_ids = { "3805", # "This is a pre-release compiler version, please do not use it in production." # The warning may or may not exist in a compiler build. "4591" # "There are more than 256 warnings. Ignoring the rest." # Due to 3805, the warning lists look different for different compiler builds. } assert len(test_ids & white_ids) == 0, "The sets are not supposed to intersect" test_ids |= white_ids test_only_ids = test_ids - source_ids source_only_ids = source_ids - test_ids if not new_ids_only: print(f"IDs in source files: {len(source_ids)}") print(f"IDs in test files : {len(test_ids)} ({len(test_ids) - len(source_ids)})") print() if len(test_only_ids) != 0: print("Error. The following error codes found in tests, but not in sources:") print_ids(test_only_ids) return False if len(source_only_ids) != 0: print("The following error codes found in sources, but not in tests:") print_ids_per_file(source_only_ids, source_id_to_file_names, top_dir) print("\n\nPlease make sure to add appropriate tests.") return False old_source_only_ids = { "1123", "1133", "1220", "1584", "1823", "1950", "1988", "2418", "2461", "2512", "2592", "2657", "2800", "2842", "2856", "3263", "3356", "3441", "3682", "3876", "3893", "3997", "4010", "4802", "4805", "4828", "4904", "4990", "5052", "5073", "5170", "5188", "5272", "5333", "5347", "5473", "5622", "6041", "6052", "6272", "6708", "6792", "6931", "7110", "7128", "7186", "7319", "7589", "7593", "7653", "7812", "7885", "8065", "8084", "8140", "8261", "8312", "8592", "8758", "9011", "9085", "9390", "9440", "9547", "9551", "9615", "9980" } new_source_only_ids = source_only_ids - old_source_only_ids if len(new_source_only_ids) != 0: print("The following new error code(s), not covered by tests, found:") print_ids(new_source_only_ids) print( "\nYou can:\n" "- create appropriate test(s);\n" "- add the error code(s) to old_source_only_ids in error_codes.py\n" " (to silence the checking script, with a promise to add a test later);\n" "- add the error code(s) to white_ids in error_codes.py\n" " (for rare cases when the error is not supposed to be tested)" ) return False return True def main(argv): # pylint: disable=too-many-branches, too-many-locals, too-many-statements check = False fix = False no_confirm = False examine_coverage = False next_id = False opts, args = getopt.getopt(argv, "", ["check", "fix", "no-confirm", "examine-coverage", "next"]) for opt, arg in opts: if opt == "--check": check = True elif opt == "--fix": fix = True elif opt == "--no-confirm": no_confirm = True elif opt == "--examine-coverage": examine_coverage = True elif opt == "--next": next_id = True if [check, fix, examine_coverage, next_id].count(True) != 1: print("usage: python error_codes.py --check | --fix [--no-confirm] | --examine-coverage | --next") exit(1) cwd = os.getcwd() source_file_names = find_files( cwd, ["libevmasm", "liblangutil", "libsolc", "libsolidity", "libsolutil", "libyul", "solc"], [".h", ".cpp"] ) source_id_to_file_names = find_ids_in_source_files(source_file_names) ok = True for id in sorted(source_id_to_file_names): if len(id) != 4: print(f"ID {id} length != 4") ok = False if id[0] == "0": print(f"ID {id} starts with zero") ok = False if len(source_id_to_file_names[id]) > 1: print(f"ID {id} appears {len(source_id_to_file_names[id])} times") ok = False if examine_coverage: if not ok: print("Incorrect IDs have to be fixed before applying --examine-coverage") exit(1) res = 0 if examine_id_coverage(cwd, source_id_to_file_names) else 1 exit(res) ok &= examine_id_coverage(cwd, source_id_to_file_names, new_ids_only=True) random.seed() if next_id: if not ok: print("Incorrect IDs have to be fixed before applying --next") exit(1) available_ids = {str(id) for id in range(1000, 10000)} - source_id_to_file_names.keys() next_id = get_next_id(available_ids) print(f"Next ID: {next_id}") exit(0) if ok: print("No incorrect IDs found") exit(0) if check: exit(1) assert fix, "Unexpected state, should not come here without --fix" if not no_confirm: answer = input( "\nDo you want to fix incorrect IDs?\n" "Please commit current changes first, and review the results when the script finishes.\n" "[Y/N]? " ) while len(answer) == 0 or answer not in "YNyn": answer = input("[Y/N]? ") if answer not in "yY": exit(1) # number of appearances for every id source_id_to_count = { id: len(file_names) for id, file_names in source_id_to_file_names.items() } fix_ids_in_source_files(source_file_names, source_id_to_count) print("Fixing completed") exit(2) if __name__ == "__main__": main(sys.argv[1:])