#!/usr/bin/env python3 """ Command to print info about makefiles remaining to be converted to soong. See usage / argument parsing below for commandline options. """ import argparse import csv import itertools import json import os import re import sys DIRECTORY_PATTERNS = [x.split("/") for x in ( "device/*", "frameworks/*", "hardware/*", "packages/*", "vendor/*", "*", )] def match_directory_group(pattern, filename): match = [] filename = filename.split("/") if len(filename) < len(pattern): return None for i in range(len(pattern)): pattern_segment = pattern[i] filename_segment = filename[i] if pattern_segment == "*" or pattern_segment == filename_segment: match.append(filename_segment) else: return None if match: return os.path.sep.join(match) else: return None def directory_group(filename): for pattern in DIRECTORY_PATTERNS: match = match_directory_group(pattern, filename) if match: return match return os.path.dirname(filename) class Analysis(object): def __init__(self, filename, line_matches): self.filename = filename; self.line_matches = line_matches def analyze_lines(filename, lines, func): line_matches = [] for i in range(len(lines)): line = lines[i] stripped = line.strip() if stripped.startswith("#"): continue if func(stripped): line_matches.append((i+1, line)) if line_matches: return Analysis(filename, line_matches); def analyze_has_conditional(line): return (line.startswith("ifeq") or line.startswith("ifneq") or line.startswith("ifdef") or line.startswith("ifndef")) NORMAL_INCLUDES = [re.compile(pattern) for pattern in ( "include \$+\(CLEAR_VARS\)", # These are in defines which are tagged separately "include \$+\(BUILD_.*\)", "include \$\(call first-makefiles-under, *\$\(LOCAL_PATH\)\)", "include \$\(call all-subdir-makefiles\)", "include \$\(all-subdir-makefiles\)", "include \$\(call all-makefiles-under, *\$\(LOCAL_PATH\)\)", "include \$\(call all-makefiles-under, *\$\(call my-dir\).*\)", "include \$\(BUILD_SYSTEM\)/base_rules.mk", # called out separately "include \$\(call all-named-subdir-makefiles,.*\)", "include \$\(subdirs\)", )] def analyze_has_wacky_include(line): if not (line.startswith("include") or line.startswith("-include") or line.startswith("sinclude")): return False for matcher in NORMAL_INCLUDES: if matcher.fullmatch(line): return False return True BASE_RULES_RE = re.compile("include \$\(BUILD_SYSTEM\)/base_rules.mk") class Analyzer(object): def __init__(self, title, func): self.title = title; self.func = func ANALYZERS = ( Analyzer("ifeq / ifneq", analyze_has_conditional), Analyzer("Wacky Includes", analyze_has_wacky_include), Analyzer("Calls base_rules", lambda line: BASE_RULES_RE.fullmatch(line)), Analyzer("Calls define", lambda line: line.startswith("define ")), Analyzer("Has ../", lambda line: "../" in line), Analyzer("dist-for-​goals", lambda line: "dist-for-goals" in line), Analyzer(".PHONY", lambda line: ".PHONY" in line), Analyzer("render-​script", lambda line: ".rscript" in line), Analyzer("vts src", lambda line: ".vts" in line), Analyzer("COPY_​HEADERS", lambda line: "LOCAL_COPY_HEADERS" in line), ) class Summary(object): def __init__(self): self.makefiles = dict() self.directories = dict() def Add(self, makefile): self.makefiles[makefile.filename] = makefile self.directories.setdefault(directory_group(makefile.filename), []).append(makefile) class Makefile(object): def __init__(self, filename): self.filename = filename # Analyze the file with open(filename, "r", errors="ignore") as f: try: lines = f.readlines() except UnicodeDecodeError as ex: sys.stderr.write("Filename: %s\n" % filename) raise ex lines = [line.strip() for line in lines] self.analyses = dict([(analyzer, analyze_lines(filename, lines, analyzer.func)) for analyzer in ANALYZERS]) def find_android_mk(): cwd = os.getcwd() for root, dirs, files in os.walk(cwd): for filename in files: if filename == "Android.mk": yield os.path.join(root, filename)[len(cwd) + 1:] for ignore in (".git", ".repo"): if ignore in dirs: dirs.remove(ignore) def is_aosp(dirname): for d in ("device/sample", "hardware/interfaces", "hardware/libhardware", "hardware/ril"): if dirname.startswith(d): return True for d in ("device/", "hardware/", "vendor/"): if dirname.startswith(d): return False return True def is_google(dirname): for d in ("device/google", "hardware/google", "test/sts", "vendor/auto", "vendor/google", "vendor/unbundled_google", "vendor/widevine", "vendor/xts"): if dirname.startswith(d): return True return False def is_clean(makefile): for analysis in makefile.analyses.values(): if analysis: return False return True def clean_and_only_blocked_by_clean(soong, all_makefiles, makefile): if not is_clean(makefile): return False modules = soong.reverse_makefiles[makefile.filename] for module in modules: for dep in soong.transitive_deps(module): for filename in soong.makefiles.get(dep, []): m = all_makefiles.get(filename) if m and not is_clean(m): return False return True class Annotations(object): def __init__(self): self.entries = [] self.count = 0 def Add(self, makefiles, modules): self.entries.append((makefiles, modules)) self.count += 1 return self.count-1 class SoongData(object): def __init__(self, reader): """Read the input file and store the modules and dependency mappings. """ self.problems = dict() self.deps = dict() self.reverse_deps = dict() self.module_types = dict() self.makefiles = dict() self.reverse_makefiles = dict() self.installed = dict() self.reverse_installed = dict() self.modules = set() for (module, module_type, problem, dependencies, makefiles, installed) in reader: self.modules.add(module) makefiles = [f for f in makefiles.strip().split(' ') if f != ""] self.module_types[module] = module_type self.problems[module] = problem self.deps[module] = [d for d in dependencies.strip().split(' ') if d != ""] for dep in self.deps[module]: if not dep in self.reverse_deps: self.reverse_deps[dep] = [] self.reverse_deps[dep].append(module) self.makefiles[module] = makefiles for f in makefiles: self.reverse_makefiles.setdefault(f, []).append(module) for f in installed.strip().split(' '): self.installed[f] = module self.reverse_installed.setdefault(module, []).append(f) def transitive_deps(self, module): results = set() def traverse(module): for dep in self.deps.get(module, []): if not dep in results: results.add(dep) traverse(module) traverse(module) return results def contains_unblocked_modules(self, filename): for m in self.reverse_makefiles[filename]: if len(self.deps[m]) == 0: return True return False def contains_blocked_modules(self, filename): for m in self.reverse_makefiles[filename]: if len(self.deps[m]) > 0: return True return False def count_deps(depsdb, module, seen): """Based on the depsdb, count the number of transitive dependencies. You can pass in an reversed dependency graph to count the number of modules that depend on the module.""" count = 0 seen.append(module) if module in depsdb: for dep in depsdb[module]: if dep in seen: continue count += 1 + count_deps(depsdb, dep, seen) return count OTHER_PARTITON = "_other" HOST_PARTITON = "_host" def get_partition_from_installed(HOST_OUT_ROOT, PRODUCT_OUT, filename): host_prefix = HOST_OUT_ROOT + "/" device_prefix = PRODUCT_OUT + "/" if filename.startswith(host_prefix): return HOST_PARTITON elif filename.startswith(device_prefix): index = filename.find("/", len(device_prefix)) if index < 0: return OTHER_PARTITON return filename[len(device_prefix):index] return OTHER_PARTITON def format_module_link(module): return "%s" % (module, module) def format_module_list(modules): return "".join(["
%s
" % format_module_link(m) for m in modules]) def print_analysis_header(link, title): print("""

%(title)s

""" % { "link": link, "title": title }) for analyzer in ANALYZERS: print("""""" % analyzer.title) print(" ") def main(): parser = argparse.ArgumentParser(description="Info about remaining Android.mk files.") parser.add_argument("--device", type=str, required=True, help="TARGET_DEVICE") parser.add_argument("--title", type=str, help="page title") parser.add_argument("--codesearch", type=str, default="https://cs.android.com/android/platform/superproject/+/master:", help="page title") parser.add_argument("--out_dir", type=str, default=None, help="Equivalent of $OUT_DIR, which will also be checked if" + " --out_dir is unset. If neither is set, default is" + " 'out'.") parser.add_argument("--mode", type=str, default="html", help="output format: csv or html") args = parser.parse_args() # Guess out directory name if not args.out_dir: args.out_dir = os.getenv("OUT_DIR", "out") while args.out_dir.endswith("/") and len(args.out_dir) > 1: args.out_dir = args.out_dir[:-1] TARGET_DEVICE = args.device global HOST_OUT_ROOT HOST_OUT_ROOT = args.out_dir + "/host" global PRODUCT_OUT PRODUCT_OUT = args.out_dir + "/target/product/%s" % TARGET_DEVICE # Read target information # TODO: Pull from configurable location. This is also slightly different because it's # only a single build, where as the tree scanning we do below is all Android.mk files. with open("%s/obj/PACKAGING/soong_conversion_intermediates/soong_conv_data" % PRODUCT_OUT, "r", errors="ignore") as csvfile: soong = SoongData(csv.reader(csvfile)) # Read the makefiles all_makefiles = dict() for filename, modules in soong.reverse_makefiles.items(): if filename.startswith(args.out_dir + "/"): continue all_makefiles[filename] = Makefile(filename) if args.mode == "html": HtmlProcessor(args=args, soong=soong, all_makefiles=all_makefiles).execute() elif args.mode == "csv": CsvProcessor(args=args, soong=soong, all_makefiles=all_makefiles).execute() class HtmlProcessor(object): def __init__(self, args, soong, all_makefiles): self.args = args self.soong = soong self.all_makefiles = all_makefiles self.annotations = Annotations() def execute(self): if self.args.title: page_title = self.args.title else: page_title = "Remaining Android.mk files" # Which modules are installed where modules_by_partition = dict() partitions = set() for installed, module in self.soong.installed.items(): partition = get_partition_from_installed(HOST_OUT_ROOT, PRODUCT_OUT, installed) modules_by_partition.setdefault(partition, []).append(module) partitions.add(partition) print(""" %(page_title)s

%(page_title)s

This page analyzes the remaining Android.mk files in the Android Source tree.

The modules are first broken down by which of the device filesystem partitions they are installed to. This also includes host tools and testcases which don't actually reside in their own partition but convenitely group together.

The makefiles for each partition are further are grouped into a set of directories aritrarily picked to break down the problem size by owners.

  • AOSP directories are colored green.
  • Google directories are colored blue.
  • Other partner directories are colored red.
Each of the makefiles are scanned for issues that are likely to come up during conversion to soong. Clicking the number in each cell shows additional information, including the line that triggered the warning.

Directory Total Easy Unblocked Clean Unblocked Blocked Clean%s
Total The total number of makefiles in this each directory.
Easy The number of makefiles that have no warnings themselves, and also none of their dependencies have warnings either.
Unblocked Clean The number of makefiles that are both Unblocked and Clean.
Unblocked Makefiles containing one or more modules that don't have any additional dependencies pending before conversion.
Blocked Makefiles containiong one or more modules which do have additional prerequesite depenedencies that are not yet converted.
Clean The number of makefiles that have none of the following warnings.
ifeq / ifneq Makefiles that use ifeq or ifneq. i.e. conditionals.
Wacky Includes Makefiles that include files other than the standard build-system defined template and macros.
Calls base_rules Makefiles that include base_rules.mk directly.
Calls define Makefiles that define their own macros. Some of these are easy to convert to soong defaults, but others are complex.
Has ../ Makefiles containing the string "../" outside of a comment. These likely access files outside their directories.
dist-for-goals Makefiles that call dist-for-goals directly.
.PHONY Makefiles that declare .PHONY targets.
renderscript Makefiles defining targets that depend on .rscript source files.
vts src Makefiles defining targets that depend on .vts source files.
COPY_HEADERS Makefiles using LOCAL_COPY_HEADERS.

Following the list of directories is a list of the modules that are installed on each partition. Potential issues from their makefiles are listed, as well as the total number of dependencies (both blocking that module and blocked by that module) and the list of direct dependencies. Note: The number is the number of all transitive dependencies and the list of modules is only the direct dependencies. """) overall_summary = Summary() # For each partition for partition in sorted(partitions): modules = modules_by_partition[partition] makefiles = set(itertools.chain.from_iterable( [self.soong.makefiles[module] for module in modules])) # Read makefiles summary = Summary() for filename in makefiles: makefile = self.all_makefiles.get(filename) if makefile: summary.Add(makefile) overall_summary.Add(makefile) # Categorize directories by who is responsible aosp_dirs = [] google_dirs = [] partner_dirs = [] for dirname in sorted(summary.directories.keys()): if is_aosp(dirname): aosp_dirs.append(dirname) elif is_google(dirname): google_dirs.append(dirname) else: partner_dirs.append(dirname) print_analysis_header("partition_" + partition, partition) for dirgroup, rowclass in [(aosp_dirs, "AospDir"), (google_dirs, "GoogleDir"), (partner_dirs, "PartnerDir"),]: for dirname in dirgroup: self.print_analysis_row(summary, modules, dirname, rowclass, summary.directories[dirname]) self.print_analysis_row(summary, modules, "Total", "TotalRow", set(itertools.chain.from_iterable(summary.directories.values()))) print(""" """) module_details = [(count_deps(self.soong.deps, m, []), -count_deps(self.soong.reverse_deps, m, []), m) for m in modules] module_details.sort() module_details = [m[2] for m in module_details] print(""" """) print("") print(" ") print(" ") print(" ") print(" ") print("") altRow = True for module in module_details: analyses = set() for filename in self.soong.makefiles[module]: makefile = summary.makefiles.get(filename) if makefile: for analyzer, analysis in makefile.analyses.items(): if analysis: analyses.add(analyzer.title) altRow = not altRow print("" % ("Alt" if altRow else "",)) print(" " % (module, module)) print(" " % " ".join(["%s" % title for title in analyses])) print(" " % count_deps(self.soong.deps, module, [])) print(" " % format_module_list(self.soong.deps.get(module, []))) print(" " % count_deps(self.soong.reverse_deps, module, [])) print(" " % format_module_list(self.soong.reverse_deps.get(module, []))) print("") print("""
Module NameIssuesBlocked ByBlocking
%s%s%s%s%s%s
""") print_analysis_header("summary", "Overall Summary") modules = [module for installed, module in self.soong.installed.items()] self.print_analysis_row(overall_summary, modules, "All Makefiles", "TotalRow", set(itertools.chain.from_iterable(overall_summary.directories.values()))) print(""" """) print(""" """) print("""

""") def traverse_ready_makefiles(self, summary, makefiles): return [Analysis(makefile.filename, []) for makefile in makefiles if clean_and_only_blocked_by_clean(self.soong, self.all_makefiles, makefile)] def print_analysis_row(self, summary, modules, rowtitle, rowclass, makefiles): all_makefiles = [Analysis(makefile.filename, []) for makefile in makefiles] clean_makefiles = [Analysis(makefile.filename, []) for makefile in makefiles if is_clean(makefile)] easy_makefiles = self.traverse_ready_makefiles(summary, makefiles) unblocked_clean_makefiles = [Analysis(makefile.filename, []) for makefile in makefiles if (self.soong.contains_unblocked_modules(makefile.filename) and is_clean(makefile))] unblocked_makefiles = [Analysis(makefile.filename, []) for makefile in makefiles if self.soong.contains_unblocked_modules(makefile.filename)] blocked_makefiles = [Analysis(makefile.filename, []) for makefile in makefiles if self.soong.contains_blocked_modules(makefile.filename)] print(""" %(rowtitle)s %(makefiles)s %(easy)s %(unblocked_clean)s %(unblocked)s %(blocked)s %(clean)s """ % { "rowclass": rowclass, "rowtitle": rowtitle, "makefiles": self.make_annotation_link(all_makefiles, modules), "unblocked": self.make_annotation_link(unblocked_makefiles, modules), "blocked": self.make_annotation_link(blocked_makefiles, modules), "clean": self.make_annotation_link(clean_makefiles, modules), "unblocked_clean": self.make_annotation_link(unblocked_clean_makefiles, modules), "easy": self.make_annotation_link(easy_makefiles, modules), }) for analyzer in ANALYZERS: analyses = [m.analyses.get(analyzer) for m in makefiles if m.analyses.get(analyzer)] print("""%s""" % self.make_annotation_link(analyses, modules)) print(" ") def make_annotation_link(self, analysis, modules): if analysis: return "%s" % ( self.annotations.Add(analysis, modules), len(analysis) ) else: return ""; class CsvProcessor(object): def __init__(self, args, soong, all_makefiles): self.args = args self.soong = soong self.all_makefiles = all_makefiles def execute(self): csvout = csv.writer(sys.stdout) # Title row row = ["Filename", "Module", "Partitions", "Easy", "Unblocked Clean", "Unblocked", "Blocked", "Clean"] for analyzer in ANALYZERS: row.append(analyzer.title) csvout.writerow(row) # Makefile & module data for filename in sorted(self.all_makefiles.keys()): makefile = self.all_makefiles[filename] for module in self.soong.reverse_makefiles[filename]: row = [filename, module] # Partitions row.append(";".join(sorted(set([get_partition_from_installed(HOST_OUT_ROOT, PRODUCT_OUT, installed) for installed in self.soong.reverse_installed.get(module, [])])))) # Easy row.append(1 if clean_and_only_blocked_by_clean(self.soong, self.all_makefiles, makefile) else "") # Unblocked Clean row.append(1 if (self.soong.contains_unblocked_modules(makefile.filename) and is_clean(makefile)) else "") # Unblocked row.append(1 if self.soong.contains_unblocked_modules(makefile.filename) else "") # Blocked row.append(1 if self.soong.contains_blocked_modules(makefile.filename) else "") # Clean row.append(1 if is_clean(makefile) else "") # Analysis for analyzer in ANALYZERS: row.append(1 if makefile.analyses.get(analyzer) else "") # Write results csvout.writerow(row) if __name__ == "__main__": main()