# Copyright (C) 2019 The Android Open Source Project # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # A collection of utilities for extracting build rule information from GN # projects. from __future__ import print_function import collections from compat import iteritems import errno import filecmp import json import os import re import shutil import subprocess import sys from typing import Dict from typing import Optional from typing import Set from typing import Tuple BUILDFLAGS_TARGET = '//gn:gen_buildflags' GEN_VERSION_TARGET = '//src/base:version_gen_h' TARGET_TOOLCHAIN = '//gn/standalone/toolchain:gcc_like_host' HOST_TOOLCHAIN = '//gn/standalone/toolchain:gcc_like_host' LINKER_UNIT_TYPES = ('executable', 'shared_library', 'static_library') # TODO(primiano): investigate these, they require further componentization. ODR_VIOLATION_IGNORE_TARGETS = { '//test/cts:perfetto_cts_deps', '//:perfetto_integrationtests', } def _check_command_output(cmd, cwd): try: output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, cwd=cwd) except subprocess.CalledProcessError as e: print( 'Command "{}" failed in {}:'.format(' '.join(cmd), cwd), file=sys.stderr) print(e.output.decode(), file=sys.stderr) sys.exit(1) else: return output.decode() def repo_root(): """Returns an absolute path to the repository root.""" return os.path.join( os.path.realpath(os.path.dirname(__file__)), os.path.pardir) def _tool_path(name, system_buildtools=False): # Pass-through to use name if the caller requests to use the system # toolchain. if system_buildtools: return [name] wrapper = os.path.abspath( os.path.join(repo_root(), 'tools', 'run_buildtools_binary.py')) return ['python3', wrapper, name] def prepare_out_directory(gn_args, name, root=repo_root(), system_buildtools=False): """Creates the JSON build description by running GN. Returns (path, desc) where |path| is the location of the output directory and |desc| is the JSON build description. """ out = os.path.join(root, 'out', name) try: os.makedirs(out) except OSError as e: if e.errno != errno.EEXIST: raise _check_command_output( _tool_path('gn', system_buildtools) + ['gen', out, '--args=%s' % gn_args], cwd=repo_root()) return out def load_build_description(out, system_buildtools=False): """Creates the JSON build description by running GN.""" desc = _check_command_output( _tool_path('gn', system_buildtools) + ['desc', out, '--format=json', '--all-toolchains', '//*'], cwd=repo_root()) return json.loads(desc) def create_build_description(gn_args, root=repo_root()): """Prepares a GN out directory and loads the build description from it. The temporary out directory is automatically deleted. """ out = prepare_out_directory(gn_args, 'tmp.gn_utils', root=root) try: return load_build_description(out) finally: shutil.rmtree(out) def build_targets(out, targets, quiet=False, system_buildtools=False): """Runs ninja to build a list of GN targets in the given out directory. Compiling these targets is required so that we can include any generated source files in the amalgamated result. """ targets = [t.replace('//', '') for t in targets] with open(os.devnull, 'w', newline='\n') as devnull: stdout = devnull if quiet else None cmd = _tool_path('ninja', system_buildtools) + targets subprocess.check_call(cmd, cwd=os.path.abspath(out), stdout=stdout) def compute_source_dependencies(out, system_buildtools=False): """For each source file, computes a set of headers it depends on.""" ninja_deps = _check_command_output( _tool_path('ninja', system_buildtools) + ['-t', 'deps'], cwd=out) deps = {} current_source = None for line in ninja_deps.split('\n'): filename = os.path.relpath(os.path.join(out, line.strip()), repo_root()) # Sanitizer builds may have a dependency of ignorelist.txt. Just skip it. if filename.endswith('gn/standalone/sanitizers/ignorelist.txt'): continue if not line or line[0] != ' ': current_source = None continue elif not current_source: # We're assuming the source file is always listed before the # headers. assert os.path.splitext(line)[1] in ['.c', '.cc', '.cpp', '.S'] current_source = filename deps[current_source] = [] else: assert current_source deps[current_source].append(filename) return deps def label_to_path(label): """Turn a GN output label (e.g., //some_dir/file.cc) into a path.""" assert label.startswith('//') return label[2:] def label_without_toolchain(label): """Strips the toolchain from a GN label. Return a GN label (e.g //buildtools:protobuf(//gn/standalone/toolchain: gcc_like_host) without the parenthesised toolchain part. """ return label.split('(')[0] def label_to_target_name_with_path(label): """ Turn a GN label into a target name involving the full path. e.g., //src/perfetto:tests -> src_perfetto_tests """ name = re.sub(r'^//:?', '', label) name = re.sub(r'[^a-zA-Z0-9_]', '_', name) return name def gen_buildflags(gn_args, target_file): """Generates the perfetto_build_flags.h for the given config. target_file: the path, relative to the repo root, where the generated buildflag header will be copied into. """ tmp_out = prepare_out_directory(gn_args, 'tmp.gen_buildflags') build_targets(tmp_out, [BUILDFLAGS_TARGET], quiet=True) src = os.path.join(tmp_out, 'gen', 'build_config', 'perfetto_build_flags.h') shutil.copy(src, os.path.join(repo_root(), target_file)) shutil.rmtree(tmp_out) def check_or_commit_generated_files(tmp_files, check): """Checks that gen files are unchanged or renames them to the final location Takes in input a list of 'xxx.swp' files that have been written. If check == False, it renames xxx.swp -> xxx. If check == True, it just checks that the contents of 'xxx.swp' == 'xxx'. Returns 0 if no diff was detected, 1 otherwise (to be used as exit code). """ res = 0 for tmp_file in tmp_files: assert (tmp_file.endswith('.swp')) target_file = os.path.relpath(tmp_file[:-4]) if check: if not filecmp.cmp(tmp_file, target_file): sys.stderr.write('%s needs to be regenerated\n' % target_file) res = 1 os.unlink(tmp_file) else: os.replace(tmp_file, target_file) return res class ODRChecker(object): """Detects ODR violations in linker units When we turn GN source sets into Soong & Bazel file groups, there is the risk to create ODR violations by including the same file group into different linker unit (this is because other build systems don't have a concept equivalent to GN's source_set). This class navigates the transitive dependencies (mostly static libraries) of a target and detects if multiple paths end up including the same file group. This is to avoid situations like: traced.exe -> base(file group) traced.exe -> libperfetto(static lib) -> base(file group) """ def __init__(self, gn: 'GnParser', target_name: str): self.gn = gn self.root = gn.get_target(target_name) self.source_sets: Dict[str, Set[str]] = collections.defaultdict(set) self.deps_visited = set() self.source_set_hdr_only = {} self._visit(target_name) num_violations = 0 if target_name in ODR_VIOLATION_IGNORE_TARGETS: return for sset, paths in self.source_sets.items(): if self.is_header_only(sset): continue if len(paths) != 1: num_violations += 1 print( 'ODR violation in target %s, multiple paths include %s:\n %s' % (target_name, sset, '\n '.join(paths)), file=sys.stderr) if num_violations > 0: raise Exception('%d ODR violations detected. Build generation aborted' % num_violations) def _visit(self, target_name: str, parent_path=''): target = self.gn.get_target(target_name) path = ((parent_path + ' > ') if parent_path else '') + target_name if not target: raise Exception('Cannot find target %s' % target_name) for ssdep in target.transitive_source_set_deps(): name_and_path = '%s (via %s)' % (target_name, path) self.source_sets[ssdep.name].add(name_and_path) deps = set(target.non_proto_or_source_set_deps()).union( target.transitive_proto_deps()) - self.deps_visited for dep in deps: if dep.type == 'executable': continue # Execs are strong boundaries and don't cause ODR violations. # static_library dependencies should reset the path. It doesn't matter if # we get to a source file via: # source_set1 > static_lib > source.cc OR # source_set1 > source_set2 > static_lib > source.cc # This is NOT an ODR violation because source.cc is linked from the same # static library next_parent_path = path if dep.type != 'static_library' else '' self.deps_visited.add(dep.name) self._visit(dep.name, next_parent_path) def is_header_only(self, source_set_name: str): cached = self.source_set_hdr_only.get(source_set_name) if cached is not None: return cached target = self.gn.get_target(source_set_name) if target.type != 'source_set': raise TypeError('%s is not a source_set' % source_set_name) res = all(src.endswith('.h') for src in target.sources) self.source_set_hdr_only[source_set_name] = res return res class GnParser(object): """A parser with some cleverness for GN json desc files The main goals of this parser are: 1) Deal with the fact that other build systems don't have an equivalent notion to GN's source_set. Conversely to Bazel's and Soong's filegroups, GN source_sets expect that dependencies, cflags and other source_set properties propagate up to the linker unit (static_library, executable or shared_library). This parser simulates the same behavior: when a source_set is encountered, some of its variables (cflags and such) are copied up to the dependent targets. This is to allow gen_xxx to create one filegroup for each source_set and then squash all the other flags onto the linker unit. 2) Detect and special-case protobuf targets, figuring out the protoc-plugin being used. """ class Target(object): """Reperesents A GN target. Maked properties are propagated up the dependency chain when a source_set dependency is encountered. """ def __init__(self, name, type): self.name = name # e.g. //src/ipc:ipc VALID_TYPES = ('static_library', 'shared_library', 'executable', 'group', 'action', 'source_set', 'proto_library', 'generated_file') assert (type in VALID_TYPES) self.type = type self.testonly = False self.toolchain = None # These are valid only for type == proto_library. # This is typically: 'proto', 'protozero', 'ipc'. self.proto_plugin: Optional[str] = None self.proto_paths = set() self.proto_exports = set() self.sources = set() # TODO(primiano): consider whether the public section should be part of # bubbled-up sources. self.public_headers = set() # 'public' # These are valid only for type == 'action' self.data = set() self.inputs = set() self.outputs = set() self.script = None self.args = [] self.custom_action_type = None self.python_main = None # These variables are propagated up when encountering a dependency # on a source_set target. self.cflags = set() self.defines = set() self.deps: Set[GnParser.Target] = set() self.transitive_deps: Set[GnParser.Target] = set() self.libs = set() self.include_dirs = set() self.ldflags = set() # Deps on //gn:xxx have this flag set to True. These dependencies # are special because they pull third_party code from buildtools/. # We don't want to keep recursing into //buildtools in generators, # this flag is used to stop the recursion and create an empty # placeholder target once we hit //gn:protoc or similar. self.is_third_party_dep_ = False def non_proto_or_source_set_deps(self): return set(d for d in self.deps if d.type != 'proto_library' and d.type != 'source_set') def proto_deps(self): return set(d for d in self.deps if d.type == 'proto_library') def transitive_proto_deps(self): return set(d for d in self.transitive_deps if d.type == 'proto_library') def transitive_cpp_proto_deps(self): return set( d for d in self.transitive_deps if d.type == 'proto_library' and d.proto_plugin != 'descriptor' and d.proto_plugin != 'source_set') def transitive_source_set_deps(self): return set(d for d in self.transitive_deps if d.type == 'source_set') def __lt__(self, other): if isinstance(other, self.__class__): return self.name < other.name raise TypeError( '\'<\' not supported between instances of \'%s\' and \'%s\'' % (type(self).__name__, type(other).__name__)) def __repr__(self): return json.dumps( { k: (list(sorted(v)) if isinstance(v, set) else v) for (k, v) in iteritems(self.__dict__) }, indent=4, sort_keys=True) def update(self, other): for key in ('cflags', 'data', 'defines', 'deps', 'include_dirs', 'ldflags', 'transitive_deps', 'libs', 'proto_paths'): self.__dict__[key].update(other.__dict__.get(key, [])) def __init__(self, gn_desc): self.gn_desc_ = gn_desc self.all_targets = {} self.linker_units = {} # Executables, shared or static libraries. self.source_sets = {} self.actions = {} self.proto_libs = {} def get_target(self, gn_target_name: str) -> Target: """Returns a Target object from the fully qualified GN target name. It bubbles up variables from source_set dependencies as described in the class-level comments. """ target = self.all_targets.get(gn_target_name) if target is not None: return target # Target already processed. desc = self.gn_desc_.get(gn_target_name) if not desc: return None target = GnParser.Target(gn_target_name, desc['type']) target.testonly = desc.get('testonly', False) target.toolchain = desc.get('toolchain', None) self.all_targets[gn_target_name] = target # We should never have GN targets directly depend on buidtools. They # should hop via //gn:xxx, so we can give generators an opportunity to # override them. assert (not gn_target_name.startswith('//buildtools')) # Don't descend further into third_party targets. Genrators are supposed # to either ignore them or route to other externally-provided targets. if gn_target_name.startswith('//gn'): target.is_third_party_dep_ = True return target proto_target_type, proto_desc = self.get_proto_target_type(target) if proto_target_type: assert proto_desc self.proto_libs[target.name] = target target.type = 'proto_library' target.proto_plugin = proto_target_type target.proto_paths.update(self.get_proto_paths(proto_desc)) target.proto_exports.update(self.get_proto_exports(proto_desc)) target.sources.update( self.get_proto_sources(proto_target_type, proto_desc)) assert (all(x.endswith('.proto') for x in target.sources)) elif target.type == 'source_set': self.source_sets[gn_target_name] = target target.sources.update(desc.get('sources', [])) target.inputs.update(desc.get('inputs', [])) elif target.type in LINKER_UNIT_TYPES: self.linker_units[gn_target_name] = target target.sources.update(desc.get('sources', [])) elif target.type == 'action': self.actions[gn_target_name] = target target.data.update(desc.get('metadata', {}).get('perfetto_data', [])) target.inputs.update(desc.get('inputs', [])) target.sources.update(desc.get('sources', [])) outs = [re.sub('^//out/.+?/gen/', '', x) for x in desc['outputs']] target.outputs.update(outs) target.script = desc['script'] # Args are typically relative to the root build dir (../../xxx) # because root build dir is typically out/xxx/). target.args = [re.sub('^../../', '//', x) for x in desc['args']] action_types = desc.get('metadata', {}).get('perfetto_action_type_for_generator', []) target.custom_action_type = action_types[0] if len( action_types) > 0 else None python_main = desc.get('metadata', {}).get('perfetto_python_main', []) target.python_main = python_main[0] if python_main else None # Default for 'public' is //* - all headers in 'sources' are public. # TODO(primiano): if a 'public' section is specified (even if empty), then # the rest of 'sources' is considered inaccessible by gn. Consider # emulating that, so that generated build files don't end up with overly # accessible headers. public_headers = [x for x in desc.get('public', []) if x != '*'] target.public_headers.update(public_headers) target.cflags.update(desc.get('cflags', []) + desc.get('cflags_cc', [])) target.libs.update(desc.get('libs', [])) target.ldflags.update(desc.get('ldflags', [])) target.defines.update(desc.get('defines', [])) target.include_dirs.update(desc.get('include_dirs', [])) # Recurse in dependencies. for dep_name in desc.get('deps', []): dep = self.get_target(dep_name) # generated_file targets only exist for GN builds: we can safely ignore # them. if dep.type == 'generated_file': continue # When a proto_library depends on an action, that is always the "_gen" # rule of the action which is "private" to the proto_library rule. # therefore, just ignore it for dep tracking purposes. if dep.type == 'action' and proto_target_type is not None: target_no_toolchain = label_without_toolchain(target.name) dep_no_toolchain = label_without_toolchain(dep.name) assert (dep_no_toolchain == f'{target_no_toolchain}_gen') continue # Non-third party groups are only used for bubbling cflags etc so don't # add a dep. if dep.type == 'group' and not dep.is_third_party_dep_: target.update(dep) # Bubble up groups's cflags/ldflags etc. continue # Linker units act as a hard boundary making all their internal deps # opaque to the outside world. For this reason, do not propogate deps # transitively across them. if dep.type in LINKER_UNIT_TYPES: target.deps.add(dep) continue if dep.type == 'source_set': target.update(dep) # Bubble up source set's cflags/ldflags etc. elif dep.type == 'proto_library': target.proto_paths.update(dep.proto_paths) target.deps.add(dep) target.transitive_deps.add(dep) target.transitive_deps.update(dep.transitive_deps) return target def get_proto_exports(self, proto_desc): # exports in metadata will be available for source_set targets. metadata = proto_desc.get('metadata', {}) return metadata.get('exports', []) def get_proto_paths(self, proto_desc): metadata = proto_desc.get('metadata', {}) return metadata.get('proto_import_dirs', []) def get_proto_sources(self, proto_target_type, proto_desc): if proto_target_type == 'source_set': metadata = proto_desc.get('metadata', {}) return metadata.get('proto_library_sources', []) return proto_desc.get('sources', []) def get_proto_target_type( self, target: Target) -> Tuple[Optional[str], Optional[Dict]]: """ Checks if the target is a proto library and return the plugin. Returns: (None, None): if the target is not a proto library. (plugin, proto_desc) where |plugin| is 'proto' in the default (lite) case or 'protozero' or 'ipc' or 'descriptor'; |proto_desc| is the GN json desc of the target with the .proto sources (_gen target for non-descriptor types or the target itself for descriptor type). """ parts = target.name.split('(', 1) name = parts[0] toolchain = '(' + parts[1] if len(parts) > 1 else '' # Descriptor targets don't have a _gen target; instead we look for the # characteristic flag in the args of the target itself. desc = self.gn_desc_.get(target.name) if '--descriptor_set_out' in desc.get('args', []): return 'descriptor', desc # Source set proto targets have a non-empty proto_library_sources in the # metadata of the description. metadata = desc.get('metadata', {}) if 'proto_library_sources' in metadata: return 'source_set', desc # In all other cases, we want to look at the _gen target as that has the # important information. gen_desc = self.gn_desc_.get('%s_gen%s' % (name, toolchain)) if gen_desc is None or gen_desc['type'] != 'action': return None, None args = gen_desc.get('args', []) if '/protoc' not in args[0]: return None, None plugin = 'proto' for arg in (arg for arg in args if arg.startswith('--plugin=')): # |arg| at this point looks like: # --plugin=protoc-gen-plugin=gcc_like_host/protozero_plugin # or # --plugin=protoc-gen-plugin=protozero_plugin plugin = arg.split('=')[-1].split('/')[-1].replace('_plugin', '') return plugin, gen_desc