blob: 17ca25f403ccd0c366b854bda4b27d4d78617715 [file] [log] [blame]
# Copyright 2020 The Pigweed Authors
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may not
# use this file except in compliance with the License. You may obtain a copy of
# the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations under
# the License.
"""Functions for building code during presubmit checks."""
import base64
import contextlib
from dataclasses import dataclass
import itertools
import json
import logging
import os
import posixpath
from pathlib import Path
import re
import subprocess
from shutil import which
import sys
import tarfile
from typing import (
Any,
Callable,
Collection,
Container,
ContextManager,
Dict,
Iterable,
Iterator,
List,
Mapping,
Optional,
Sequence,
Set,
Tuple,
Union,
)
from pw_presubmit.presubmit import (
call,
Check,
FileFilter,
filter_paths,
install_package,
PresubmitResult,
SubStep,
)
from pw_presubmit.presubmit_context import (
PresubmitContext,
PresubmitFailure,
)
from pw_presubmit import (
bazel_parser,
format_code,
ninja_parser,
)
from pw_presubmit.tools import (
plural,
log_run,
format_command,
)
_LOG = logging.getLogger(__name__)
def bazel(ctx: PresubmitContext, cmd: str, *args: str, **kwargs) -> None:
"""Invokes Bazel with some common flags set.
Intended for use with bazel build and test. May not work with others.
"""
num_jobs: List[str] = []
if ctx.num_jobs is not None:
num_jobs.extend(('--jobs', str(ctx.num_jobs)))
keep_going: List[str] = []
if ctx.continue_after_build_error:
keep_going.append('--keep_going')
bazel_stdout = ctx.output_dir / 'bazel.stdout'
ctx.output_dir.mkdir(exist_ok=True, parents=True)
try:
with bazel_stdout.open('w') as outs:
call(
'bazel',
cmd,
'--verbose_failures',
'--verbose_explanations',
'--worker_verbose',
f'--symlink_prefix={ctx.output_dir / ".bazel-"}',
*num_jobs,
*keep_going,
*args,
cwd=ctx.root,
tee=outs,
call_annotation={'build_system': 'bazel'},
**kwargs,
)
except PresubmitFailure as exc:
failure = bazel_parser.parse_bazel_stdout(bazel_stdout)
if failure:
with ctx.failure_summary_log.open('w') as outs:
outs.write(failure)
raise exc
def _gn_value(value) -> str:
if isinstance(value, bool):
return str(value).lower()
if (
isinstance(value, str)
and '"' not in value
and not value.startswith("{")
and not value.startswith("[")
):
return f'"{value}"'
if isinstance(value, (list, tuple)):
return f'[{", ".join(_gn_value(a) for a in value)}]'
# Fall-back case handles integers as well as strings that already
# contain double quotation marks, or look like scopes or lists.
return str(value)
def gn_args_list(**kwargs) -> List[str]:
"""Return a list of formatted strings to use as gn args.
Currently supports bool, int, and str values. In the case of str values,
quotation marks will be added automatically, unless the string already
contains one or more double quotation marks, or starts with a { or [
character, in which case it will be passed through as-is.
"""
transformed_args = []
for arg, val in kwargs.items():
transformed_args.append(f'{arg}={_gn_value(val)}')
# Use ccache if available for faster repeat presubmit runs.
if which('ccache') and 'pw_command_launcher' not in kwargs:
transformed_args.append('pw_command_launcher="ccache"')
return transformed_args
def gn_args(**kwargs) -> str:
"""Builds a string to use for the --args argument to gn gen.
Currently supports bool, int, and str values. In the case of str values,
quotation marks will be added automatically, unless the string already
contains one or more double quotation marks, or starts with a { or [
character, in which case it will be passed through as-is.
"""
return '--args=' + ' '.join(gn_args_list(**kwargs))
def write_gn_args_file(destination_file: Path, **kwargs) -> str:
"""Write gn args to a file.
Currently supports bool, int, and str values. In the case of str values,
quotation marks will be added automatically, unless the string already
contains one or more double quotation marks, or starts with a { or [
character, in which case it will be passed through as-is.
Returns:
The contents of the written file.
"""
contents = '\n'.join(gn_args_list(**kwargs))
# Add a trailing linebreak
contents += '\n'
destination_file.parent.mkdir(exist_ok=True, parents=True)
if (
destination_file.is_file()
and destination_file.read_text(encoding='utf-8') == contents
):
# File is identical, don't re-write.
return contents
destination_file.write_text(contents, encoding='utf-8')
return contents
def gn_gen(
ctx: PresubmitContext,
*args: str,
gn_check: bool = True, # pylint: disable=redefined-outer-name
gn_fail_on_unused: bool = True,
export_compile_commands: Union[bool, str] = True,
preserve_args_gn: bool = False,
**gn_arguments,
) -> None:
"""Runs gn gen in the specified directory with optional GN args.
Runs with --check=system if gn_check=True. Note that this does not cover
generated files. Run gn_check() after building to check generated files.
"""
all_gn_args = dict(gn_arguments)
all_gn_args.update(ctx.override_gn_args)
_LOG.debug('%r', all_gn_args)
args_option = gn_args(**all_gn_args)
if not ctx.dry_run and not preserve_args_gn:
# Delete args.gn to ensure this is a clean build.
args_gn = ctx.output_dir / 'args.gn'
if args_gn.is_file():
args_gn.unlink()
export_commands_arg = ''
if export_compile_commands:
export_commands_arg = '--export-compile-commands'
if isinstance(export_compile_commands, str):
export_commands_arg += f'={export_compile_commands}'
call(
'gn',
'gen',
ctx.output_dir,
'--color=always',
*(['--check=system'] if gn_check else []),
*(['--fail-on-unused-args'] if gn_fail_on_unused else []),
*([export_commands_arg] if export_commands_arg else []),
*args,
*([args_option] if all_gn_args else []),
cwd=ctx.root,
call_annotation={
'gn_gen_args': all_gn_args,
'gn_gen_args_option': args_option,
},
)
def gn_check(ctx: PresubmitContext) -> PresubmitResult:
"""Runs gn check, including on generated and system files."""
call(
'gn',
'check',
ctx.output_dir,
'--check-generated',
'--check-system',
cwd=ctx.root,
)
return PresubmitResult.PASS
def ninja(
ctx: PresubmitContext,
*args,
save_compdb: bool = True,
save_graph: bool = True,
**kwargs,
) -> None:
"""Runs ninja in the specified directory."""
num_jobs: List[str] = []
if ctx.num_jobs is not None:
num_jobs.extend(('-j', str(ctx.num_jobs)))
keep_going: List[str] = []
if ctx.continue_after_build_error:
keep_going.extend(('-k', '0'))
if save_compdb:
proc = log_run(
['ninja', '-C', ctx.output_dir, '-t', 'compdb', *args],
capture_output=True,
**kwargs,
)
if not ctx.dry_run:
(ctx.output_dir / 'ninja.compdb').write_bytes(proc.stdout)
if save_graph:
proc = log_run(
['ninja', '-C', ctx.output_dir, '-t', 'graph', *args],
capture_output=True,
**kwargs,
)
if not ctx.dry_run:
(ctx.output_dir / 'ninja.graph').write_bytes(proc.stdout)
ninja_stdout = ctx.output_dir / 'ninja.stdout'
ctx.output_dir.mkdir(exist_ok=True, parents=True)
try:
with ninja_stdout.open('w') as outs:
if sys.platform == 'win32':
# Windows doesn't support pw-wrap-ninja.
ninja_command = ['ninja']
else:
ninja_command = ['pw-wrap-ninja', '--log-actions']
call(
*ninja_command,
'-C',
ctx.output_dir,
*num_jobs,
*keep_going,
*args,
tee=outs,
propagate_sigterm=True,
call_annotation={'build_system': 'ninja'},
**kwargs,
)
except PresubmitFailure as exc:
failure = ninja_parser.parse_ninja_stdout(ninja_stdout)
if failure:
with ctx.failure_summary_log.open('w') as outs:
outs.write(failure)
raise exc
def get_gn_args(directory: Path) -> List[Dict[str, Dict[str, str]]]:
"""Dumps GN variables to JSON."""
proc = log_run(
['gn', 'args', directory, '--list', '--json'], stdout=subprocess.PIPE
)
return json.loads(proc.stdout)
def cmake(
ctx: PresubmitContext,
*args: str,
env: Optional[Mapping['str', 'str']] = None,
) -> None:
"""Runs CMake for Ninja on the given source and output directories."""
call(
'cmake',
'-B',
ctx.output_dir,
'-S',
ctx.root,
'-G',
'Ninja',
*args,
env=env,
)
def env_with_clang_vars() -> Mapping[str, str]:
"""Returns the environment variables with CC, CXX, etc. set for clang."""
env = os.environ.copy()
env['CC'] = env['LD'] = env['AS'] = 'clang'
env['CXX'] = 'clang++'
return env
def _get_paths_from_command(source_dir: Path, *args, **kwargs) -> Set[Path]:
"""Runs a command and reads Bazel or GN //-style paths from it."""
process = log_run(args, capture_output=True, cwd=source_dir, **kwargs)
if process.returncode:
_LOG.error(
'Build invocation failed with return code %d!', process.returncode
)
_LOG.error(
'[COMMAND] %s\n%s\n%s',
*format_command(args, kwargs),
process.stderr.decode(),
)
raise PresubmitFailure
files = set()
for line in process.stdout.splitlines():
path = line.strip().lstrip(b'/').replace(b':', b'/').decode()
path = source_dir.joinpath(path)
if path.is_file():
files.add(path)
return files
# Finds string literals with '.' in them.
_MAYBE_A_PATH = re.compile(
r'"' # Starting double quote.
# Start capture group 1 - the whole filename:
# File basename, a single period, file extension.
r'([^\n" ]+\.[^\n" ]+)'
# Non-capturing group 2 (optional).
r'(?: > [^\n"]+)?' # pw_zip style string "input_file.txt > output_file.txt"
r'"' # Ending double quote.
)
def _search_files_for_paths(build_files: Iterable[Path]) -> Iterable[Path]:
for build_file in build_files:
directory = build_file.parent
for string in _MAYBE_A_PATH.finditer(build_file.read_text()):
path = directory / string.group(1)
if path.is_file():
yield path
def _read_compile_commands(compile_commands: Path) -> dict:
with compile_commands.open('rb') as fd:
return json.load(fd)
def compiled_files(compile_commands: Path) -> Iterable[Path]:
for command in _read_compile_commands(compile_commands):
file = Path(command['file'])
if file.is_absolute():
yield file
else:
yield file.joinpath(command['directory']).resolve()
def check_compile_commands_for_files(
compile_commands: Union[Path, Iterable[Path]],
files: Iterable[Path],
extensions: Collection[str] = format_code.CPP_SOURCE_EXTS,
) -> List[Path]:
"""Checks for paths in one or more compile_commands.json files.
Only checks C and C++ source files by default.
"""
if isinstance(compile_commands, Path):
compile_commands = [compile_commands]
compiled = frozenset(
itertools.chain.from_iterable(
compiled_files(cmds) for cmds in compile_commands
)
)
return [f for f in files if f not in compiled and f.suffix in extensions]
def check_bazel_build_for_files(
bazel_extensions_to_check: Container[str],
files: Iterable[Path],
bazel_dirs: Iterable[Path] = (),
) -> List[Path]:
"""Checks that source files are in the Bazel builds.
Args:
bazel_extensions_to_check: which file suffixes to look for in Bazel
files: the files that should be checked
bazel_dirs: directories in which to run bazel query
Returns:
a list of missing files; will be empty if there were no missing files
"""
# Collect all paths in the Bazel builds.
bazel_builds: Set[Path] = set()
for directory in bazel_dirs:
bazel_builds.update(
_get_paths_from_command(
directory, 'bazel', 'query', 'kind("source file", //...:*)'
)
)
missing: List[Path] = []
if bazel_dirs:
for path in (p for p in files if p.suffix in bazel_extensions_to_check):
if path not in bazel_builds:
# TODO: b/234883555 - Replace this workaround for fuzzers.
if 'fuzz' not in str(path):
missing.append(path)
if missing:
_LOG.warning(
'%s missing from the Bazel build:\n%s',
plural(missing, 'file', are=True),
'\n'.join(str(x) for x in missing),
)
return missing
def check_gn_build_for_files(
gn_extensions_to_check: Container[str],
files: Iterable[Path],
gn_dirs: Iterable[Tuple[Path, Path]] = (),
gn_build_files: Iterable[Path] = (),
) -> List[Path]:
"""Checks that source files are in the GN build.
Args:
gn_extensions_to_check: which file suffixes to look for in GN
files: the files that should be checked
gn_dirs: (source_dir, output_dir) tuples with which to run gn desc
gn_build_files: paths to BUILD.gn files to directly search for paths
Returns:
a list of missing files; will be empty if there were no missing files
"""
# Collect all paths in GN builds.
gn_builds: Set[Path] = set()
for source_dir, output_dir in gn_dirs:
gn_builds.update(
_get_paths_from_command(source_dir, 'gn', 'desc', output_dir, '*')
)
gn_builds.update(_search_files_for_paths(gn_build_files))
missing: List[Path] = []
if gn_dirs or gn_build_files:
for path in (p for p in files if p.suffix in gn_extensions_to_check):
if path not in gn_builds:
missing.append(path)
if missing:
_LOG.warning(
'%s missing from the GN build:\n%s',
plural(missing, 'file', are=True),
'\n'.join(str(x) for x in missing),
)
return missing
def check_builds_for_files(
bazel_extensions_to_check: Container[str],
gn_extensions_to_check: Container[str],
files: Iterable[Path],
bazel_dirs: Iterable[Path] = (),
gn_dirs: Iterable[Tuple[Path, Path]] = (),
gn_build_files: Iterable[Path] = (),
) -> Dict[str, List[Path]]:
"""Checks that source files are in the GN and Bazel builds.
Args:
bazel_extensions_to_check: which file suffixes to look for in Bazel
gn_extensions_to_check: which file suffixes to look for in GN
files: the files that should be checked
bazel_dirs: directories in which to run bazel query
gn_dirs: (source_dir, output_dir) tuples with which to run gn desc
gn_build_files: paths to BUILD.gn files to directly search for paths
Returns:
a dictionary mapping build system ('Bazel' or 'GN' to a list of missing
files; will be empty if there were no missing files
"""
bazel_missing = check_bazel_build_for_files(
bazel_extensions_to_check=bazel_extensions_to_check,
files=files,
bazel_dirs=bazel_dirs,
)
gn_missing = check_gn_build_for_files(
gn_extensions_to_check=gn_extensions_to_check,
files=files,
gn_dirs=gn_dirs,
gn_build_files=gn_build_files,
)
result = {}
if bazel_missing:
result['Bazel'] = bazel_missing
if gn_missing:
result['GN'] = gn_missing
return result
@contextlib.contextmanager
def test_server(executable: str, output_dir: Path):
"""Context manager that runs a test server executable.
Args:
executable: name of the test server executable
output_dir: path to the output directory (for logs)
"""
with open(output_dir / 'test_server.log', 'w') as outs:
try:
proc = subprocess.Popen(
[executable, '--verbose'],
stdout=outs,
stderr=subprocess.STDOUT,
)
yield
finally:
proc.terminate() # pylint: disable=used-before-assignment
@contextlib.contextmanager
def modified_env(**envvars):
"""Context manager that sets environment variables.
Use by assigning values to variable names in the argument list, e.g.:
`modified_env(MY_FLAG="some value")`
Args:
envvars: Keyword arguments
"""
saved_env = os.environ.copy()
os.environ.update(envvars)
try:
yield
finally:
os.environ = saved_env
def fuzztest_prng_seed(ctx: PresubmitContext) -> str:
"""Convert the RNG seed to the format expected by FuzzTest.
FuzzTest can be configured to use the seed by setting the
`FUZZTEST_PRNG_SEED` environment variable to this value.
Args:
ctx: The context that includes a pseudorandom number generator seed.
"""
rng_bytes = ctx.rng_seed.to_bytes(32, sys.byteorder)
return base64.urlsafe_b64encode(rng_bytes).decode('ascii').rstrip('=')
@filter_paths(
file_filter=FileFilter(endswith=('.bzl', '.bazel'), name=('WORKSPACE',))
)
def bazel_lint(ctx: PresubmitContext):
"""Runs buildifier with lint on Bazel files.
Should be run after bazel_format since that will give more useful output
for formatting-only issues.
"""
failure = False
for path in ctx.paths:
try:
call('buildifier', '--lint=warn', '--mode=check', path)
except PresubmitFailure:
failure = True
if failure:
raise PresubmitFailure
@Check
def gn_gen_check(ctx: PresubmitContext):
"""Runs gn gen --check to enforce correct header dependencies."""
gn_gen(ctx, gn_check=True)
Item = Union[int, str]
Value = Union[Item, Sequence[Item]]
ValueCallable = Callable[[PresubmitContext], Value]
InputItem = Union[Item, ValueCallable]
InputValue = Union[InputItem, Sequence[InputItem]]
def _value(ctx: PresubmitContext, val: InputValue) -> Value:
"""Process any lambdas inside val
val is a single value or a list of values, any of which might be a lambda
that needs to be resolved. Call each of these lambdas with ctx and replace
the lambda with the result. Return the updated top-level structure.
"""
# TODO(mohrr) Use typing.TypeGuard instead of "type: ignore"
if isinstance(val, (str, int)):
return val
if callable(val):
return val(ctx)
result: List[Item] = []
for item in val:
if callable(item):
call_result = item(ctx)
if isinstance(item, (int, str)):
result.append(call_result)
else: # Sequence.
result.extend(call_result) # type: ignore
elif isinstance(item, (int, str)):
result.append(item)
else: # Sequence.
result.extend(item)
return result
_CtxMgrLambda = Callable[[PresubmitContext], ContextManager]
_CtxMgrOrLambda = Union[ContextManager, _CtxMgrLambda]
_INCREMENTAL_COVERAGE_TOOL = 'cloud_client'
_ABSOLUTE_COVERAGE_TOOL = 'raw_coverage_cloud_uploader'
@dataclass(frozen=True)
class CoverageOptions:
"""Coverage collection configuration.
For Google use only. See go/kalypsi-abs and go/kalypsi-inc for documentation
of the metadata fields.
"""
target_bucket_root: str
target_bucket_project: str
project: str
trace_type: str
ref: str
source: str
owner: str
bug_component: str
trim_prefix: str = ''
add_prefix: str = ''
class _NinjaBase(Check):
"""Thin wrapper of Check for steps that call ninja."""
def __init__(
self,
*args,
packages: Sequence[str] = (),
ninja_contexts: Sequence[_CtxMgrOrLambda] = (),
ninja_targets: Union[str, Sequence[str], Sequence[Sequence[str]]] = (),
coverage_options: Optional[CoverageOptions] = None,
**kwargs,
):
"""Initializes a _NinjaBase object.
Args:
*args: Passed on to superclass.
packages: List of 'pw package' packages to install.
ninja_contexts: List of context managers to apply around ninja
calls.
ninja_targets: Single ninja target, list of Ninja targets, or list
of list of ninja targets. If a list of a list, ninja will be
called multiple times with the same build directory.
coverage_options: Coverage collection options (or None, if not
collecting coverage data).
**kwargs: Passed on to superclass.
"""
super().__init__(*args, **kwargs)
self._packages: Sequence[str] = packages
self._ninja_contexts: Tuple[_CtxMgrOrLambda, ...] = tuple(
ninja_contexts
)
self._coverage_options = coverage_options
if isinstance(ninja_targets, str):
ninja_targets = (ninja_targets,)
ninja_targets = list(ninja_targets)
all_strings = all(isinstance(x, str) for x in ninja_targets)
any_strings = any(isinstance(x, str) for x in ninja_targets)
if ninja_targets and all_strings != any_strings:
raise ValueError(repr(ninja_targets))
self._ninja_target_lists: Tuple[Tuple[str, ...], ...]
if all_strings:
targets: List[str] = []
for target in ninja_targets:
targets.append(target) # type: ignore
self._ninja_target_lists = (tuple(targets),)
else:
self._ninja_target_lists = tuple(tuple(x) for x in ninja_targets)
@property
def ninja_targets(self) -> List[str]:
return list(itertools.chain(*self._ninja_target_lists))
def _install_package( # pylint: disable=no-self-use
self,
ctx: PresubmitContext,
package: str,
) -> PresubmitResult:
install_package(ctx, package)
return PresubmitResult.PASS
@contextlib.contextmanager
def _context(self, ctx: PresubmitContext):
"""Apply any context managers necessary for building."""
with contextlib.ExitStack() as stack:
for mgr in self._ninja_contexts:
if isinstance(mgr, contextlib.AbstractContextManager):
stack.enter_context(mgr)
else:
stack.enter_context(mgr(ctx)) # type: ignore
yield
def _ninja(
self, ctx: PresubmitContext, targets: Sequence[str]
) -> PresubmitResult:
with self._context(ctx):
ninja(ctx, *targets)
return PresubmitResult.PASS
def _coverage(
self, ctx: PresubmitContext, options: CoverageOptions
) -> PresubmitResult:
"""Archive and (on LUCI) upload coverage reports."""
reports = ctx.output_dir / 'coverage_reports'
os.makedirs(reports, exist_ok=True)
coverage_jsons: List[Path] = []
for path in ctx.output_dir.rglob('coverage_report'):
_LOG.debug('exploring %s', path)
name = str(path.relative_to(ctx.output_dir))
name = name.replace('_', '').replace('/', '_')
with tarfile.open(reports / f'{name}.tar.gz', 'w:gz') as tar:
tar.add(path, arcname=name, recursive=True)
json_path = path / 'json' / 'report.json'
if json_path.is_file():
_LOG.debug('found json %s', json_path)
coverage_jsons.append(json_path)
if not coverage_jsons:
ctx.fail('No coverage json file found')
return PresubmitResult.FAIL
if len(coverage_jsons) > 1:
_LOG.warning(
'More than one coverage json file, selecting first: %r',
coverage_jsons,
)
coverage_json = coverage_jsons[0]
if ctx.luci:
if ctx.luci.is_dev:
_LOG.warning('Not uploading coverage since running in dev')
return PresubmitResult.PASS
with self._context(ctx):
# GCS bucket paths are POSIX-like.
coverage_gcs_path = posixpath.join(
options.target_bucket_root,
'incremental' if ctx.luci.is_try else 'absolute',
options.target_bucket_project,
str(ctx.luci.buildbucket_id),
)
_copy_to_gcs(
ctx,
coverage_json,
posixpath.join(coverage_gcs_path, 'report.json'),
)
metadata_json = _write_coverage_metadata(ctx, options)
_copy_to_gcs(
ctx,
metadata_json,
posixpath.join(coverage_gcs_path, 'metadata.json'),
)
return PresubmitResult.PASS
_LOG.warning('Not uploading coverage since running locally')
return PresubmitResult.PASS
def _package_substeps(self) -> Iterator[SubStep]:
for package in self._packages:
yield SubStep(
f'install {package} package',
self._install_package,
(package,),
)
def _ninja_substeps(self) -> Iterator[SubStep]:
targets_parts = set()
for targets in self._ninja_target_lists:
targets_part = " ".join(targets)
maxlen = 70
if len(targets_part) > maxlen:
targets_part = f'{targets_part[0:maxlen-3]}...'
assert targets_part not in targets_parts
targets_parts.add(targets_part)
yield SubStep(f'ninja {targets_part}', self._ninja, (targets,))
def _coverage_substeps(self) -> Iterator[SubStep]:
if self._coverage_options is not None:
yield SubStep('coverage', self._coverage, (self._coverage_options,))
def _copy_to_gcs(ctx: PresubmitContext, filepath: Path, gcs_dst: str):
cmd = [
"gsutil",
"cp",
filepath,
gcs_dst,
]
upload_stdout = ctx.output_dir / (filepath.name + '.stdout')
with upload_stdout.open('w') as outs:
call(*cmd, tee=outs)
def _write_coverage_metadata(
ctx: PresubmitContext, options: CoverageOptions
) -> Path:
"""Write out Kalypsi coverage metadata and return the path to it."""
assert ctx.luci is not None
assert len(ctx.luci.triggers) == 1
change = ctx.luci.triggers[0]
metadata = {
'host': change.gerrit_host,
'project': options.project,
'trace_type': options.trace_type,
'trim_prefix': options.trim_prefix,
'patchset_num': change.patchset,
'change_id': change.number,
'owner': options.owner,
'bug_component': options.bug_component,
}
if ctx.luci.is_try:
# Running in CQ: uploading incremental coverage
metadata.update(
{
# Note: no `add_prefix`. According to the documentation, that's
# only supported for absolute coverage.
#
# TODO(tpudlik): Follow up with Kalypsi team, since this is
# surprising (given that trim_prefix is supported for both types
# of coverage). This might be an error in the docs.
'patchset_num': change.patchset,
'change_id': change.number,
}
)
else:
# Running in CI: uploading absolute coverage
metadata.update(
{
'add_prefix': options.add_prefix,
'commit_id': change.ref,
'ref': options.ref,
'source': options.source,
}
)
metadata_json = ctx.output_dir / "metadata.json"
with metadata_json.open('w') as metadata_file:
json.dump(metadata, metadata_file)
return metadata_json
class GnGenNinja(_NinjaBase):
"""Thin wrapper of Check for steps that just call gn/ninja.
Runs gn gen, ninja, then gn check.
"""
def __init__(
self,
*args,
gn_args: Optional[ # pylint: disable=redefined-outer-name
Dict[str, Any]
] = None,
**kwargs,
):
"""Initializes a GnGenNinja object.
Args:
*args: Passed on to superclass.
gn_args: Dict of GN args.
**kwargs: Passed on to superclass.
"""
super().__init__(self._substeps(), *args, **kwargs)
self._gn_args: Dict[str, Any] = gn_args or {}
@property
def gn_args(self) -> Dict[str, Any]:
return self._gn_args
def _gn_gen(self, ctx: PresubmitContext) -> PresubmitResult:
args: Dict[str, Any] = {}
if self._coverage_options is not None:
args['pw_toolchain_COVERAGE_ENABLED'] = True
args['pw_build_PYTHON_TEST_COVERAGE'] = True
args['pw_C_OPTIMIZATION_LEVELS'] = ('debug',)
if ctx.incremental:
args['pw_toolchain_PROFILE_SOURCE_FILES'] = [
f'//{x.relative_to(ctx.root)}' for x in ctx.paths
]
args.update({k: _value(ctx, v) for k, v in self._gn_args.items()})
gn_gen(ctx, gn_check=False, **args) # type: ignore
return PresubmitResult.PASS
def _substeps(self) -> Iterator[SubStep]:
yield from self._package_substeps()
yield SubStep('gn gen', self._gn_gen)
yield from self._ninja_substeps()
# Run gn check after building so it can check generated files.
yield SubStep('gn check', gn_check)
yield from self._coverage_substeps()