| # |
| # Copyright (C) 2019 The Android Open Source Project |
| # |
| # Licensed under the Apache License, Version 2.0 (the "License"); |
| # you may not use this file except in compliance with the License. |
| # You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| import argparse |
| import collections |
| import fnmatch |
| import hashlib |
| import os |
| import pathlib |
| import subprocess |
| import tempfile |
| import zipfile |
| def silent_call(cmd): |
| return subprocess.call( |
| cmd, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL) == 0 |
| def sha1sum(f): |
| with open(f, "rb") as fin: |
| return hashlib.sha1(fin.read()).hexdigest() |
| def sha1sum_without_signing_key(filepath): |
| apk = zipfile.ZipFile(filepath) |
| l = [] |
| for f in sorted(apk.namelist()): |
| if f.startswith("META-INF/"): |
| continue |
| l.append(hashlib.sha1(apk.read(f)).hexdigest()) |
| l.append(f) |
| return hashlib.sha1(",".join(l).encode()).hexdigest() |
| def strip_and_sha1sum(filepath): |
| """Strip informations of elf file and calculate sha1 hash.""" |
| tmp_filepath = filepath + ".tmp.no-build-id" |
| llvm_strip = [ |
| "llvm-strip", "--strip-all", "--keep-section=.ARM.attributes", |
| "--remove-section=.note.gnu.build-id", filepath, "-o", tmp_filepath |
| ] |
| strip_all_and_remove_build_id = lambda: silent_call(llvm_strip) |
| try: |
| if strip_all_and_remove_build_id(): |
| return sha1sum(tmp_filepath) |
| else: |
| return sha1sum(filepath) |
| finally: |
| if os.path.exists(tmp_filepath): |
| os.remove(tmp_filepath) |
| return sha1sum(filepath) |
| def make_filter_from_allowlists(allowlists, all_targets): |
| """Creates a callable filter from a list of allowlist files. |
| Allowlist can contain pathname patterns or skipped lines. Pathnames are case |
| insensitive. |
| Allowlist can contain single-line comments. Comment lines begin with # |
| For example, this ignores the file "system/build.prop": |
| SYSTEM/build.prop |
| This ignores txt files: |
| *.txt |
| This ignores files in directory "system/dontcare/" |
| SYSTEM/dontcare/* |
| This ignores lines prefixed with pat1 or pat2 in file "system/build.prop": |
| SYSTEM/build.prop=pat1 pat2 |
| Args: |
| allowlists: A list of allowlist filenames. |
| all_targets: A list of targets to compare. |
| Returns: |
| A callable object that accepts a file pathname and returns True if the file |
| is skipped by the allowlists and False when it is not. |
| """ |
| skipped_patterns = set() |
| skipped_lines = collections.defaultdict(list) |
| for allowlist in allowlists: |
| if not os.path.isfile(allowlist): |
| continue |
| with open(allowlist, "rb") as f: |
| for line in f: |
| pat = line.strip().decode() |
| if pat.startswith("#"): |
| continue |
| if pat and pat[-1] == "\\": |
| pat = pat.rstrip("\\") |
| if "=" in pat: |
| filename, prefixes = pat.split("=", 1) |
| prefixes = prefixes.split() |
| if prefixes: |
| skipped_lines[filename.lower()].extend(prefixes) |
| elif pat: |
| skipped_patterns.add(pat.lower()) |
| def diff_with_skipped_lines(filename, prefixes): |
| """Compares sha1 digest of file while ignoring lines. |
| Args: |
| filename: File to compare among each target. |
| prefixes: A list of prefixes. Lines that start with prefix are skipped. |
| Returns: |
| True if file is identical among each target. |
| """ |
| file_digest_respect_ignore = [] |
| for target in all_targets: |
| pathname = os.path.join(target, filename) |
| if not os.path.isfile(pathname): |
| return False |
| sha1 = hashlib.sha1() |
| with open(pathname, "rb") as f: |
| for line in f: |
| line_text = line.decode() |
| if not any(line_text.startswith(prefix) for prefix in prefixes): |
| sha1.update(line) |
| file_digest_respect_ignore.append(sha1.hexdigest()) |
| return (len(file_digest_respect_ignore) == len(all_targets) and |
| len(set(file_digest_respect_ignore)) == 1) |
| def allowlist_filter(filename): |
| norm_filename = filename.lower() |
| for pattern in skipped_patterns: |
| if fnmatch.fnmatch(norm_filename, pattern): |
| return True |
| if norm_filename in skipped_lines: |
| skipped_prefixes = skipped_lines[norm_filename] |
| return diff_with_skipped_lines(filename, skipped_prefixes) |
| return False |
| return allowlist_filter |
| def main(all_targets, |
| search_paths, |
| allowlists, |
| ignore_signing_key=False, |
| list_only=False): |
| def run(path): |
| is_executable_component = silent_call(["llvm-objdump", "-a", path]) |
| is_apk = path.endswith(".apk") |
| if is_executable_component: |
| return strip_and_sha1sum(path) |
| elif is_apk and ignore_signing_key: |
| return sha1sum_without_signing_key(path) |
| else: |
| return sha1sum(path) |
| # artifact_sha1_target_map[filename][sha1] = list of targets |
| artifact_sha1_target_map = collections.defaultdict( |
| lambda: collections.defaultdict(list)) |
| for target in all_targets: |
| paths = [] |
| for search_path in search_paths: |
| for path in pathlib.Path(target, search_path).glob("**/*"): |
| if path.exists() and not path.is_dir(): |
| paths.append((str(path), str(path.relative_to(target)))) |
| target_basename = os.path.basename(os.path.normpath(target)) |
| for path, filename in paths: |
| sha1 = 0 |
| if not list_only: |
| sha1 = run(path) |
| artifact_sha1_target_map[filename][sha1].append(target_basename) |
| def pretty_print(sha1, filename, targets, exclude_sha1): |
| if exclude_sha1: |
| return "{}, {}\n".format(filename, ";".join(targets)) |
| return "{}, {}, {}\n".format(filename, sha1[:10], ";".join(targets)) |
| def is_common(sha1_target_map): |
| for _, targets in sha1_target_map.items(): |
| return len(sha1_target_map) == 1 and len(targets) == len(all_targets) |
| return False |
| allowlist_filter = make_filter_from_allowlists(allowlists, all_targets) |
| common = [] |
| diff = [] |
| allowlisted_diff = [] |
| for filename, sha1_target_map in artifact_sha1_target_map.items(): |
| if is_common(sha1_target_map): |
| for sha1, targets in sha1_target_map.items(): |
| common.append(pretty_print(sha1, filename, targets, list_only)) |
| else: |
| if allowlist_filter(filename): |
| for sha1, targets in sha1_target_map.items(): |
| allowlisted_diff.append( |
| pretty_print(sha1, filename, targets, list_only)) |
| else: |
| for sha1, targets in sha1_target_map.items(): |
| diff.append(pretty_print(sha1, filename, targets, list_only)) |
| common = sorted(common) |
| diff = sorted(diff) |
| allowlisted_diff = sorted(allowlisted_diff) |
| header = "filename, sha1sum, targets\n" |
| if list_only: |
| header = "filename, targets\n" |
| with open("common.csv", "w") as fout: |
| fout.write(header) |
| fout.writelines(common) |
| with open("diff.csv", "w") as fout: |
| fout.write(header) |
| fout.writelines(diff) |
| with open("allowlisted_diff.csv", "w") as fout: |
| fout.write(header) |
| fout.writelines(allowlisted_diff) |
| def main_with_zip(extracted_paths, main_args): |
| for origin_path, tmp_path in zip(main_args.target, extracted_paths): |
| unzip_cmd = ["unzip", "-qd", tmp_path, os.path.join(origin_path, "*.zip")] |
| unzip_cmd.extend([os.path.join(s, "*") for s in main_args.search_path]) |
| subprocess.call(unzip_cmd) |
| main( |
| extracted_paths, |
| main_args.search_path, |
| main_args.allowlist, |
| main_args.ignore_signing_key, |
| list_only=main_args.list_only) |
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser( |
| prog="compare_images", |
| usage="compare_images -t model1 model2 [model...] -s dir1 [dir...] [-i] [-u] [-p] [-w allowlist1] [-w allowlist2]" |
| ) |
| parser.add_argument("-t", "--target", nargs="+", required=True) |
| parser.add_argument("-s", "--search_path", nargs="+", required=True) |
| parser.add_argument("-i", "--ignore_signing_key", action="store_true") |
| parser.add_argument( |
| "-l", |
| "--list_only", |
| action="store_true", |
| help="Compare file list only and ignore SHA-1 diff") |
| parser.add_argument("-u", "--unzip", action="store_true") |
| parser.add_argument("-p", "--preserve_extracted_files", action="store_true") |
| parser.add_argument("-w", "--allowlist", action="append", default=[]) |
| args = parser.parse_args() |
| if len(args.target) < 2: |
| parser.error("The number of targets has to be at least two.") |
| if args.unzip: |
| if args.preserve_extracted_files: |
| main_with_zip(args.target, args) |
| else: |
| with tempfile.TemporaryDirectory() as tmpdir: |
| target_in_tmp = [os.path.join(tmpdir, t) for t in args.target] |
| for p in target_in_tmp: |
| os.makedirs(p) |
| main_with_zip(target_in_tmp, args) |
| else: |
| main( |
| args.target, |
| args.search_path, |
| args.allowlist, |
| args.ignore_signing_key, |
| list_only=args.list_only) |