2015-03-24 03:13:21 +01:00
|
|
|
#!/usr/bin/env python
|
|
|
|
#
|
|
|
|
# Copyright (C) 2008 The Android Open Source Project
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
"""
|
|
|
|
Signs all the APK files in a target-files zipfile, producing a new
|
|
|
|
target-files zip.
|
|
|
|
|
|
|
|
Usage: sign_target_files_apks [flags] input_target_files output_target_files
|
|
|
|
|
|
|
|
-e (--extra_apks) <name,name,...=key>
|
2019-03-15 17:37:01 +01:00
|
|
|
Add extra APK/APEX name/key pairs as though they appeared in apkcerts.txt
|
|
|
|
or apexkeys.txt (so mappings specified by -k and -d are applied). Keys
|
|
|
|
specified in -e override any value for that app contained in the
|
|
|
|
apkcerts.txt file, or the container key for an APEX. Option may be
|
|
|
|
repeated to give multiple extra packages.
|
|
|
|
|
|
|
|
--extra_apex_payload_key <name=key>
|
|
|
|
Add a mapping for APEX package name to payload signing key, which will
|
|
|
|
override the default payload signing key in apexkeys.txt. Note that the
|
|
|
|
container key should be overridden via the `--extra_apks` flag above.
|
|
|
|
Option may be repeated for multiple APEXes.
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2018-06-19 21:19:35 +02:00
|
|
|
--skip_apks_with_path_prefix <prefix>
|
|
|
|
Skip signing an APK if it has the matching prefix in its path. The prefix
|
|
|
|
should be matching the entry name, which has partition names in upper
|
|
|
|
case, e.g. "VENDOR/app/", or "SYSTEM_OTHER/preloads/". Option may be
|
|
|
|
repeated to give multiple prefixes.
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
-k (--key_mapping) <src_key=dest_key>
|
|
|
|
Add a mapping from the key name as specified in apkcerts.txt (the
|
|
|
|
src_key) to the real key you wish to sign the package with
|
|
|
|
(dest_key). Option may be repeated to give multiple key
|
|
|
|
mappings.
|
|
|
|
|
|
|
|
-d (--default_key_mappings) <dir>
|
|
|
|
Set up the following key mappings:
|
|
|
|
|
|
|
|
$devkey/devkey ==> $dir/releasekey
|
|
|
|
$devkey/testkey ==> $dir/releasekey
|
|
|
|
$devkey/media ==> $dir/media
|
|
|
|
$devkey/shared ==> $dir/shared
|
|
|
|
$devkey/platform ==> $dir/platform
|
|
|
|
|
|
|
|
where $devkey is the directory part of the value of
|
|
|
|
default_system_dev_certificate from the input target-files's
|
2019-04-10 06:35:37 +02:00
|
|
|
META/misc_info.txt. (Defaulting to "build/make/target/product/security"
|
2015-03-24 03:13:21 +01:00
|
|
|
if the value is not present in misc_info.
|
|
|
|
|
|
|
|
-d and -k options are added to the set of mappings in the order
|
|
|
|
in which they appear on the command line.
|
|
|
|
|
|
|
|
-o (--replace_ota_keys)
|
2016-06-16 23:41:24 +02:00
|
|
|
Replace the certificate (public key) used by OTA package verification
|
|
|
|
with the ones specified in the input target_files zip (in the
|
|
|
|
META/otakeys.txt file). Key remapping (-k and -d) is performed on the
|
|
|
|
keys. For A/B devices, the payload verification key will be replaced
|
|
|
|
as well. If there're multiple OTA keys, only the first one will be used
|
|
|
|
for payload verification.
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
-t (--tag_changes) <+tag>,<-tag>,...
|
|
|
|
Comma-separated list of changes to make to the set of tags (in
|
|
|
|
the last component of the build fingerprint). Prefix each with
|
|
|
|
'+' or '-' to indicate whether that tag should be added or
|
|
|
|
removed. Changes are processed in the order they appear.
|
|
|
|
Default value is "-test-keys,-dev-keys,+release-keys".
|
|
|
|
|
2016-06-18 02:01:22 +02:00
|
|
|
--replace_verity_private_key <key>
|
|
|
|
Replace the private key used for verity signing. It expects a filename
|
|
|
|
WITHOUT the extension (e.g. verity_key).
|
|
|
|
|
|
|
|
--replace_verity_public_key <key>
|
|
|
|
Replace the certificate (public key) used for verity verification. The
|
|
|
|
key file replaces the one at BOOT/RAMDISK/verity_key (or ROOT/verity_key
|
|
|
|
for devices using system_root_image). It expects the key filename WITH
|
|
|
|
the extension (e.g. verity_key.pub).
|
|
|
|
|
2016-06-17 04:58:44 +02:00
|
|
|
--replace_verity_keyid <path_to_X509_PEM_cert_file>
|
|
|
|
Replace the veritykeyid in BOOT/cmdline of input_target_file_zip
|
2016-06-18 02:01:22 +02:00
|
|
|
with keyid of the cert pointed by <path_to_X509_PEM_cert_file>.
|
2017-06-20 00:48:02 +02:00
|
|
|
|
2019-02-27 03:15:51 +01:00
|
|
|
--avb_{boot,system,system_other,vendor,dtbo,vbmeta}_algorithm <algorithm>
|
|
|
|
--avb_{boot,system,system_other,vendor,dtbo,vbmeta}_key <key>
|
2017-06-20 00:48:02 +02:00
|
|
|
Use the specified algorithm (e.g. SHA256_RSA4096) and the key to AVB-sign
|
|
|
|
the specified image. Otherwise it uses the existing values in info dict.
|
|
|
|
|
2019-02-27 03:15:51 +01:00
|
|
|
--avb_{apex,boot,system,system_other,vendor,dtbo,vbmeta}_extra_args <args>
|
2017-06-20 00:48:02 +02:00
|
|
|
Specify any additional args that are needed to AVB-sign the image
|
|
|
|
(e.g. "--signing_helper /path/to/helper"). The args will be appended to
|
|
|
|
the existing ones in info dict.
|
2015-03-24 03:13:21 +01:00
|
|
|
"""
|
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
from __future__ import print_function
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
import base64
|
|
|
|
import copy
|
|
|
|
import errno
|
2017-08-14 15:49:21 +02:00
|
|
|
import gzip
|
2019-03-15 17:37:01 +01:00
|
|
|
import itertools
|
2019-03-15 17:33:43 +01:00
|
|
|
import logging
|
2015-03-24 03:13:21 +01:00
|
|
|
import os
|
|
|
|
import re
|
2017-08-14 15:49:21 +02:00
|
|
|
import shutil
|
2017-07-12 20:57:05 +02:00
|
|
|
import stat
|
2015-03-24 03:13:21 +01:00
|
|
|
import subprocess
|
2017-12-24 19:37:38 +01:00
|
|
|
import sys
|
2015-03-24 03:13:21 +01:00
|
|
|
import tempfile
|
|
|
|
import zipfile
|
2017-12-05 02:16:36 +01:00
|
|
|
from xml.etree import ElementTree
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
import add_img_to_target_files
|
2019-03-15 17:37:01 +01:00
|
|
|
import apex_utils
|
2015-03-24 03:13:21 +01:00
|
|
|
import common
|
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
|
|
|
|
if sys.hexversion < 0x02070000:
|
|
|
|
print("Python 2.7 or newer is required.", file=sys.stderr)
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
|
2019-03-15 17:33:43 +01:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
OPTIONS = common.OPTIONS
|
|
|
|
|
|
|
|
OPTIONS.extra_apks = {}
|
2019-03-15 17:37:01 +01:00
|
|
|
OPTIONS.extra_apex_payload_keys = {}
|
2018-06-19 21:19:35 +02:00
|
|
|
OPTIONS.skip_apks_with_path_prefix = set()
|
2015-03-24 03:13:21 +01:00
|
|
|
OPTIONS.key_map = {}
|
2017-05-23 23:51:02 +02:00
|
|
|
OPTIONS.rebuild_recovery = False
|
2015-03-24 03:13:21 +01:00
|
|
|
OPTIONS.replace_ota_keys = False
|
|
|
|
OPTIONS.replace_verity_public_key = False
|
|
|
|
OPTIONS.replace_verity_private_key = False
|
2016-06-17 04:58:44 +02:00
|
|
|
OPTIONS.replace_verity_keyid = False
|
2015-03-24 03:13:21 +01:00
|
|
|
OPTIONS.tag_changes = ("-test-keys", "-dev-keys", "+release-keys")
|
2017-06-20 00:48:02 +02:00
|
|
|
OPTIONS.avb_keys = {}
|
|
|
|
OPTIONS.avb_algorithms = {}
|
|
|
|
OPTIONS.avb_extra_args = {}
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
|
2017-08-14 15:49:21 +02:00
|
|
|
def GetApkCerts(certmap):
|
2015-03-24 03:13:21 +01:00
|
|
|
# apply the key remapping to the contents of the file
|
|
|
|
for apk, cert in certmap.iteritems():
|
|
|
|
certmap[apk] = OPTIONS.key_map.get(cert, cert)
|
|
|
|
|
|
|
|
# apply all the -e options, overriding anything in the file
|
|
|
|
for apk, cert in OPTIONS.extra_apks.iteritems():
|
|
|
|
if not cert:
|
|
|
|
cert = "PRESIGNED"
|
|
|
|
certmap[apk] = OPTIONS.key_map.get(cert, cert)
|
|
|
|
|
|
|
|
return certmap
|
|
|
|
|
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
def GetApexKeys(keys_info, key_map):
|
|
|
|
"""Gets APEX payload and container signing keys by applying the mapping rules.
|
|
|
|
|
2019-03-19 20:24:03 +01:00
|
|
|
Presigned payload / container keys will be set accordingly.
|
2019-03-15 17:37:01 +01:00
|
|
|
|
|
|
|
Args:
|
|
|
|
keys_info: A dict that maps from APEX filenames to a tuple of (payload_key,
|
|
|
|
container_key).
|
|
|
|
key_map: A dict that overrides the keys, specified via command-line input.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
A dict that contains the updated APEX key mapping, which should be used for
|
|
|
|
the current signing.
|
|
|
|
"""
|
|
|
|
# Apply all the --extra_apex_payload_key options to override the payload
|
|
|
|
# signing keys in the given keys_info.
|
|
|
|
for apex, key in OPTIONS.extra_apex_payload_keys.items():
|
2019-03-19 20:24:03 +01:00
|
|
|
if not key:
|
|
|
|
key = 'PRESIGNED'
|
2019-03-15 17:37:01 +01:00
|
|
|
keys_info[apex] = (key, keys_info[apex][1])
|
|
|
|
|
|
|
|
# Apply the key remapping to container keys.
|
|
|
|
for apex, (payload_key, container_key) in keys_info.items():
|
|
|
|
keys_info[apex] = (payload_key, key_map.get(container_key, container_key))
|
|
|
|
|
|
|
|
# Apply all the --extra_apks options to override the container keys.
|
|
|
|
for apex, key in OPTIONS.extra_apks.items():
|
|
|
|
# Skip non-APEX containers.
|
|
|
|
if apex not in keys_info:
|
|
|
|
continue
|
2019-03-19 20:24:03 +01:00
|
|
|
if not key:
|
|
|
|
key = 'PRESIGNED'
|
2019-03-18 18:24:17 +01:00
|
|
|
keys_info[apex] = (keys_info[apex][0], key_map.get(key, key))
|
2019-03-15 17:37:01 +01:00
|
|
|
|
|
|
|
return keys_info
|
|
|
|
|
|
|
|
|
2018-06-19 21:19:35 +02:00
|
|
|
def GetApkFileInfo(filename, compressed_extension, skipped_prefixes):
|
2018-06-19 21:19:35 +02:00
|
|
|
"""Returns the APK info based on the given filename.
|
|
|
|
|
|
|
|
Checks if the given filename (with path) looks like an APK file, by taking the
|
2018-06-19 21:19:35 +02:00
|
|
|
compressed extension into consideration. If it appears to be an APK file,
|
|
|
|
further checks if the APK file should be skipped when signing, based on the
|
|
|
|
given path prefixes.
|
2018-06-19 21:19:35 +02:00
|
|
|
|
|
|
|
Args:
|
|
|
|
filename: Path to the file.
|
|
|
|
compressed_extension: The extension string of compressed APKs (e.g. ".gz"),
|
|
|
|
or None if there's no compressed APKs.
|
2018-06-19 21:19:35 +02:00
|
|
|
skipped_prefixes: A set/list/tuple of the path prefixes to be skipped.
|
2018-06-19 21:19:35 +02:00
|
|
|
|
|
|
|
Returns:
|
2018-06-19 21:19:35 +02:00
|
|
|
(is_apk, is_compressed, should_be_skipped): is_apk indicates whether the
|
|
|
|
given filename is an APK file. is_compressed indicates whether the APK file
|
|
|
|
is compressed (only meaningful when is_apk is True). should_be_skipped
|
|
|
|
indicates whether the filename matches any of the given prefixes to be
|
|
|
|
skipped.
|
2018-06-19 21:19:35 +02:00
|
|
|
|
|
|
|
Raises:
|
2018-06-19 21:19:35 +02:00
|
|
|
AssertionError: On invalid compressed_extension or skipped_prefixes inputs.
|
2018-06-19 21:19:35 +02:00
|
|
|
"""
|
|
|
|
assert compressed_extension is None or compressed_extension.startswith('.'), \
|
|
|
|
"Invalid compressed_extension arg: '{}'".format(compressed_extension)
|
|
|
|
|
2018-06-19 21:19:35 +02:00
|
|
|
# skipped_prefixes should be one of set/list/tuple types. Other types such as
|
|
|
|
# str shouldn't be accepted.
|
2019-03-15 17:33:43 +01:00
|
|
|
assert isinstance(skipped_prefixes, (set, list, tuple)), \
|
|
|
|
"Invalid skipped_prefixes input type: {}".format(type(skipped_prefixes))
|
2018-06-19 21:19:35 +02:00
|
|
|
|
2018-06-19 21:19:35 +02:00
|
|
|
compressed_apk_extension = (
|
|
|
|
".apk" + compressed_extension if compressed_extension else None)
|
|
|
|
is_apk = (filename.endswith(".apk") or
|
|
|
|
(compressed_apk_extension and
|
|
|
|
filename.endswith(compressed_apk_extension)))
|
|
|
|
if not is_apk:
|
2018-06-19 21:19:35 +02:00
|
|
|
return (False, False, False)
|
2018-06-19 21:19:35 +02:00
|
|
|
|
|
|
|
is_compressed = (compressed_apk_extension and
|
|
|
|
filename.endswith(compressed_apk_extension))
|
2018-06-19 21:19:35 +02:00
|
|
|
should_be_skipped = filename.startswith(tuple(skipped_prefixes))
|
|
|
|
return (True, is_compressed, should_be_skipped)
|
2018-06-19 21:19:35 +02:00
|
|
|
|
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
def CheckApkAndApexKeysAvailable(input_tf_zip, known_keys,
|
2019-03-19 20:24:03 +01:00
|
|
|
compressed_extension, apex_keys):
|
2019-03-15 17:37:01 +01:00
|
|
|
"""Checks that all the APKs and APEXes have keys specified.
|
2018-06-19 21:19:35 +02:00
|
|
|
|
|
|
|
Args:
|
|
|
|
input_tf_zip: An open target_files zip file.
|
2019-03-15 17:37:01 +01:00
|
|
|
known_keys: A set of APKs and APEXes that have known signing keys.
|
2018-06-19 21:19:35 +02:00
|
|
|
compressed_extension: The extension string of compressed APKs, such as
|
2019-03-15 17:37:01 +01:00
|
|
|
'.gz', or None if there's no compressed APKs.
|
2019-03-19 20:24:03 +01:00
|
|
|
apex_keys: A dict that contains the key mapping from APEX name to
|
|
|
|
(payload_key, container_key).
|
2018-06-19 21:19:35 +02:00
|
|
|
|
|
|
|
Raises:
|
2019-03-15 17:37:01 +01:00
|
|
|
AssertionError: On finding unknown APKs and APEXes.
|
2018-06-19 21:19:35 +02:00
|
|
|
"""
|
2019-03-15 17:37:01 +01:00
|
|
|
unknown_files = []
|
2015-03-24 03:13:21 +01:00
|
|
|
for info in input_tf_zip.infolist():
|
2019-03-15 17:37:01 +01:00
|
|
|
# Handle APEXes first, e.g. SYSTEM/apex/com.android.tzdata.apex.
|
|
|
|
if (info.filename.startswith('SYSTEM/apex') and
|
|
|
|
info.filename.endswith('.apex')):
|
|
|
|
name = os.path.basename(info.filename)
|
|
|
|
if name not in known_keys:
|
|
|
|
unknown_files.append(name)
|
|
|
|
continue
|
|
|
|
|
|
|
|
# And APKs.
|
2018-06-19 21:19:35 +02:00
|
|
|
(is_apk, is_compressed, should_be_skipped) = GetApkFileInfo(
|
|
|
|
info.filename, compressed_extension, OPTIONS.skip_apks_with_path_prefix)
|
|
|
|
if not is_apk or should_be_skipped:
|
2018-06-19 21:19:35 +02:00
|
|
|
continue
|
2019-03-15 17:37:01 +01:00
|
|
|
|
2018-06-19 21:19:35 +02:00
|
|
|
name = os.path.basename(info.filename)
|
|
|
|
if is_compressed:
|
|
|
|
name = name[:-len(compressed_extension)]
|
2019-03-15 17:37:01 +01:00
|
|
|
if name not in known_keys:
|
|
|
|
unknown_files.append(name)
|
2018-06-19 21:19:35 +02:00
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
assert not unknown_files, \
|
2018-06-19 21:19:35 +02:00
|
|
|
("No key specified for:\n {}\n"
|
|
|
|
"Use '-e <apkname>=' to specify a key (which may be an empty string to "
|
2019-03-15 17:37:01 +01:00
|
|
|
"not sign this apk).".format("\n ".join(unknown_files)))
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2019-03-19 20:24:03 +01:00
|
|
|
# For all the APEXes, double check that we won't have an APEX that has only
|
|
|
|
# one of the payload / container keys set.
|
|
|
|
if not apex_keys:
|
|
|
|
return
|
|
|
|
|
|
|
|
invalid_apexes = []
|
|
|
|
for info in input_tf_zip.infolist():
|
|
|
|
if (not info.filename.startswith('SYSTEM/apex') or
|
|
|
|
not info.filename.endswith('.apex')):
|
|
|
|
continue
|
|
|
|
|
|
|
|
name = os.path.basename(info.filename)
|
|
|
|
(payload_key, container_key) = apex_keys[name]
|
|
|
|
if ((payload_key in common.SPECIAL_CERT_STRINGS and
|
|
|
|
container_key not in common.SPECIAL_CERT_STRINGS) or
|
|
|
|
(payload_key not in common.SPECIAL_CERT_STRINGS and
|
|
|
|
container_key in common.SPECIAL_CERT_STRINGS)):
|
|
|
|
invalid_apexes.append(
|
|
|
|
"{}: payload_key {}, container_key {}".format(
|
|
|
|
name, payload_key, container_key))
|
|
|
|
|
|
|
|
assert not invalid_apexes, \
|
|
|
|
"Invalid APEX keys specified:\n {}\n".format(
|
|
|
|
"\n ".join(invalid_apexes))
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2017-08-14 15:49:21 +02:00
|
|
|
def SignApk(data, keyname, pw, platform_api_level, codename_to_api_level_map,
|
|
|
|
is_compressed):
|
2015-03-24 03:13:21 +01:00
|
|
|
unsigned = tempfile.NamedTemporaryFile()
|
|
|
|
unsigned.write(data)
|
|
|
|
unsigned.flush()
|
|
|
|
|
2017-08-14 15:49:21 +02:00
|
|
|
if is_compressed:
|
|
|
|
uncompressed = tempfile.NamedTemporaryFile()
|
2017-12-24 19:37:38 +01:00
|
|
|
with gzip.open(unsigned.name, "rb") as in_file, \
|
|
|
|
open(uncompressed.name, "wb") as out_file:
|
2017-08-14 15:49:21 +02:00
|
|
|
shutil.copyfileobj(in_file, out_file)
|
|
|
|
|
|
|
|
# Finally, close the "unsigned" file (which is gzip compressed), and then
|
|
|
|
# replace it with the uncompressed version.
|
|
|
|
#
|
|
|
|
# TODO(narayan): All this nastiness can be avoided if python 3.2 is in use,
|
|
|
|
# we could just gzip / gunzip in-memory buffers instead.
|
|
|
|
unsigned.close()
|
|
|
|
unsigned = uncompressed
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
signed = tempfile.NamedTemporaryFile()
|
|
|
|
|
2016-01-13 19:32:47 +01:00
|
|
|
# For pre-N builds, don't upgrade to SHA-256 JAR signatures based on the APK's
|
|
|
|
# minSdkVersion to avoid increasing incremental OTA update sizes. If an APK
|
|
|
|
# didn't change, we don't want its signature to change due to the switch
|
|
|
|
# from SHA-1 to SHA-256.
|
|
|
|
# By default, APK signer chooses SHA-256 signatures if the APK's minSdkVersion
|
|
|
|
# is 18 or higher. For pre-N builds we disable this mechanism by pretending
|
|
|
|
# that the APK's minSdkVersion is 1.
|
|
|
|
# For N+ builds, we let APK signer rely on the APK's minSdkVersion to
|
|
|
|
# determine whether to use SHA-256.
|
|
|
|
min_api_level = None
|
|
|
|
if platform_api_level > 23:
|
|
|
|
# Let APK signer choose whether to use SHA-1 or SHA-256, based on the APK's
|
|
|
|
# minSdkVersion attribute
|
|
|
|
min_api_level = None
|
|
|
|
else:
|
|
|
|
# Force APK signer to use SHA-1
|
|
|
|
min_api_level = 1
|
|
|
|
|
|
|
|
common.SignFile(unsigned.name, signed.name, keyname, pw,
|
2017-12-24 19:37:38 +01:00
|
|
|
min_api_level=min_api_level,
|
|
|
|
codename_to_api_level_map=codename_to_api_level_map)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
data = None
|
2017-08-14 15:49:21 +02:00
|
|
|
if is_compressed:
|
|
|
|
# Recompress the file after it has been signed.
|
|
|
|
compressed = tempfile.NamedTemporaryFile()
|
2017-12-24 19:37:38 +01:00
|
|
|
with open(signed.name, "rb") as in_file, \
|
|
|
|
gzip.open(compressed.name, "wb") as out_file:
|
2017-08-14 15:49:21 +02:00
|
|
|
shutil.copyfileobj(in_file, out_file)
|
|
|
|
|
|
|
|
data = compressed.read()
|
|
|
|
compressed.close()
|
|
|
|
else:
|
|
|
|
data = signed.read()
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
unsigned.close()
|
|
|
|
signed.close()
|
|
|
|
|
|
|
|
return data
|
|
|
|
|
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
def SignApex(apex_data, payload_key, container_key, container_pw,
|
|
|
|
codename_to_api_level_map, signing_args=None):
|
|
|
|
"""Signs the current APEX with the given payload/container keys.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
apex_data: Raw APEX data.
|
2019-04-02 06:25:05 +02:00
|
|
|
payload_key: The path to payload signing key (w/ extension).
|
2019-03-15 17:37:01 +01:00
|
|
|
container_key: The path to container signing key (w/o extension).
|
|
|
|
container_pw: The matching password of the container_key, or None.
|
|
|
|
codename_to_api_level_map: A dict that maps from codename to API level.
|
|
|
|
signing_args: Additional args to be passed to the payload signer.
|
|
|
|
|
|
|
|
Returns:
|
2019-04-02 06:25:05 +02:00
|
|
|
The path to the signed APEX file.
|
2019-03-15 17:37:01 +01:00
|
|
|
"""
|
|
|
|
apex_file = common.MakeTempFile(prefix='apex-', suffix='.apex')
|
|
|
|
with open(apex_file, 'wb') as apex_fp:
|
|
|
|
apex_fp.write(apex_data)
|
|
|
|
|
|
|
|
APEX_PAYLOAD_IMAGE = 'apex_payload.img'
|
2019-04-02 06:25:05 +02:00
|
|
|
APEX_PUBKEY = 'apex_pubkey'
|
2019-03-15 17:37:01 +01:00
|
|
|
|
2019-04-02 06:25:05 +02:00
|
|
|
# 1a. Extract and sign the APEX_PAYLOAD_IMAGE entry with the given
|
|
|
|
# payload_key.
|
2019-03-15 17:37:01 +01:00
|
|
|
payload_dir = common.MakeTempDir(prefix='apex-payload-')
|
|
|
|
with zipfile.ZipFile(apex_file) as apex_fd:
|
|
|
|
payload_file = apex_fd.extract(APEX_PAYLOAD_IMAGE, payload_dir)
|
|
|
|
|
|
|
|
payload_info = apex_utils.ParseApexPayloadInfo(payload_file)
|
|
|
|
apex_utils.SignApexPayload(
|
|
|
|
payload_file,
|
|
|
|
payload_key,
|
|
|
|
payload_info['apex.key'],
|
|
|
|
payload_info['Algorithm'],
|
|
|
|
payload_info['Salt'],
|
|
|
|
signing_args)
|
|
|
|
|
2019-04-02 06:25:05 +02:00
|
|
|
# 1b. Update the embedded payload public key.
|
|
|
|
payload_public_key = common.ExtractAvbPublicKey(payload_key)
|
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
common.ZipDelete(apex_file, APEX_PAYLOAD_IMAGE)
|
2019-04-02 06:25:05 +02:00
|
|
|
common.ZipDelete(apex_file, APEX_PUBKEY)
|
2019-03-15 17:37:01 +01:00
|
|
|
apex_zip = zipfile.ZipFile(apex_file, 'a')
|
|
|
|
common.ZipWrite(apex_zip, payload_file, arcname=APEX_PAYLOAD_IMAGE)
|
2019-04-02 06:25:05 +02:00
|
|
|
common.ZipWrite(apex_zip, payload_public_key, arcname=APEX_PUBKEY)
|
2019-03-15 17:37:01 +01:00
|
|
|
common.ZipClose(apex_zip)
|
|
|
|
|
2019-03-23 07:16:58 +01:00
|
|
|
# 2. Align the files at page boundary (same as in apexer).
|
|
|
|
aligned_apex = common.MakeTempFile(
|
|
|
|
prefix='apex-container-', suffix='.apex')
|
|
|
|
common.RunAndCheckOutput(
|
|
|
|
['zipalign', '-f', '4096', apex_file, aligned_apex])
|
|
|
|
|
|
|
|
# 3. Sign the APEX container with container_key.
|
2019-03-15 17:37:01 +01:00
|
|
|
signed_apex = common.MakeTempFile(prefix='apex-container-', suffix='.apex')
|
2019-03-23 07:16:58 +01:00
|
|
|
|
|
|
|
# Specify the 4K alignment when calling SignApk.
|
|
|
|
extra_signapk_args = OPTIONS.extra_signapk_args[:]
|
|
|
|
extra_signapk_args.extend(['-a', '4096'])
|
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
common.SignFile(
|
2019-03-23 07:16:58 +01:00
|
|
|
aligned_apex,
|
2019-03-15 17:37:01 +01:00
|
|
|
signed_apex,
|
|
|
|
container_key,
|
|
|
|
container_pw,
|
2019-03-23 07:16:58 +01:00
|
|
|
codename_to_api_level_map=codename_to_api_level_map,
|
|
|
|
extra_signapk_args=extra_signapk_args)
|
2019-03-15 17:37:01 +01:00
|
|
|
|
2019-04-02 06:25:05 +02:00
|
|
|
return signed_apex
|
2019-03-15 17:37:01 +01:00
|
|
|
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
def ProcessTargetFiles(input_tf_zip, output_tf_zip, misc_info,
|
2019-03-15 17:37:01 +01:00
|
|
|
apk_keys, apex_keys, key_passwords,
|
|
|
|
platform_api_level, codename_to_api_level_map,
|
2017-08-14 15:49:21 +02:00
|
|
|
compressed_extension):
|
2018-06-19 21:19:35 +02:00
|
|
|
# maxsize measures the maximum filename length, including the ones to be
|
|
|
|
# skipped.
|
2017-12-24 19:37:38 +01:00
|
|
|
maxsize = max(
|
|
|
|
[len(os.path.basename(i.filename)) for i in input_tf_zip.infolist()
|
2018-06-19 21:19:35 +02:00
|
|
|
if GetApkFileInfo(i.filename, compressed_extension, [])[0]])
|
2016-06-16 23:41:24 +02:00
|
|
|
system_root_image = misc_info.get("system_root_image") == "true"
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
for info in input_tf_zip.infolist():
|
2018-06-19 21:19:35 +02:00
|
|
|
filename = info.filename
|
|
|
|
if filename.startswith("IMAGES/"):
|
2015-03-24 03:13:21 +01:00
|
|
|
continue
|
|
|
|
|
2019-01-11 21:37:35 +01:00
|
|
|
# Skip split super images, which will be re-generated during signing.
|
|
|
|
if filename.startswith("OTA/") and filename.endswith(".img"):
|
|
|
|
continue
|
|
|
|
|
2018-06-19 21:19:35 +02:00
|
|
|
data = input_tf_zip.read(filename)
|
2015-03-24 03:13:21 +01:00
|
|
|
out_info = copy.copy(info)
|
2018-06-19 21:19:35 +02:00
|
|
|
(is_apk, is_compressed, should_be_skipped) = GetApkFileInfo(
|
|
|
|
filename, compressed_extension, OPTIONS.skip_apks_with_path_prefix)
|
|
|
|
|
|
|
|
if is_apk and should_be_skipped:
|
|
|
|
# Copy skipped APKs verbatim.
|
|
|
|
print(
|
|
|
|
"NOT signing: %s\n"
|
|
|
|
" (skipped due to matching prefix)" % (filename,))
|
|
|
|
common.ZipWriteStr(output_tf_zip, out_info, data)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2015-07-22 21:33:18 +02:00
|
|
|
# Sign APKs.
|
2018-06-19 21:19:35 +02:00
|
|
|
elif is_apk:
|
2018-06-19 21:19:35 +02:00
|
|
|
name = os.path.basename(filename)
|
2017-08-14 15:49:21 +02:00
|
|
|
if is_compressed:
|
|
|
|
name = name[:-len(compressed_extension)]
|
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
key = apk_keys[name]
|
2015-03-24 03:13:21 +01:00
|
|
|
if key not in common.SPECIAL_CERT_STRINGS:
|
2017-12-24 19:37:38 +01:00
|
|
|
print(" signing: %-*s (%s)" % (maxsize, name, key))
|
2016-01-13 19:32:47 +01:00
|
|
|
signed_data = SignApk(data, key, key_passwords[key], platform_api_level,
|
2017-12-24 19:37:38 +01:00
|
|
|
codename_to_api_level_map, is_compressed)
|
2015-04-01 20:21:55 +02:00
|
|
|
common.ZipWriteStr(output_tf_zip, out_info, signed_data)
|
2015-03-24 03:13:21 +01:00
|
|
|
else:
|
|
|
|
# an APK we're not supposed to sign.
|
2018-06-19 21:19:35 +02:00
|
|
|
print(
|
|
|
|
"NOT signing: %s\n"
|
|
|
|
" (skipped due to special cert string)" % (name,))
|
2015-04-01 20:21:55 +02:00
|
|
|
common.ZipWriteStr(output_tf_zip, out_info, data)
|
2016-06-16 23:41:24 +02:00
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
# Sign bundled APEX files.
|
|
|
|
elif filename.startswith("SYSTEM/apex") and filename.endswith(".apex"):
|
|
|
|
name = os.path.basename(filename)
|
|
|
|
payload_key, container_key = apex_keys[name]
|
|
|
|
|
2019-03-19 20:24:03 +01:00
|
|
|
# We've asserted not having a case with only one of them PRESIGNED.
|
|
|
|
if (payload_key not in common.SPECIAL_CERT_STRINGS and
|
|
|
|
container_key not in common.SPECIAL_CERT_STRINGS):
|
|
|
|
print(" signing: %-*s container (%s)" % (
|
|
|
|
maxsize, name, container_key))
|
|
|
|
print(" : %-*s payload (%s)" % (
|
|
|
|
maxsize, name, payload_key))
|
|
|
|
|
2019-04-02 06:25:05 +02:00
|
|
|
signed_apex = SignApex(
|
2019-03-19 20:24:03 +01:00
|
|
|
data,
|
|
|
|
payload_key,
|
|
|
|
container_key,
|
|
|
|
key_passwords[container_key],
|
|
|
|
codename_to_api_level_map,
|
|
|
|
OPTIONS.avb_extra_args.get('apex'))
|
|
|
|
common.ZipWrite(output_tf_zip, signed_apex, filename)
|
2019-03-15 17:37:01 +01:00
|
|
|
|
2019-03-19 20:24:03 +01:00
|
|
|
else:
|
|
|
|
print(
|
|
|
|
"NOT signing: %s\n"
|
|
|
|
" (skipped due to special cert string)" % (name,))
|
|
|
|
common.ZipWriteStr(output_tf_zip, out_info, data)
|
2019-03-15 17:37:01 +01:00
|
|
|
|
|
|
|
# AVB public keys for the installed APEXes, which will be updated later.
|
|
|
|
elif (os.path.dirname(filename) == 'SYSTEM/etc/security/apex' and
|
|
|
|
filename != 'SYSTEM/etc/security/apex/'):
|
|
|
|
continue
|
|
|
|
|
2016-06-16 23:41:24 +02:00
|
|
|
# System properties.
|
2018-06-19 21:19:35 +02:00
|
|
|
elif filename in ("SYSTEM/build.prop",
|
|
|
|
"VENDOR/build.prop",
|
|
|
|
"SYSTEM/etc/prop.default",
|
|
|
|
"BOOT/RAMDISK/prop.default",
|
|
|
|
"BOOT/RAMDISK/default.prop", # legacy
|
|
|
|
"ROOT/default.prop", # legacy
|
|
|
|
"RECOVERY/RAMDISK/prop.default",
|
|
|
|
"RECOVERY/RAMDISK/default.prop"): # legacy
|
|
|
|
print("Rewriting %s:" % (filename,))
|
2017-05-01 15:56:26 +02:00
|
|
|
if stat.S_ISLNK(info.external_attr >> 16):
|
|
|
|
new_data = data
|
|
|
|
else:
|
2017-12-08 23:42:16 +01:00
|
|
|
new_data = RewriteProps(data)
|
2015-04-01 20:21:55 +02:00
|
|
|
common.ZipWriteStr(output_tf_zip, out_info, new_data)
|
2016-06-16 23:41:24 +02:00
|
|
|
|
2017-12-05 02:16:36 +01:00
|
|
|
# Replace the certs in *mac_permissions.xml (there could be multiple, such
|
|
|
|
# as {system,vendor}/etc/selinux/{plat,nonplat}_mac_permissions.xml).
|
2018-06-19 21:19:35 +02:00
|
|
|
elif filename.endswith("mac_permissions.xml"):
|
|
|
|
print("Rewriting %s with new keys." % (filename,))
|
2015-03-24 03:13:21 +01:00
|
|
|
new_data = ReplaceCerts(data)
|
2015-04-01 20:21:55 +02:00
|
|
|
common.ZipWriteStr(output_tf_zip, out_info, new_data)
|
2016-06-16 23:41:24 +02:00
|
|
|
|
2017-05-23 23:51:02 +02:00
|
|
|
# Ask add_img_to_target_files to rebuild the recovery patch if needed.
|
2018-06-19 21:19:35 +02:00
|
|
|
elif filename in ("SYSTEM/recovery-from-boot.p",
|
|
|
|
"SYSTEM/etc/recovery.img",
|
|
|
|
"SYSTEM/bin/install-recovery.sh"):
|
2017-05-23 23:51:02 +02:00
|
|
|
OPTIONS.rebuild_recovery = True
|
2016-06-16 23:41:24 +02:00
|
|
|
|
2018-10-19 23:34:15 +02:00
|
|
|
# Don't copy OTA certs if we're replacing them.
|
2018-08-18 01:27:01 +02:00
|
|
|
elif (
|
|
|
|
OPTIONS.replace_ota_keys and
|
|
|
|
filename in (
|
2018-10-19 23:34:15 +02:00
|
|
|
"BOOT/RAMDISK/system/etc/security/otacerts.zip",
|
2018-08-18 01:27:01 +02:00
|
|
|
"BOOT/RAMDISK/system/etc/update_engine/update-payload-key.pub.pem",
|
2018-10-19 23:34:15 +02:00
|
|
|
"RECOVERY/RAMDISK/system/etc/security/otacerts.zip",
|
2018-08-18 01:27:01 +02:00
|
|
|
"SYSTEM/etc/security/otacerts.zip",
|
|
|
|
"SYSTEM/etc/update_engine/update-payload-key.pub.pem")):
|
2015-03-24 03:13:21 +01:00
|
|
|
pass
|
2016-06-16 23:41:24 +02:00
|
|
|
|
2017-06-05 20:55:16 +02:00
|
|
|
# Skip META/misc_info.txt since we will write back the new values later.
|
2018-06-19 21:19:35 +02:00
|
|
|
elif filename == "META/misc_info.txt":
|
2015-03-24 03:13:21 +01:00
|
|
|
pass
|
2016-06-18 02:01:22 +02:00
|
|
|
|
|
|
|
# Skip verity public key if we will replace it.
|
2015-03-24 03:13:21 +01:00
|
|
|
elif (OPTIONS.replace_verity_public_key and
|
2018-06-19 21:19:35 +02:00
|
|
|
filename in ("BOOT/RAMDISK/verity_key",
|
|
|
|
"ROOT/verity_key")):
|
2015-03-24 03:13:21 +01:00
|
|
|
pass
|
2016-06-16 23:41:24 +02:00
|
|
|
|
2016-06-18 02:01:22 +02:00
|
|
|
# Skip verity keyid (for system_root_image use) if we will replace it.
|
2018-06-19 21:19:35 +02:00
|
|
|
elif OPTIONS.replace_verity_keyid and filename == "BOOT/cmdline":
|
2016-06-17 04:58:44 +02:00
|
|
|
pass
|
|
|
|
|
2016-08-12 03:04:27 +02:00
|
|
|
# Skip the care_map as we will regenerate the system/vendor images.
|
2018-09-15 01:24:41 +02:00
|
|
|
elif filename == "META/care_map.pb" or filename == "META/care_map.txt":
|
2016-08-12 03:04:27 +02:00
|
|
|
pass
|
|
|
|
|
2019-02-27 03:15:51 +01:00
|
|
|
# Updates system_other.avbpubkey in /product/etc/.
|
|
|
|
elif filename in (
|
|
|
|
"PRODUCT/etc/security/avb/system_other.avbpubkey",
|
|
|
|
"SYSTEM/product/etc/security/avb/system_other.avbpubkey"):
|
|
|
|
# Only update system_other's public key, if the corresponding signing
|
|
|
|
# key is specified via --avb_system_other_key.
|
|
|
|
signing_key = OPTIONS.avb_keys.get("system_other")
|
|
|
|
if signing_key:
|
|
|
|
public_key = common.ExtractAvbPublicKey(signing_key)
|
|
|
|
print(" Rewriting AVB public key of system_other in /product")
|
|
|
|
common.ZipWrite(output_tf_zip, public_key, filename)
|
|
|
|
|
2019-04-23 06:28:44 +02:00
|
|
|
# Should NOT sign boot-debug.img.
|
|
|
|
elif filename in (
|
|
|
|
"BOOT/RAMDISK/force_debuggable",
|
|
|
|
"RECOVERY/RAMDISK/force_debuggable"
|
|
|
|
"RECOVERY/RAMDISK/first_stage_ramdisk/force_debuggable"):
|
|
|
|
raise common.ExternalError("debuggable boot.img cannot be signed")
|
|
|
|
|
2016-06-16 23:41:24 +02:00
|
|
|
# A non-APK file; copy it verbatim.
|
2015-03-24 03:13:21 +01:00
|
|
|
else:
|
2015-04-01 20:21:55 +02:00
|
|
|
common.ZipWriteStr(output_tf_zip, out_info, data)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
if OPTIONS.replace_ota_keys:
|
2017-05-23 23:51:02 +02:00
|
|
|
ReplaceOtaKeys(input_tf_zip, output_tf_zip, misc_info)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2017-06-05 20:55:16 +02:00
|
|
|
# Replace the keyid string in misc_info dict.
|
2016-06-18 02:01:22 +02:00
|
|
|
if OPTIONS.replace_verity_private_key:
|
2017-06-05 20:55:16 +02:00
|
|
|
ReplaceVerityPrivateKey(misc_info, OPTIONS.replace_verity_private_key[1])
|
2016-06-18 02:01:22 +02:00
|
|
|
|
|
|
|
if OPTIONS.replace_verity_public_key:
|
2017-12-24 19:37:38 +01:00
|
|
|
dest = "ROOT/verity_key" if system_root_image else "BOOT/RAMDISK/verity_key"
|
2016-06-18 02:01:22 +02:00
|
|
|
# We are replacing the one in boot image only, since the one under
|
|
|
|
# recovery won't ever be needed.
|
2017-05-23 23:51:02 +02:00
|
|
|
ReplaceVerityPublicKey(
|
2016-06-18 02:01:22 +02:00
|
|
|
output_tf_zip, dest, OPTIONS.replace_verity_public_key[1])
|
|
|
|
|
|
|
|
# Replace the keyid string in BOOT/cmdline.
|
|
|
|
if OPTIONS.replace_verity_keyid:
|
2017-05-23 23:51:02 +02:00
|
|
|
ReplaceVerityKeyId(input_tf_zip, output_tf_zip,
|
|
|
|
OPTIONS.replace_verity_keyid[1])
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2017-06-20 00:48:02 +02:00
|
|
|
# Replace the AVB signing keys, if any.
|
|
|
|
ReplaceAvbSigningKeys(misc_info)
|
|
|
|
|
2017-06-05 20:55:16 +02:00
|
|
|
# Write back misc_info with the latest values.
|
|
|
|
ReplaceMiscInfoTxt(input_tf_zip, output_tf_zip, misc_info)
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
def ReplaceCerts(data):
|
2017-12-05 02:16:36 +01:00
|
|
|
"""Replaces all the occurences of X.509 certs with the new ones.
|
|
|
|
|
|
|
|
The mapping info is read from OPTIONS.key_map. Non-existent certificate will
|
|
|
|
be skipped. After the replacement, it additionally checks for duplicate
|
|
|
|
entries, which would otherwise fail the policy loading code in
|
|
|
|
frameworks/base/services/core/java/com/android/server/pm/SELinuxMMAC.java.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
data: Input string that contains a set of X.509 certs.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
A string after the replacement.
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
AssertionError: On finding duplicate entries.
|
|
|
|
"""
|
2015-03-24 03:13:21 +01:00
|
|
|
for old, new in OPTIONS.key_map.iteritems():
|
2017-12-05 02:16:36 +01:00
|
|
|
if OPTIONS.verbose:
|
|
|
|
print(" Replacing %s.x509.pem with %s.x509.pem" % (old, new))
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
try:
|
2017-12-05 02:16:36 +01:00
|
|
|
with open(old + ".x509.pem") as old_fp:
|
|
|
|
old_cert16 = base64.b16encode(
|
|
|
|
common.ParseCertificate(old_fp.read())).lower()
|
|
|
|
with open(new + ".x509.pem") as new_fp:
|
|
|
|
new_cert16 = base64.b16encode(
|
|
|
|
common.ParseCertificate(new_fp.read())).lower()
|
2015-03-24 03:13:21 +01:00
|
|
|
except IOError as e:
|
2017-12-05 02:16:36 +01:00
|
|
|
if OPTIONS.verbose or e.errno != errno.ENOENT:
|
|
|
|
print(" Error accessing %s: %s.\nSkip replacing %s.x509.pem with "
|
|
|
|
"%s.x509.pem." % (e.filename, e.strerror, old, new))
|
|
|
|
continue
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2017-12-05 02:16:36 +01:00
|
|
|
# Only match entire certs.
|
|
|
|
pattern = "\\b" + old_cert16 + "\\b"
|
|
|
|
(data, num) = re.subn(pattern, new_cert16, data, flags=re.IGNORECASE)
|
|
|
|
|
|
|
|
if OPTIONS.verbose:
|
|
|
|
print(" Replaced %d occurence(s) of %s.x509.pem with %s.x509.pem" % (
|
|
|
|
num, old, new))
|
|
|
|
|
|
|
|
# Verify that there're no duplicate entries after the replacement. Note that
|
|
|
|
# it's only checking entries with global seinfo at the moment (i.e. ignoring
|
|
|
|
# the ones with inner packages). (Bug: 69479366)
|
|
|
|
root = ElementTree.fromstring(data)
|
|
|
|
signatures = [signer.attrib['signature'] for signer in root.findall('signer')]
|
|
|
|
assert len(signatures) == len(set(signatures)), \
|
|
|
|
"Found duplicate entries after cert replacement: {}".format(data)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
return data
|
|
|
|
|
|
|
|
|
|
|
|
def EditTags(tags):
|
2017-12-08 23:42:16 +01:00
|
|
|
"""Applies the edits to the tag string as specified in OPTIONS.tag_changes.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
tags: The input string that contains comma-separated tags.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The updated tags (comma-separated and sorted).
|
|
|
|
"""
|
2015-03-24 03:13:21 +01:00
|
|
|
tags = set(tags.split(","))
|
|
|
|
for ch in OPTIONS.tag_changes:
|
|
|
|
if ch[0] == "-":
|
|
|
|
tags.discard(ch[1:])
|
|
|
|
elif ch[0] == "+":
|
|
|
|
tags.add(ch[1:])
|
|
|
|
return ",".join(sorted(tags))
|
|
|
|
|
|
|
|
|
2017-12-08 23:42:16 +01:00
|
|
|
def RewriteProps(data):
|
|
|
|
"""Rewrites the system properties in the given string.
|
|
|
|
|
|
|
|
Each property is expected in 'key=value' format. The properties that contain
|
|
|
|
build tags (i.e. test-keys, dev-keys) will be updated accordingly by calling
|
|
|
|
EditTags().
|
|
|
|
|
|
|
|
Args:
|
|
|
|
data: Input string, separated by newlines.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The string with modified properties.
|
|
|
|
"""
|
2015-03-24 03:13:21 +01:00
|
|
|
output = []
|
|
|
|
for line in data.split("\n"):
|
|
|
|
line = line.strip()
|
|
|
|
original_line = line
|
|
|
|
if line and line[0] != '#' and "=" in line:
|
|
|
|
key, value = line.split("=", 1)
|
2017-12-08 23:42:16 +01:00
|
|
|
if key in ("ro.build.fingerprint", "ro.build.thumbprint",
|
|
|
|
"ro.vendor.build.fingerprint", "ro.vendor.build.thumbprint"):
|
2015-03-24 03:13:21 +01:00
|
|
|
pieces = value.split("/")
|
|
|
|
pieces[-1] = EditTags(pieces[-1])
|
|
|
|
value = "/".join(pieces)
|
2015-09-12 00:27:56 +02:00
|
|
|
elif key == "ro.bootimage.build.fingerprint":
|
|
|
|
pieces = value.split("/")
|
|
|
|
pieces[-1] = EditTags(pieces[-1])
|
|
|
|
value = "/".join(pieces)
|
2015-03-24 03:13:21 +01:00
|
|
|
elif key == "ro.build.description":
|
|
|
|
pieces = value.split(" ")
|
|
|
|
assert len(pieces) == 5
|
|
|
|
pieces[-1] = EditTags(pieces[-1])
|
|
|
|
value = " ".join(pieces)
|
|
|
|
elif key == "ro.build.tags":
|
|
|
|
value = EditTags(value)
|
|
|
|
elif key == "ro.build.display.id":
|
|
|
|
# change, eg, "JWR66N dev-keys" to "JWR66N"
|
|
|
|
value = value.split()
|
|
|
|
if len(value) > 1 and value[-1].endswith("-keys"):
|
|
|
|
value.pop()
|
|
|
|
value = " ".join(value)
|
|
|
|
line = key + "=" + value
|
|
|
|
if line != original_line:
|
2017-12-24 19:37:38 +01:00
|
|
|
print(" replace: ", original_line)
|
|
|
|
print(" with: ", line)
|
2015-03-24 03:13:21 +01:00
|
|
|
output.append(line)
|
|
|
|
return "\n".join(output) + "\n"
|
|
|
|
|
|
|
|
|
2018-10-19 23:34:15 +02:00
|
|
|
def WriteOtacerts(output_zip, filename, keys):
|
|
|
|
"""Constructs a zipfile from given keys; and writes it to output_zip.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
output_zip: The output target_files zip.
|
|
|
|
filename: The archive name in the output zip.
|
|
|
|
keys: A list of public keys to use during OTA package verification.
|
|
|
|
"""
|
|
|
|
|
|
|
|
try:
|
|
|
|
from StringIO import StringIO
|
|
|
|
except ImportError:
|
|
|
|
from io import StringIO
|
|
|
|
temp_file = StringIO()
|
|
|
|
certs_zip = zipfile.ZipFile(temp_file, "w")
|
|
|
|
for k in keys:
|
|
|
|
common.ZipWrite(certs_zip, k)
|
|
|
|
common.ZipClose(certs_zip)
|
|
|
|
common.ZipWriteStr(output_zip, filename, temp_file.getvalue())
|
|
|
|
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
def ReplaceOtaKeys(input_tf_zip, output_tf_zip, misc_info):
|
|
|
|
try:
|
|
|
|
keylist = input_tf_zip.read("META/otakeys.txt").split()
|
|
|
|
except KeyError:
|
|
|
|
raise common.ExternalError("can't read META/otakeys.txt from input")
|
|
|
|
|
2017-11-09 19:10:10 +01:00
|
|
|
extra_recovery_keys = misc_info.get("extra_recovery_keys")
|
2015-03-24 03:13:21 +01:00
|
|
|
if extra_recovery_keys:
|
|
|
|
extra_recovery_keys = [OPTIONS.key_map.get(k, k) + ".x509.pem"
|
|
|
|
for k in extra_recovery_keys.split()]
|
|
|
|
if extra_recovery_keys:
|
2017-12-24 19:37:38 +01:00
|
|
|
print("extra recovery-only key(s): " + ", ".join(extra_recovery_keys))
|
2015-03-24 03:13:21 +01:00
|
|
|
else:
|
|
|
|
extra_recovery_keys = []
|
|
|
|
|
|
|
|
mapped_keys = []
|
|
|
|
for k in keylist:
|
|
|
|
m = re.match(r"^(.*)\.x509\.pem$", k)
|
|
|
|
if not m:
|
|
|
|
raise common.ExternalError(
|
|
|
|
"can't parse \"%s\" from META/otakeys.txt" % (k,))
|
|
|
|
k = m.group(1)
|
|
|
|
mapped_keys.append(OPTIONS.key_map.get(k, k) + ".x509.pem")
|
|
|
|
|
|
|
|
if mapped_keys:
|
2017-12-24 19:37:38 +01:00
|
|
|
print("using:\n ", "\n ".join(mapped_keys))
|
|
|
|
print("for OTA package verification")
|
2015-03-24 03:13:21 +01:00
|
|
|
else:
|
|
|
|
devkey = misc_info.get("default_system_dev_certificate",
|
2019-04-10 06:35:37 +02:00
|
|
|
"build/make/target/product/security/testkey")
|
2017-11-09 19:10:10 +01:00
|
|
|
mapped_devkey = OPTIONS.key_map.get(devkey, devkey)
|
|
|
|
if mapped_devkey != devkey:
|
|
|
|
misc_info["default_system_dev_certificate"] = mapped_devkey
|
|
|
|
mapped_keys.append(mapped_devkey + ".x509.pem")
|
2016-06-16 23:41:24 +02:00
|
|
|
print("META/otakeys.txt has no keys; using %s for OTA package"
|
|
|
|
" verification." % (mapped_keys[0],))
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2018-10-19 23:34:15 +02:00
|
|
|
# recovery now uses the same x509.pem version of the keys.
|
2015-03-24 03:13:21 +01:00
|
|
|
# extra_recovery_keys are used only in recovery.
|
2018-09-20 20:04:37 +02:00
|
|
|
if misc_info.get("recovery_as_boot") == "true":
|
2018-10-19 23:34:15 +02:00
|
|
|
recovery_keys_location = "BOOT/RAMDISK/system/etc/security/otacerts.zip"
|
2016-06-16 23:41:24 +02:00
|
|
|
else:
|
2018-10-19 23:34:15 +02:00
|
|
|
recovery_keys_location = "RECOVERY/RAMDISK/system/etc/security/otacerts.zip"
|
|
|
|
|
|
|
|
WriteOtacerts(output_tf_zip, recovery_keys_location,
|
|
|
|
mapped_keys + extra_recovery_keys)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
# SystemUpdateActivity uses the x509.pem version of the keys, but
|
|
|
|
# put into a zipfile system/etc/security/otacerts.zip.
|
|
|
|
# We DO NOT include the extra_recovery_keys (if any) here.
|
2018-10-19 23:34:15 +02:00
|
|
|
WriteOtacerts(output_tf_zip, "SYSTEM/etc/security/otacerts.zip", mapped_keys)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2016-06-16 23:41:24 +02:00
|
|
|
# For A/B devices, update the payload verification key.
|
|
|
|
if misc_info.get("ab_update") == "true":
|
|
|
|
# Unlike otacerts.zip that may contain multiple keys, we can only specify
|
|
|
|
# ONE payload verification key.
|
|
|
|
if len(mapped_keys) > 1:
|
|
|
|
print("\n WARNING: Found more than one OTA keys; Using the first one"
|
|
|
|
" as payload verification key.\n\n")
|
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
print("Using %s for payload verification." % (mapped_keys[0],))
|
2018-02-04 21:13:35 +01:00
|
|
|
pubkey = common.ExtractPublicKey(mapped_keys[0])
|
2016-07-07 00:28:59 +02:00
|
|
|
common.ZipWriteStr(
|
2016-06-16 23:41:24 +02:00
|
|
|
output_tf_zip,
|
2016-07-07 00:28:59 +02:00
|
|
|
"SYSTEM/etc/update_engine/update-payload-key.pub.pem",
|
|
|
|
pubkey)
|
2016-08-05 01:06:12 +02:00
|
|
|
common.ZipWriteStr(
|
|
|
|
output_tf_zip,
|
2018-08-18 01:27:01 +02:00
|
|
|
"BOOT/RAMDISK/system/etc/update_engine/update-payload-key.pub.pem",
|
2016-08-05 01:06:12 +02:00
|
|
|
pubkey)
|
2016-06-16 23:41:24 +02:00
|
|
|
|
2016-06-18 02:01:22 +02:00
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
def ReplaceVerityPublicKey(output_zip, filename, key_path):
|
|
|
|
"""Replaces the verity public key at the given path in the given zip.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
output_zip: The output target_files zip.
|
|
|
|
filename: The archive name in the output zip.
|
|
|
|
key_path: The path to the public key.
|
|
|
|
"""
|
|
|
|
print("Replacing verity public key with %s" % (key_path,))
|
|
|
|
common.ZipWrite(output_zip, key_path, arcname=filename)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2016-06-18 02:01:22 +02:00
|
|
|
|
2017-06-05 20:55:16 +02:00
|
|
|
def ReplaceVerityPrivateKey(misc_info, key_path):
|
2017-12-24 19:37:38 +01:00
|
|
|
"""Replaces the verity private key in misc_info dict.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
misc_info: The info dict.
|
|
|
|
key_path: The path to the private key in PKCS#8 format.
|
|
|
|
"""
|
|
|
|
print("Replacing verity private key with %s" % (key_path,))
|
2015-03-24 03:13:21 +01:00
|
|
|
misc_info["verity_key"] = key_path
|
|
|
|
|
2016-06-18 02:01:22 +02:00
|
|
|
|
2017-12-24 08:44:48 +01:00
|
|
|
def ReplaceVerityKeyId(input_zip, output_zip, key_path):
|
|
|
|
"""Replaces the veritykeyid parameter in BOOT/cmdline.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
input_zip: The input target_files zip, which should be already open.
|
|
|
|
output_zip: The output target_files zip, which should be already open and
|
|
|
|
writable.
|
|
|
|
key_path: The path to the PEM encoded X.509 certificate.
|
|
|
|
"""
|
|
|
|
in_cmdline = input_zip.read("BOOT/cmdline")
|
|
|
|
# Copy in_cmdline to output_zip if veritykeyid is not present.
|
2016-06-17 04:58:44 +02:00
|
|
|
if "veritykeyid" not in in_cmdline:
|
2017-12-24 08:44:48 +01:00
|
|
|
common.ZipWriteStr(output_zip, "BOOT/cmdline", in_cmdline)
|
|
|
|
return
|
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
out_buffer = []
|
2016-06-17 04:58:44 +02:00
|
|
|
for param in in_cmdline.split():
|
2017-12-24 08:44:48 +01:00
|
|
|
if "veritykeyid" not in param:
|
2017-12-24 19:37:38 +01:00
|
|
|
out_buffer.append(param)
|
2017-12-24 08:44:48 +01:00
|
|
|
continue
|
2016-06-17 04:58:44 +02:00
|
|
|
|
2017-12-24 08:44:48 +01:00
|
|
|
# Extract keyid using openssl command.
|
|
|
|
p = common.Run(["openssl", "x509", "-in", key_path, "-text"],
|
2018-02-20 19:05:46 +01:00
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
2017-12-24 08:44:48 +01:00
|
|
|
keyid, stderr = p.communicate()
|
|
|
|
assert p.returncode == 0, "Failed to dump certificate: {}".format(stderr)
|
|
|
|
keyid = re.search(
|
|
|
|
r'keyid:([0-9a-fA-F:]*)', keyid).group(1).replace(':', '').lower()
|
|
|
|
print("Replacing verity keyid with {}".format(keyid))
|
|
|
|
out_buffer.append("veritykeyid=id:%s" % (keyid,))
|
|
|
|
|
|
|
|
out_cmdline = ' '.join(out_buffer).strip() + '\n'
|
|
|
|
common.ZipWriteStr(output_zip, "BOOT/cmdline", out_cmdline)
|
2017-06-05 20:55:16 +02:00
|
|
|
|
|
|
|
|
|
|
|
def ReplaceMiscInfoTxt(input_zip, output_zip, misc_info):
|
|
|
|
"""Replaces META/misc_info.txt.
|
|
|
|
|
|
|
|
Only writes back the ones in the original META/misc_info.txt. Because the
|
|
|
|
current in-memory dict contains additional items computed at runtime.
|
|
|
|
"""
|
|
|
|
misc_info_old = common.LoadDictionaryFromLines(
|
|
|
|
input_zip.read('META/misc_info.txt').split('\n'))
|
|
|
|
items = []
|
|
|
|
for key in sorted(misc_info):
|
|
|
|
if key in misc_info_old:
|
|
|
|
items.append('%s=%s' % (key, misc_info[key]))
|
|
|
|
common.ZipWriteStr(output_zip, "META/misc_info.txt", '\n'.join(items))
|
2016-06-17 04:58:44 +02:00
|
|
|
|
2016-06-18 02:01:22 +02:00
|
|
|
|
2017-06-20 00:48:02 +02:00
|
|
|
def ReplaceAvbSigningKeys(misc_info):
|
|
|
|
"""Replaces the AVB signing keys."""
|
|
|
|
|
|
|
|
AVB_FOOTER_ARGS_BY_PARTITION = {
|
2017-12-24 19:37:38 +01:00
|
|
|
'boot' : 'avb_boot_add_hash_footer_args',
|
|
|
|
'dtbo' : 'avb_dtbo_add_hash_footer_args',
|
|
|
|
'recovery' : 'avb_recovery_add_hash_footer_args',
|
|
|
|
'system' : 'avb_system_add_hashtree_footer_args',
|
2019-02-27 03:15:51 +01:00
|
|
|
'system_other' : 'avb_system_other_add_hashtree_footer_args',
|
2017-12-24 19:37:38 +01:00
|
|
|
'vendor' : 'avb_vendor_add_hashtree_footer_args',
|
|
|
|
'vbmeta' : 'avb_vbmeta_args',
|
2017-06-20 00:48:02 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
def ReplaceAvbPartitionSigningKey(partition):
|
|
|
|
key = OPTIONS.avb_keys.get(partition)
|
|
|
|
if not key:
|
|
|
|
return
|
|
|
|
|
|
|
|
algorithm = OPTIONS.avb_algorithms.get(partition)
|
|
|
|
assert algorithm, 'Missing AVB signing algorithm for %s' % (partition,)
|
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
print('Replacing AVB signing key for %s with "%s" (%s)' % (
|
|
|
|
partition, key, algorithm))
|
2017-06-20 00:48:02 +02:00
|
|
|
misc_info['avb_' + partition + '_algorithm'] = algorithm
|
|
|
|
misc_info['avb_' + partition + '_key_path'] = key
|
|
|
|
|
|
|
|
extra_args = OPTIONS.avb_extra_args.get(partition)
|
|
|
|
if extra_args:
|
2017-12-24 19:37:38 +01:00
|
|
|
print('Setting extra AVB signing args for %s to "%s"' % (
|
|
|
|
partition, extra_args))
|
2017-06-20 00:48:02 +02:00
|
|
|
args_key = AVB_FOOTER_ARGS_BY_PARTITION[partition]
|
|
|
|
misc_info[args_key] = (misc_info.get(args_key, '') + ' ' + extra_args)
|
|
|
|
|
|
|
|
for partition in AVB_FOOTER_ARGS_BY_PARTITION:
|
|
|
|
ReplaceAvbPartitionSigningKey(partition)
|
|
|
|
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
def BuildKeyMap(misc_info, key_mapping_options):
|
|
|
|
for s, d in key_mapping_options:
|
|
|
|
if s is None: # -d option
|
|
|
|
devkey = misc_info.get("default_system_dev_certificate",
|
2019-04-10 06:35:37 +02:00
|
|
|
"build/make/target/product/security/testkey")
|
2015-03-24 03:13:21 +01:00
|
|
|
devkeydir = os.path.dirname(devkey)
|
|
|
|
|
|
|
|
OPTIONS.key_map.update({
|
|
|
|
devkeydir + "/testkey": d + "/releasekey",
|
|
|
|
devkeydir + "/devkey": d + "/releasekey",
|
|
|
|
devkeydir + "/media": d + "/media",
|
|
|
|
devkeydir + "/shared": d + "/shared",
|
|
|
|
devkeydir + "/platform": d + "/platform",
|
|
|
|
})
|
|
|
|
else:
|
|
|
|
OPTIONS.key_map[s] = d
|
|
|
|
|
|
|
|
|
2016-01-13 19:32:47 +01:00
|
|
|
def GetApiLevelAndCodename(input_tf_zip):
|
|
|
|
data = input_tf_zip.read("SYSTEM/build.prop")
|
|
|
|
api_level = None
|
|
|
|
codename = None
|
|
|
|
for line in data.split("\n"):
|
|
|
|
line = line.strip()
|
|
|
|
if line and line[0] != '#' and "=" in line:
|
|
|
|
key, value = line.split("=", 1)
|
|
|
|
key = key.strip()
|
|
|
|
if key == "ro.build.version.sdk":
|
|
|
|
api_level = int(value.strip())
|
|
|
|
elif key == "ro.build.version.codename":
|
|
|
|
codename = value.strip()
|
|
|
|
|
|
|
|
if api_level is None:
|
|
|
|
raise ValueError("No ro.build.version.sdk in SYSTEM/build.prop")
|
|
|
|
if codename is None:
|
|
|
|
raise ValueError("No ro.build.version.codename in SYSTEM/build.prop")
|
|
|
|
|
|
|
|
return (api_level, codename)
|
|
|
|
|
|
|
|
|
|
|
|
def GetCodenameToApiLevelMap(input_tf_zip):
|
|
|
|
data = input_tf_zip.read("SYSTEM/build.prop")
|
|
|
|
api_level = None
|
|
|
|
codenames = None
|
|
|
|
for line in data.split("\n"):
|
|
|
|
line = line.strip()
|
|
|
|
if line and line[0] != '#' and "=" in line:
|
|
|
|
key, value = line.split("=", 1)
|
|
|
|
key = key.strip()
|
|
|
|
if key == "ro.build.version.sdk":
|
|
|
|
api_level = int(value.strip())
|
|
|
|
elif key == "ro.build.version.all_codenames":
|
|
|
|
codenames = value.strip().split(",")
|
|
|
|
|
|
|
|
if api_level is None:
|
|
|
|
raise ValueError("No ro.build.version.sdk in SYSTEM/build.prop")
|
|
|
|
if codenames is None:
|
|
|
|
raise ValueError("No ro.build.version.all_codenames in SYSTEM/build.prop")
|
|
|
|
|
|
|
|
result = dict()
|
|
|
|
for codename in codenames:
|
|
|
|
codename = codename.strip()
|
2019-03-15 17:33:43 +01:00
|
|
|
if codename:
|
2016-01-13 19:32:47 +01:00
|
|
|
result[codename] = api_level
|
|
|
|
return result
|
|
|
|
|
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
def ReadApexKeysInfo(tf_zip):
|
|
|
|
"""Parses the APEX keys info from a given target-files zip.
|
|
|
|
|
|
|
|
Given a target-files ZipFile, parses the META/apexkeys.txt entry and returns a
|
|
|
|
dict that contains the mapping from APEX names (e.g. com.android.tzdata) to a
|
|
|
|
tuple of (payload_key, container_key).
|
|
|
|
|
|
|
|
Args:
|
|
|
|
tf_zip: The input target_files ZipFile (already open).
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
(payload_key, container_key): payload_key contains the path to the payload
|
|
|
|
signing key; container_key contains the path to the container signing
|
|
|
|
key.
|
|
|
|
"""
|
|
|
|
keys = {}
|
|
|
|
for line in tf_zip.read("META/apexkeys.txt").split("\n"):
|
|
|
|
line = line.strip()
|
|
|
|
if not line:
|
|
|
|
continue
|
|
|
|
matches = re.match(
|
|
|
|
r'^name="(?P<NAME>.*)"\s+'
|
|
|
|
r'public_key="(?P<PAYLOAD_PUBLIC_KEY>.*)"\s+'
|
|
|
|
r'private_key="(?P<PAYLOAD_PRIVATE_KEY>.*)"\s+'
|
|
|
|
r'container_certificate="(?P<CONTAINER_CERT>.*)"\s+'
|
|
|
|
r'container_private_key="(?P<CONTAINER_PRIVATE_KEY>.*)"$',
|
|
|
|
line)
|
|
|
|
if not matches:
|
|
|
|
continue
|
|
|
|
|
|
|
|
name = matches.group('NAME')
|
|
|
|
payload_private_key = matches.group("PAYLOAD_PRIVATE_KEY")
|
|
|
|
|
|
|
|
def CompareKeys(pubkey, pubkey_suffix, privkey, privkey_suffix):
|
|
|
|
pubkey_suffix_len = len(pubkey_suffix)
|
|
|
|
privkey_suffix_len = len(privkey_suffix)
|
|
|
|
return (pubkey.endswith(pubkey_suffix) and
|
|
|
|
privkey.endswith(privkey_suffix) and
|
|
|
|
pubkey[:-pubkey_suffix_len] == privkey[:-privkey_suffix_len])
|
|
|
|
|
2019-03-26 20:59:25 +01:00
|
|
|
# Sanity check on the container key names, as we'll carry them without the
|
|
|
|
# extensions. This doesn't apply to payload keys though, which we will use
|
|
|
|
# full names only.
|
2019-03-15 17:37:01 +01:00
|
|
|
container_cert = matches.group("CONTAINER_CERT")
|
|
|
|
container_private_key = matches.group("CONTAINER_PRIVATE_KEY")
|
|
|
|
if not CompareKeys(
|
|
|
|
container_cert, OPTIONS.public_key_suffix,
|
|
|
|
container_private_key, OPTIONS.private_key_suffix):
|
|
|
|
raise ValueError("Failed to parse container keys: \n{}".format(line))
|
|
|
|
|
|
|
|
keys[name] = (payload_private_key,
|
|
|
|
container_cert[:-len(OPTIONS.public_key_suffix)])
|
|
|
|
|
|
|
|
return keys
|
|
|
|
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
def main(argv):
|
|
|
|
|
|
|
|
key_mapping_options = []
|
|
|
|
|
|
|
|
def option_handler(o, a):
|
|
|
|
if o in ("-e", "--extra_apks"):
|
|
|
|
names, key = a.split("=")
|
|
|
|
names = names.split(",")
|
|
|
|
for n in names:
|
|
|
|
OPTIONS.extra_apks[n] = key
|
2019-03-15 17:37:01 +01:00
|
|
|
elif o == "--extra_apex_payload_key":
|
|
|
|
apex_name, key = a.split("=")
|
|
|
|
OPTIONS.extra_apex_payload_keys[apex_name] = key
|
2018-06-19 21:19:35 +02:00
|
|
|
elif o == "--skip_apks_with_path_prefix":
|
|
|
|
# Sanity check the prefix, which must be in all upper case.
|
|
|
|
prefix = a.split('/')[0]
|
|
|
|
if not prefix or prefix != prefix.upper():
|
|
|
|
raise ValueError("Invalid path prefix '%s'" % (a,))
|
|
|
|
OPTIONS.skip_apks_with_path_prefix.add(a)
|
2015-03-24 03:13:21 +01:00
|
|
|
elif o in ("-d", "--default_key_mappings"):
|
|
|
|
key_mapping_options.append((None, a))
|
|
|
|
elif o in ("-k", "--key_mapping"):
|
|
|
|
key_mapping_options.append(a.split("=", 1))
|
|
|
|
elif o in ("-o", "--replace_ota_keys"):
|
|
|
|
OPTIONS.replace_ota_keys = True
|
|
|
|
elif o in ("-t", "--tag_changes"):
|
|
|
|
new = []
|
|
|
|
for i in a.split(","):
|
|
|
|
i = i.strip()
|
|
|
|
if not i or i[0] not in "-+":
|
|
|
|
raise ValueError("Bad tag change '%s'" % (i,))
|
|
|
|
new.append(i[0] + i[1:].strip())
|
|
|
|
OPTIONS.tag_changes = tuple(new)
|
|
|
|
elif o == "--replace_verity_public_key":
|
|
|
|
OPTIONS.replace_verity_public_key = (True, a)
|
|
|
|
elif o == "--replace_verity_private_key":
|
|
|
|
OPTIONS.replace_verity_private_key = (True, a)
|
2016-06-17 04:58:44 +02:00
|
|
|
elif o == "--replace_verity_keyid":
|
|
|
|
OPTIONS.replace_verity_keyid = (True, a)
|
2017-06-20 00:48:02 +02:00
|
|
|
elif o == "--avb_vbmeta_key":
|
|
|
|
OPTIONS.avb_keys['vbmeta'] = a
|
|
|
|
elif o == "--avb_vbmeta_algorithm":
|
|
|
|
OPTIONS.avb_algorithms['vbmeta'] = a
|
|
|
|
elif o == "--avb_vbmeta_extra_args":
|
|
|
|
OPTIONS.avb_extra_args['vbmeta'] = a
|
|
|
|
elif o == "--avb_boot_key":
|
|
|
|
OPTIONS.avb_keys['boot'] = a
|
|
|
|
elif o == "--avb_boot_algorithm":
|
|
|
|
OPTIONS.avb_algorithms['boot'] = a
|
|
|
|
elif o == "--avb_boot_extra_args":
|
|
|
|
OPTIONS.avb_extra_args['boot'] = a
|
|
|
|
elif o == "--avb_dtbo_key":
|
|
|
|
OPTIONS.avb_keys['dtbo'] = a
|
|
|
|
elif o == "--avb_dtbo_algorithm":
|
|
|
|
OPTIONS.avb_algorithms['dtbo'] = a
|
|
|
|
elif o == "--avb_dtbo_extra_args":
|
|
|
|
OPTIONS.avb_extra_args['dtbo'] = a
|
|
|
|
elif o == "--avb_system_key":
|
|
|
|
OPTIONS.avb_keys['system'] = a
|
|
|
|
elif o == "--avb_system_algorithm":
|
|
|
|
OPTIONS.avb_algorithms['system'] = a
|
|
|
|
elif o == "--avb_system_extra_args":
|
|
|
|
OPTIONS.avb_extra_args['system'] = a
|
2019-02-27 03:15:51 +01:00
|
|
|
elif o == "--avb_system_other_key":
|
|
|
|
OPTIONS.avb_keys['system_other'] = a
|
|
|
|
elif o == "--avb_system_other_algorithm":
|
|
|
|
OPTIONS.avb_algorithms['system_other'] = a
|
|
|
|
elif o == "--avb_system_other_extra_args":
|
|
|
|
OPTIONS.avb_extra_args['system_other'] = a
|
2017-06-20 00:48:02 +02:00
|
|
|
elif o == "--avb_vendor_key":
|
|
|
|
OPTIONS.avb_keys['vendor'] = a
|
|
|
|
elif o == "--avb_vendor_algorithm":
|
|
|
|
OPTIONS.avb_algorithms['vendor'] = a
|
|
|
|
elif o == "--avb_vendor_extra_args":
|
|
|
|
OPTIONS.avb_extra_args['vendor'] = a
|
2019-03-15 17:37:01 +01:00
|
|
|
elif o == "--avb_apex_extra_args":
|
|
|
|
OPTIONS.avb_extra_args['apex'] = a
|
2015-03-24 03:13:21 +01:00
|
|
|
else:
|
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
2017-06-20 00:48:02 +02:00
|
|
|
args = common.ParseOptions(
|
|
|
|
argv, __doc__,
|
|
|
|
extra_opts="e:d:k:ot:",
|
|
|
|
extra_long_opts=[
|
2017-12-24 19:37:38 +01:00
|
|
|
"extra_apks=",
|
2019-03-15 17:37:01 +01:00
|
|
|
"extra_apex_payload_key=",
|
2018-06-19 21:19:35 +02:00
|
|
|
"skip_apks_with_path_prefix=",
|
2017-12-24 19:37:38 +01:00
|
|
|
"default_key_mappings=",
|
|
|
|
"key_mapping=",
|
|
|
|
"replace_ota_keys",
|
|
|
|
"tag_changes=",
|
|
|
|
"replace_verity_public_key=",
|
|
|
|
"replace_verity_private_key=",
|
|
|
|
"replace_verity_keyid=",
|
2019-03-15 17:37:01 +01:00
|
|
|
"avb_apex_extra_args=",
|
2017-12-24 19:37:38 +01:00
|
|
|
"avb_vbmeta_algorithm=",
|
|
|
|
"avb_vbmeta_key=",
|
|
|
|
"avb_vbmeta_extra_args=",
|
|
|
|
"avb_boot_algorithm=",
|
|
|
|
"avb_boot_key=",
|
|
|
|
"avb_boot_extra_args=",
|
|
|
|
"avb_dtbo_algorithm=",
|
|
|
|
"avb_dtbo_key=",
|
|
|
|
"avb_dtbo_extra_args=",
|
|
|
|
"avb_system_algorithm=",
|
|
|
|
"avb_system_key=",
|
|
|
|
"avb_system_extra_args=",
|
2019-02-27 03:15:51 +01:00
|
|
|
"avb_system_other_algorithm=",
|
|
|
|
"avb_system_other_key=",
|
|
|
|
"avb_system_other_extra_args=",
|
2017-12-24 19:37:38 +01:00
|
|
|
"avb_vendor_algorithm=",
|
|
|
|
"avb_vendor_key=",
|
|
|
|
"avb_vendor_extra_args=",
|
2017-06-20 00:48:02 +02:00
|
|
|
],
|
|
|
|
extra_option_handler=option_handler)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
if len(args) != 2:
|
|
|
|
common.Usage(__doc__)
|
|
|
|
sys.exit(1)
|
|
|
|
|
2019-03-15 17:33:43 +01:00
|
|
|
common.InitLogging()
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
input_zip = zipfile.ZipFile(args[0], "r")
|
2017-06-13 21:54:58 +02:00
|
|
|
output_zip = zipfile.ZipFile(args[1], "w",
|
|
|
|
compression=zipfile.ZIP_DEFLATED,
|
|
|
|
allowZip64=True)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
misc_info = common.LoadInfoDict(input_zip)
|
|
|
|
|
|
|
|
BuildKeyMap(misc_info, key_mapping_options)
|
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
apk_keys_info, compressed_extension = common.ReadApkCerts(input_zip)
|
|
|
|
apk_keys = GetApkCerts(apk_keys_info)
|
|
|
|
|
|
|
|
apex_keys_info = ReadApexKeysInfo(input_zip)
|
|
|
|
apex_keys = GetApexKeys(apex_keys_info, apk_keys)
|
|
|
|
|
|
|
|
CheckApkAndApexKeysAvailable(
|
|
|
|
input_zip,
|
|
|
|
set(apk_keys.keys()) | set(apex_keys.keys()),
|
2019-03-19 20:24:03 +01:00
|
|
|
compressed_extension,
|
|
|
|
apex_keys)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2019-03-15 17:37:01 +01:00
|
|
|
key_passwords = common.GetKeyPasswords(
|
|
|
|
set(apk_keys.values()) | set(itertools.chain(*apex_keys.values())))
|
2016-09-30 02:53:56 +02:00
|
|
|
platform_api_level, _ = GetApiLevelAndCodename(input_zip)
|
2016-01-13 19:32:47 +01:00
|
|
|
codename_to_api_level_map = GetCodenameToApiLevelMap(input_zip)
|
|
|
|
|
2015-03-24 03:13:21 +01:00
|
|
|
ProcessTargetFiles(input_zip, output_zip, misc_info,
|
2019-03-15 17:37:01 +01:00
|
|
|
apk_keys, apex_keys, key_passwords,
|
|
|
|
platform_api_level, codename_to_api_level_map,
|
2017-08-14 15:49:21 +02:00
|
|
|
compressed_extension)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2015-04-01 20:21:55 +02:00
|
|
|
common.ZipClose(input_zip)
|
|
|
|
common.ZipClose(output_zip)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2016-08-04 04:21:52 +02:00
|
|
|
# Skip building userdata.img and cache.img when signing the target files.
|
2017-05-23 23:51:02 +02:00
|
|
|
new_args = ["--is_signing"]
|
|
|
|
# add_img_to_target_files builds the system image from scratch, so the
|
|
|
|
# recovery patch is guaranteed to be regenerated there.
|
|
|
|
if OPTIONS.rebuild_recovery:
|
|
|
|
new_args.append("--rebuild_recovery")
|
|
|
|
new_args.append(args[1])
|
2016-08-04 04:21:52 +02:00
|
|
|
add_img_to_target_files.main(new_args)
|
2015-03-24 03:13:21 +01:00
|
|
|
|
2017-12-24 19:37:38 +01:00
|
|
|
print("done.")
|
2015-03-24 03:13:21 +01:00
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
try:
|
|
|
|
main(sys.argv[1:])
|
2017-12-24 19:37:38 +01:00
|
|
|
except common.ExternalError as e:
|
|
|
|
print("\n ERROR: %s\n" % (e,))
|
2015-03-24 03:13:21 +01:00
|
|
|
sys.exit(1)
|
2017-06-20 00:48:02 +02:00
|
|
|
finally:
|
|
|
|
common.Cleanup()
|