1550 lines
65 KiB
Python
Executable file
1550 lines
65 KiB
Python
Executable file
#!/usr/bin/env python
|
|
#
|
|
# Copyright (C) 2008 The Android Open Source Project
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
"""
|
|
Given a target-files zipfile, produces an OTA package that installs that build.
|
|
An incremental OTA is produced if -i is given, otherwise a full OTA is produced.
|
|
|
|
Usage: ota_from_target_files [options] input_target_files output_ota_package
|
|
|
|
Common options that apply to both of non-A/B and A/B OTAs
|
|
|
|
--downgrade
|
|
Intentionally generate an incremental OTA that updates from a newer build
|
|
to an older one (e.g. downgrading from P preview back to O MR1).
|
|
"ota-downgrade=yes" will be set in the package metadata file. A data wipe
|
|
will always be enforced when using this flag, so "ota-wipe=yes" will also
|
|
be included in the metadata file. The update-binary in the source build
|
|
will be used in the OTA package, unless --binary flag is specified. Please
|
|
also check the comment for --override_timestamp below.
|
|
|
|
-i (--incremental_from) <file>
|
|
Generate an incremental OTA using the given target-files zip as the
|
|
starting build.
|
|
|
|
-k (--package_key) <key>
|
|
Key to use to sign the package (default is the value of
|
|
default_system_dev_certificate from the input target-files's
|
|
META/misc_info.txt, or "build/make/target/product/security/testkey" if
|
|
that value is not specified).
|
|
|
|
For incremental OTAs, the default value is based on the source
|
|
target-file, not the target build.
|
|
|
|
--override_timestamp
|
|
Intentionally generate an incremental OTA that updates from a newer build
|
|
to an older one (based on timestamp comparison), by setting the downgrade
|
|
flag in the package metadata. This differs from --downgrade flag, as we
|
|
don't enforce a data wipe with this flag. Because we know for sure this is
|
|
NOT an actual downgrade case, but two builds happen to be cut in a reverse
|
|
order (e.g. from two branches). A legit use case is that we cut a new
|
|
build C (after having A and B), but want to enfore an update path of A ->
|
|
C -> B. Specifying --downgrade may not help since that would enforce a
|
|
data wipe for C -> B update.
|
|
|
|
We used to set a fake timestamp in the package metadata for this flow. But
|
|
now we consolidate the two cases (i.e. an actual downgrade, or a downgrade
|
|
based on timestamp) with the same "ota-downgrade=yes" flag, with the
|
|
difference being whether "ota-wipe=yes" is set.
|
|
|
|
--wipe_user_data
|
|
Generate an OTA package that will wipe the user data partition when
|
|
installed.
|
|
|
|
--retrofit_dynamic_partitions
|
|
Generates an OTA package that updates a device to support dynamic
|
|
partitions (default False). This flag is implied when generating
|
|
an incremental OTA where the base build does not support dynamic
|
|
partitions but the target build does. For A/B, when this flag is set,
|
|
--skip_postinstall is implied.
|
|
|
|
--skip_compatibility_check
|
|
Skip checking compatibility of the input target files package.
|
|
|
|
--output_metadata_path
|
|
Write a copy of the metadata to a separate file. Therefore, users can
|
|
read the post build fingerprint without extracting the OTA package.
|
|
|
|
--force_non_ab
|
|
This flag can only be set on an A/B device that also supports non-A/B
|
|
updates. Implies --two_step.
|
|
If set, generate that non-A/B update package.
|
|
If not set, generates A/B package for A/B device and non-A/B package for
|
|
non-A/B device.
|
|
|
|
-o (--oem_settings) <main_file[,additional_files...]>
|
|
Comma separated list of files used to specify the expected OEM-specific
|
|
properties on the OEM partition of the intended device. Multiple expected
|
|
values can be used by providing multiple files. Only the first dict will
|
|
be used to compute fingerprint, while the rest will be used to assert
|
|
OEM-specific properties.
|
|
|
|
Non-A/B OTA specific options
|
|
|
|
-b (--binary) <file>
|
|
Use the given binary as the update-binary in the output package, instead
|
|
of the binary in the build's target_files. Use for development only.
|
|
|
|
--block
|
|
Generate a block-based OTA for non-A/B device. We have deprecated the
|
|
support for file-based OTA since O. Block-based OTA will be used by
|
|
default for all non-A/B devices. Keeping this flag here to not break
|
|
existing callers.
|
|
|
|
-e (--extra_script) <file>
|
|
Insert the contents of file at the end of the update script.
|
|
|
|
--full_bootloader
|
|
Similar to --full_radio. When generating an incremental OTA, always
|
|
include a full copy of bootloader image.
|
|
|
|
--full_radio
|
|
When generating an incremental OTA, always include a full copy of radio
|
|
image. This option is only meaningful when -i is specified, because a full
|
|
radio is always included in a full OTA if applicable.
|
|
|
|
--log_diff <file>
|
|
Generate a log file that shows the differences in the source and target
|
|
builds for an incremental package. This option is only meaningful when -i
|
|
is specified.
|
|
|
|
--oem_no_mount
|
|
For devices with OEM-specific properties but without an OEM partition, do
|
|
not mount the OEM partition in the updater-script. This should be very
|
|
rarely used, since it's expected to have a dedicated OEM partition for
|
|
OEM-specific properties. Only meaningful when -o is specified.
|
|
|
|
--stash_threshold <float>
|
|
Specify the threshold that will be used to compute the maximum allowed
|
|
stash size (defaults to 0.8).
|
|
|
|
-t (--worker_threads) <int>
|
|
Specify the number of worker-threads that will be used when generating
|
|
patches for incremental updates (defaults to 3).
|
|
|
|
--verify
|
|
Verify the checksums of the updated system and vendor (if any) partitions.
|
|
Non-A/B incremental OTAs only.
|
|
|
|
-2 (--two_step)
|
|
Generate a 'two-step' OTA package, where recovery is updated first, so
|
|
that any changes made to the system partition are done using the new
|
|
recovery (new kernel, etc.).
|
|
|
|
A/B OTA specific options
|
|
|
|
--disable_fec_computation
|
|
Disable the on device FEC data computation for incremental updates. OTA will be larger but installation will be faster.
|
|
|
|
--include_secondary
|
|
Additionally include the payload for secondary slot images (default:
|
|
False). Only meaningful when generating A/B OTAs.
|
|
|
|
By default, an A/B OTA package doesn't contain the images for the
|
|
secondary slot (e.g. system_other.img). Specifying this flag allows
|
|
generating a separate payload that will install secondary slot images.
|
|
|
|
Such a package needs to be applied in a two-stage manner, with a reboot
|
|
in-between. During the first stage, the updater applies the primary
|
|
payload only. Upon finishing, it reboots the device into the newly updated
|
|
slot. It then continues to install the secondary payload to the inactive
|
|
slot, but without switching the active slot at the end (needs the matching
|
|
support in update_engine, i.e. SWITCH_SLOT_ON_REBOOT flag).
|
|
|
|
Due to the special install procedure, the secondary payload will be always
|
|
generated as a full payload.
|
|
|
|
--payload_signer <signer>
|
|
Specify the signer when signing the payload and metadata for A/B OTAs.
|
|
By default (i.e. without this flag), it calls 'openssl pkeyutl' to sign
|
|
with the package private key. If the private key cannot be accessed
|
|
directly, a payload signer that knows how to do that should be specified.
|
|
The signer will be supplied with "-inkey <path_to_key>",
|
|
"-in <input_file>" and "-out <output_file>" parameters.
|
|
|
|
--payload_signer_args <args>
|
|
Specify the arguments needed for payload signer.
|
|
|
|
--payload_signer_maximum_signature_size <signature_size>
|
|
The maximum signature size (in bytes) that would be generated by the given
|
|
payload signer. Only meaningful when custom payload signer is specified
|
|
via '--payload_signer'.
|
|
If the signer uses a RSA key, this should be the number of bytes to
|
|
represent the modulus. If it uses an EC key, this is the size of a
|
|
DER-encoded ECDSA signature.
|
|
|
|
--payload_signer_key_size <key_size>
|
|
Deprecated. Use the '--payload_signer_maximum_signature_size' instead.
|
|
|
|
--boot_variable_file <path>
|
|
A file that contains the possible values of ro.boot.* properties. It's
|
|
used to calculate the possible runtime fingerprints when some
|
|
ro.product.* properties are overridden by the 'import' statement.
|
|
The file expects one property per line, and each line has the following
|
|
format: 'prop_name=value1,value2'. e.g. 'ro.boot.product.sku=std,pro'
|
|
The path specified can either be relative to the current working directory
|
|
or the path to a file inside of input_target_files.
|
|
|
|
--skip_postinstall
|
|
Skip the postinstall hooks when generating an A/B OTA package (default:
|
|
False). Note that this discards ALL the hooks, including non-optional
|
|
ones. Should only be used if caller knows it's safe to do so (e.g. all the
|
|
postinstall work is to dexopt apps and a data wipe will happen immediately
|
|
after). Only meaningful when generating A/B OTAs.
|
|
|
|
--partial "<PARTITION> [<PARTITION>[...]]"
|
|
Generate partial updates, overriding ab_partitions list with the given
|
|
list. Specify --partial= without partition list to let tooling auto detect
|
|
partial partition list.
|
|
|
|
--custom_image <custom_partition=custom_image>
|
|
Use the specified custom_image to update custom_partition when generating
|
|
an A/B OTA package. e.g. "--custom_image oem=oem.img --custom_image
|
|
cus=cus_test.img"
|
|
|
|
--disable_vabc
|
|
Disable Virtual A/B Compression, for builds that have compression enabled
|
|
by default.
|
|
|
|
--vabc_downgrade
|
|
Don't disable Virtual A/B Compression for downgrading OTAs.
|
|
For VABC downgrades, we must finish merging before doing data wipe, and
|
|
since data wipe is required for downgrading OTA, this might cause long
|
|
wait time in recovery.
|
|
|
|
--enable_vabc_xor
|
|
Enable the VABC xor feature. Will reduce space requirements for OTA, but OTA installation will be slower.
|
|
|
|
--force_minor_version
|
|
Override the update_engine minor version for delta generation.
|
|
|
|
--compressor_types
|
|
A colon ':' separated list of compressors. Allowed values are bz2 and brotli.
|
|
|
|
--enable_zucchini
|
|
Whether to enable to zucchini feature. Will generate smaller OTA but uses more memory, OTA generation will take longer.
|
|
|
|
--enable_puffdiff
|
|
Whether to enable to puffdiff feature. Will generate smaller OTA but uses more memory, OTA generation will take longer.
|
|
|
|
--enable_lz4diff
|
|
Whether to enable lz4diff feature. Will generate smaller OTA for EROFS but
|
|
uses more memory.
|
|
|
|
--spl_downgrade
|
|
Force generate an SPL downgrade OTA. Only needed if target build has an
|
|
older SPL.
|
|
|
|
--vabc_compression_param
|
|
Compression algorithm to be used for VABC. Available options: gz, lz4, zstd, brotli, none.
|
|
Compression level can be specified by appending ",$LEVEL" to option.
|
|
e.g. --vabc_compression_param=gz,9 specifies level 9 compression with gz algorithm
|
|
|
|
--security_patch_level
|
|
Override the security patch level in target files
|
|
|
|
--max_threads
|
|
Specify max number of threads allowed when generating A/B OTA
|
|
|
|
--vabc_cow_version
|
|
Specify the VABC cow version to be used
|
|
|
|
--compression_factor
|
|
Specify the maximum block size to be compressed at once during OTA. supported options: 4k, 8k, 16k, 32k, 64k, 128k, 256k
|
|
"""
|
|
|
|
from __future__ import print_function
|
|
|
|
import logging
|
|
import multiprocessing
|
|
import os
|
|
import os.path
|
|
import re
|
|
import shutil
|
|
import subprocess
|
|
import sys
|
|
import zipfile
|
|
|
|
import care_map_pb2
|
|
import common
|
|
import ota_utils
|
|
import payload_signer
|
|
from ota_utils import (VABC_COMPRESSION_PARAM_SUPPORT, FinalizeMetadata, GetPackageMetadata,
|
|
PayloadGenerator, SECURITY_PATCH_LEVEL_PROP_NAME, ExtractTargetFiles, CopyTargetFilesDir)
|
|
from common import DoesInputFileContain, IsSparseImage
|
|
import target_files_diff
|
|
from non_ab_ota import GenerateNonAbOtaPackage
|
|
from payload_signer import PayloadSigner
|
|
|
|
if sys.hexversion < 0x02070000:
|
|
print("Python 2.7 or newer is required.", file=sys.stderr)
|
|
sys.exit(1)
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
OPTIONS = ota_utils.OPTIONS
|
|
OPTIONS.verify = False
|
|
OPTIONS.patch_threshold = 0.95
|
|
OPTIONS.wipe_user_data = False
|
|
OPTIONS.extra_script = None
|
|
OPTIONS.worker_threads = multiprocessing.cpu_count() // 2
|
|
if OPTIONS.worker_threads == 0:
|
|
OPTIONS.worker_threads = 1
|
|
OPTIONS.two_step = False
|
|
OPTIONS.include_secondary = False
|
|
OPTIONS.block_based = True
|
|
OPTIONS.updater_binary = None
|
|
OPTIONS.oem_dicts = None
|
|
OPTIONS.oem_source = None
|
|
OPTIONS.oem_no_mount = False
|
|
OPTIONS.full_radio = False
|
|
OPTIONS.full_bootloader = False
|
|
# Stash size cannot exceed cache_size * threshold.
|
|
OPTIONS.cache_size = None
|
|
OPTIONS.stash_threshold = 0.8
|
|
OPTIONS.log_diff = None
|
|
OPTIONS.extracted_input = None
|
|
OPTIONS.skip_postinstall = False
|
|
OPTIONS.skip_compatibility_check = False
|
|
OPTIONS.disable_fec_computation = False
|
|
OPTIONS.disable_verity_computation = False
|
|
OPTIONS.partial = None
|
|
OPTIONS.custom_images = {}
|
|
OPTIONS.disable_vabc = False
|
|
OPTIONS.spl_downgrade = False
|
|
OPTIONS.vabc_downgrade = False
|
|
OPTIONS.enable_vabc_xor = True
|
|
OPTIONS.force_minor_version = None
|
|
OPTIONS.compressor_types = None
|
|
OPTIONS.enable_zucchini = False
|
|
OPTIONS.enable_puffdiff = None
|
|
OPTIONS.enable_lz4diff = False
|
|
OPTIONS.vabc_compression_param = None
|
|
OPTIONS.security_patch_level = None
|
|
OPTIONS.max_threads = None
|
|
OPTIONS.vabc_cow_version = None
|
|
OPTIONS.compression_factor = None
|
|
|
|
|
|
POSTINSTALL_CONFIG = 'META/postinstall_config.txt'
|
|
DYNAMIC_PARTITION_INFO = 'META/dynamic_partitions_info.txt'
|
|
MISC_INFO = 'META/misc_info.txt'
|
|
AB_PARTITIONS = 'META/ab_partitions.txt'
|
|
|
|
# Files to be unzipped for target diffing purpose.
|
|
TARGET_DIFFING_UNZIP_PATTERN = ['BOOT', 'RECOVERY', 'SYSTEM/*', 'VENDOR/*',
|
|
'PRODUCT/*', 'SYSTEM_EXT/*', 'ODM/*',
|
|
'VENDOR_DLKM/*', 'ODM_DLKM/*', 'SYSTEM_DLKM/*']
|
|
RETROFIT_DAP_UNZIP_PATTERN = ['OTA/super_*.img', AB_PARTITIONS]
|
|
|
|
# Images to be excluded from secondary payload. We essentially only keep
|
|
# 'system_other' and bootloader partitions.
|
|
SECONDARY_PAYLOAD_SKIPPED_IMAGES = [
|
|
'boot', 'dtbo', 'modem', 'odm', 'odm_dlkm', 'product', 'radio', 'recovery',
|
|
'system_dlkm', 'system_ext', 'vbmeta', 'vbmeta_system', 'vbmeta_vendor',
|
|
'vendor', 'vendor_boot']
|
|
|
|
|
|
def _LoadOemDicts(oem_source):
|
|
"""Returns the list of loaded OEM properties dict."""
|
|
if not oem_source:
|
|
return None
|
|
|
|
oem_dicts = []
|
|
for oem_file in oem_source:
|
|
oem_dicts.append(common.LoadDictionaryFromFile(oem_file))
|
|
return oem_dicts
|
|
|
|
def ModifyKeyvalueList(content: str, key: str, value: str):
|
|
""" Update update the key value list with specified key and value
|
|
Args:
|
|
content: The string content of dynamic_partitions_info.txt. Each line
|
|
should be a key valur pair, where string before the first '=' are keys,
|
|
remaining parts are values.
|
|
key: the key of the key value pair to modify
|
|
value: the new value to replace with
|
|
|
|
Returns:
|
|
Updated content of the key value list
|
|
"""
|
|
output_list = []
|
|
for line in content.splitlines():
|
|
if line.startswith(key+"="):
|
|
continue
|
|
output_list.append(line)
|
|
output_list.append("{}={}".format(key, value))
|
|
return "\n".join(output_list)
|
|
|
|
def ModifyVABCCompressionParam(content, algo):
|
|
""" Update update VABC Compression Param in dynamic_partitions_info.txt
|
|
Args:
|
|
content: The string content of dynamic_partitions_info.txt
|
|
algo: The compression algorithm should be used for VABC. See
|
|
https://cs.android.com/android/platform/superproject/+/master:system/core/fs_mgr/libsnapshot/cow_writer.cpp;l=127;bpv=1;bpt=1?q=CowWriter::ParseOptions&sq=
|
|
Returns:
|
|
Updated content of dynamic_partitions_info.txt , with custom compression algo
|
|
"""
|
|
return ModifyKeyvalueList(content, "virtual_ab_compression_method", algo)
|
|
|
|
|
|
def UpdatesInfoForSpecialUpdates(content, partitions_filter,
|
|
delete_keys=None):
|
|
""" Updates info file for secondary payload generation, partial update, etc.
|
|
|
|
Scan each line in the info file, and remove the unwanted partitions from
|
|
the dynamic partition list in the related properties. e.g.
|
|
"super_google_dynamic_partitions_partition_list=system vendor product"
|
|
will become "super_google_dynamic_partitions_partition_list=system".
|
|
|
|
Args:
|
|
content: The content of the input info file. e.g. misc_info.txt.
|
|
partitions_filter: A function to filter the desired partitions from a given
|
|
list
|
|
delete_keys: A list of keys to delete in the info file
|
|
|
|
Returns:
|
|
A string of the updated info content.
|
|
"""
|
|
|
|
output_list = []
|
|
# The suffix in partition_list variables that follows the name of the
|
|
# partition group.
|
|
list_suffix = 'partition_list'
|
|
for line in content.splitlines():
|
|
if line.startswith('#') or '=' not in line:
|
|
output_list.append(line)
|
|
continue
|
|
key, value = line.strip().split('=', 1)
|
|
|
|
if delete_keys and key in delete_keys:
|
|
pass
|
|
elif key.endswith(list_suffix):
|
|
partitions = value.split()
|
|
# TODO for partial update, partitions in the same group must be all
|
|
# updated or all omitted
|
|
partitions = filter(partitions_filter, partitions)
|
|
output_list.append('{}={}'.format(key, ' '.join(partitions)))
|
|
else:
|
|
output_list.append(line)
|
|
return '\n'.join(output_list)
|
|
|
|
|
|
def GetTargetFilesZipForSecondaryImages(input_file, skip_postinstall=False):
|
|
"""Returns a target-files.zip file for generating secondary payload.
|
|
|
|
Although the original target-files.zip already contains secondary slot
|
|
images (i.e. IMAGES/system_other.img), we need to rename the files to the
|
|
ones without _other suffix. Note that we cannot instead modify the names in
|
|
META/ab_partitions.txt, because there are no matching partitions on device.
|
|
|
|
For the partitions that don't have secondary images, the ones for primary
|
|
slot will be used. This is to ensure that we always have valid boot, vbmeta,
|
|
bootloader images in the inactive slot.
|
|
|
|
After writing system_other to inactive slot's system partiiton,
|
|
PackageManagerService will read `ro.cp_system_other_odex`, and set
|
|
`sys.cppreopt` to "requested". Then, according to
|
|
system/extras/cppreopts/cppreopts.rc , init will mount system_other at
|
|
/postinstall, and execute `cppreopts` to copy optimized APKs from
|
|
/postinstall to /data .
|
|
|
|
Args:
|
|
input_file: The input target-files.zip file.
|
|
skip_postinstall: Whether to skip copying the postinstall config file.
|
|
|
|
Returns:
|
|
The filename of the target-files.zip for generating secondary payload.
|
|
"""
|
|
|
|
def GetInfoForSecondaryImages(info_file):
|
|
"""Updates info file for secondary payload generation."""
|
|
with open(info_file) as f:
|
|
content = f.read()
|
|
# Remove virtual_ab flag from secondary payload so that OTA client
|
|
# don't use snapshots for secondary update
|
|
delete_keys = ['virtual_ab', "virtual_ab_retrofit"]
|
|
return UpdatesInfoForSpecialUpdates(
|
|
content, lambda p: p not in SECONDARY_PAYLOAD_SKIPPED_IMAGES,
|
|
delete_keys)
|
|
|
|
target_file = common.MakeTempFile(prefix="targetfiles-", suffix=".zip")
|
|
target_zip = zipfile.ZipFile(target_file, 'w', allowZip64=True)
|
|
|
|
fileslist = []
|
|
for (root, dirs, files) in os.walk(input_file):
|
|
root = root.lstrip(input_file).lstrip("/")
|
|
fileslist.extend([os.path.join(root, d) for d in dirs])
|
|
fileslist.extend([os.path.join(root, d) for d in files])
|
|
|
|
input_tmp = input_file
|
|
for filename in fileslist:
|
|
unzipped_file = os.path.join(input_tmp, *filename.split('/'))
|
|
if filename == 'IMAGES/system_other.img':
|
|
common.ZipWrite(target_zip, unzipped_file, arcname='IMAGES/system.img')
|
|
|
|
# Primary images and friends need to be skipped explicitly.
|
|
elif filename in ('IMAGES/system.img',
|
|
'IMAGES/system.map'):
|
|
pass
|
|
|
|
# Copy images that are not in SECONDARY_PAYLOAD_SKIPPED_IMAGES.
|
|
elif filename.startswith(('IMAGES/', 'RADIO/')):
|
|
image_name = os.path.basename(filename)
|
|
if image_name not in ['{}.img'.format(partition) for partition in
|
|
SECONDARY_PAYLOAD_SKIPPED_IMAGES]:
|
|
common.ZipWrite(target_zip, unzipped_file, arcname=filename)
|
|
|
|
# Skip copying the postinstall config if requested.
|
|
elif skip_postinstall and filename == POSTINSTALL_CONFIG:
|
|
pass
|
|
|
|
elif filename.startswith('META/'):
|
|
# Remove the unnecessary partitions for secondary images from the
|
|
# ab_partitions file.
|
|
if filename == AB_PARTITIONS:
|
|
with open(unzipped_file) as f:
|
|
partition_list = f.read().splitlines()
|
|
partition_list = [partition for partition in partition_list if partition
|
|
and partition not in SECONDARY_PAYLOAD_SKIPPED_IMAGES]
|
|
common.ZipWriteStr(target_zip, filename,
|
|
'\n'.join(partition_list))
|
|
# Remove the unnecessary partitions from the dynamic partitions list.
|
|
elif (filename == 'META/misc_info.txt' or
|
|
filename == DYNAMIC_PARTITION_INFO):
|
|
modified_info = GetInfoForSecondaryImages(unzipped_file)
|
|
common.ZipWriteStr(target_zip, filename, modified_info)
|
|
else:
|
|
common.ZipWrite(target_zip, unzipped_file, arcname=filename)
|
|
|
|
common.ZipClose(target_zip)
|
|
|
|
return target_file
|
|
|
|
|
|
def GetTargetFilesZipWithoutPostinstallConfig(input_file):
|
|
"""Returns a target-files.zip that's not containing postinstall_config.txt.
|
|
|
|
This allows brillo_update_payload script to skip writing all the postinstall
|
|
hooks in the generated payload. The input target-files.zip file will be
|
|
duplicated, with 'META/postinstall_config.txt' skipped. If input_file doesn't
|
|
contain the postinstall_config.txt entry, the input file will be returned.
|
|
|
|
Args:
|
|
input_file: The input target-files.zip filename.
|
|
|
|
Returns:
|
|
The filename of target-files.zip that doesn't contain postinstall config.
|
|
"""
|
|
config_path = os.path.join(input_file, POSTINSTALL_CONFIG)
|
|
if os.path.exists(config_path):
|
|
os.unlink(config_path)
|
|
return input_file
|
|
|
|
|
|
def ParseInfoDict(target_file_path):
|
|
return common.LoadInfoDict(target_file_path)
|
|
|
|
def ModifyTargetFilesDynamicPartitionInfo(input_file, key, value):
|
|
"""Returns a target-files.zip with a custom VABC compression param.
|
|
Args:
|
|
input_file: The input target-files.zip path
|
|
vabc_compression_param: Custom Virtual AB Compression algorithm
|
|
|
|
Returns:
|
|
The path to modified target-files.zip
|
|
"""
|
|
if os.path.isdir(input_file):
|
|
dynamic_partition_info_path = os.path.join(
|
|
input_file, *DYNAMIC_PARTITION_INFO.split("/"))
|
|
with open(dynamic_partition_info_path, "r") as fp:
|
|
dynamic_partition_info = fp.read()
|
|
dynamic_partition_info = ModifyKeyvalueList(
|
|
dynamic_partition_info, key, value)
|
|
with open(dynamic_partition_info_path, "w") as fp:
|
|
fp.write(dynamic_partition_info)
|
|
return input_file
|
|
|
|
target_file = common.MakeTempFile(prefix="targetfiles-", suffix=".zip")
|
|
shutil.copyfile(input_file, target_file)
|
|
common.ZipDelete(target_file, DYNAMIC_PARTITION_INFO)
|
|
with zipfile.ZipFile(input_file, 'r', allowZip64=True) as zfp:
|
|
dynamic_partition_info = zfp.read(DYNAMIC_PARTITION_INFO).decode()
|
|
dynamic_partition_info = ModifyKeyvalueList(
|
|
dynamic_partition_info, key, value)
|
|
with zipfile.ZipFile(target_file, "a", allowZip64=True) as output_zip:
|
|
output_zip.writestr(DYNAMIC_PARTITION_INFO, dynamic_partition_info)
|
|
return target_file
|
|
|
|
def GetTargetFilesZipForCustomVABCCompression(input_file, vabc_compression_param):
|
|
"""Returns a target-files.zip with a custom VABC compression param.
|
|
Args:
|
|
input_file: The input target-files.zip path
|
|
vabc_compression_param: Custom Virtual AB Compression algorithm
|
|
|
|
Returns:
|
|
The path to modified target-files.zip
|
|
"""
|
|
return ModifyTargetFilesDynamicPartitionInfo(input_file, "virtual_ab_compression_method", vabc_compression_param)
|
|
|
|
|
|
def GetTargetFilesZipForPartialUpdates(input_file, ab_partitions):
|
|
"""Returns a target-files.zip for partial ota update package generation.
|
|
|
|
This function modifies ab_partitions list with the desired partitions before
|
|
calling the brillo_update_payload script. It also cleans up the reference to
|
|
the excluded partitions in the info file, e.g. misc_info.txt.
|
|
|
|
Args:
|
|
input_file: The input target-files.zip filename.
|
|
ab_partitions: A list of partitions to include in the partial update
|
|
|
|
Returns:
|
|
The filename of target-files.zip used for partial ota update.
|
|
"""
|
|
|
|
original_ab_partitions = common.ReadFromInputFile(input_file, AB_PARTITIONS)
|
|
|
|
unrecognized_partitions = [partition for partition in ab_partitions if
|
|
partition not in original_ab_partitions]
|
|
if unrecognized_partitions:
|
|
raise ValueError("Unrecognized partitions when generating partial updates",
|
|
unrecognized_partitions)
|
|
|
|
logger.info("Generating partial updates for %s", ab_partitions)
|
|
for subdir in ["IMAGES", "RADIO", "PREBUILT_IMAGES"]:
|
|
image_dir = os.path.join(subdir)
|
|
if not os.path.exists(image_dir):
|
|
continue
|
|
for filename in os.listdir(image_dir):
|
|
filepath = os.path.join(image_dir, filename)
|
|
if filename.endswith(".img"):
|
|
partition_name = filename.removesuffix(".img")
|
|
if partition_name not in ab_partitions:
|
|
os.unlink(filepath)
|
|
|
|
common.WriteToInputFile(input_file, 'META/ab_partitions.txt',
|
|
'\n'.join(ab_partitions))
|
|
CARE_MAP_ENTRY = "META/care_map.pb"
|
|
if DoesInputFileContain(input_file, CARE_MAP_ENTRY):
|
|
caremap = care_map_pb2.CareMap()
|
|
caremap.ParseFromString(
|
|
common.ReadBytesFromInputFile(input_file, CARE_MAP_ENTRY))
|
|
filtered = [
|
|
part for part in caremap.partitions if part.name in ab_partitions]
|
|
del caremap.partitions[:]
|
|
caremap.partitions.extend(filtered)
|
|
common.WriteBytesToInputFile(input_file, CARE_MAP_ENTRY,
|
|
caremap.SerializeToString())
|
|
|
|
for info_file in ['META/misc_info.txt', DYNAMIC_PARTITION_INFO]:
|
|
if not DoesInputFileContain(input_file, info_file):
|
|
logger.warning('Cannot find %s in input zipfile', info_file)
|
|
continue
|
|
|
|
content = common.ReadFromInputFile(input_file, info_file)
|
|
modified_info = UpdatesInfoForSpecialUpdates(
|
|
content, lambda p: p in ab_partitions)
|
|
if OPTIONS.vabc_compression_param and info_file == DYNAMIC_PARTITION_INFO:
|
|
modified_info = ModifyVABCCompressionParam(
|
|
modified_info, OPTIONS.vabc_compression_param)
|
|
common.WriteToInputFile(input_file, info_file, modified_info)
|
|
|
|
def IsInPartialList(postinstall_line: str):
|
|
idx = postinstall_line.find("=")
|
|
if idx < 0:
|
|
return False
|
|
key = postinstall_line[:idx]
|
|
logger.info("%s %s", key, ab_partitions)
|
|
for part in ab_partitions:
|
|
if key.endswith("_" + part):
|
|
return True
|
|
return False
|
|
|
|
if common.DoesInputFileContain(input_file, POSTINSTALL_CONFIG):
|
|
postinstall_config = common.ReadFromInputFile(
|
|
input_file, POSTINSTALL_CONFIG)
|
|
postinstall_config = [
|
|
line for line in postinstall_config.splitlines() if IsInPartialList(line)]
|
|
if postinstall_config:
|
|
postinstall_config = "\n".join(postinstall_config)
|
|
common.WriteToInputFile(
|
|
input_file, POSTINSTALL_CONFIG, postinstall_config)
|
|
else:
|
|
os.unlink(os.path.join(input_file, POSTINSTALL_CONFIG))
|
|
|
|
return input_file
|
|
|
|
|
|
def GetTargetFilesZipForRetrofitDynamicPartitions(input_file,
|
|
super_block_devices,
|
|
dynamic_partition_list):
|
|
"""Returns a target-files.zip for retrofitting dynamic partitions.
|
|
|
|
This allows brillo_update_payload to generate an OTA based on the exact
|
|
bits on the block devices. Postinstall is disabled.
|
|
|
|
Args:
|
|
input_file: The input target-files.zip filename.
|
|
super_block_devices: The list of super block devices
|
|
dynamic_partition_list: The list of dynamic partitions
|
|
|
|
Returns:
|
|
The filename of target-files.zip with *.img replaced with super_*.img for
|
|
each block device in super_block_devices.
|
|
"""
|
|
assert super_block_devices, "No super_block_devices are specified."
|
|
|
|
replace = {'OTA/super_{}.img'.format(dev): 'IMAGES/{}.img'.format(dev)
|
|
for dev in super_block_devices}
|
|
|
|
# Remove partitions from META/ab_partitions.txt that is in
|
|
# dynamic_partition_list but not in super_block_devices so that
|
|
# brillo_update_payload won't generate update for those logical partitions.
|
|
ab_partitions_lines = common.ReadFromInputFile(
|
|
input_file, AB_PARTITIONS).split("\n")
|
|
ab_partitions = [line.strip() for line in ab_partitions_lines]
|
|
# Assert that all super_block_devices are in ab_partitions
|
|
super_device_not_updated = [partition for partition in super_block_devices
|
|
if partition not in ab_partitions]
|
|
assert not super_device_not_updated, \
|
|
"{} is in super_block_devices but not in {}".format(
|
|
super_device_not_updated, AB_PARTITIONS)
|
|
# ab_partitions -= (dynamic_partition_list - super_block_devices)
|
|
to_delete = [AB_PARTITIONS]
|
|
|
|
# Always skip postinstall for a retrofit update.
|
|
to_delete += [POSTINSTALL_CONFIG]
|
|
|
|
# Delete dynamic_partitions_info.txt so that brillo_update_payload thinks this
|
|
# is a regular update on devices without dynamic partitions support.
|
|
to_delete += [DYNAMIC_PARTITION_INFO]
|
|
|
|
# Remove the existing partition images as well as the map files.
|
|
to_delete += list(replace.values())
|
|
to_delete += ['IMAGES/{}.map'.format(dev) for dev in super_block_devices]
|
|
for item in to_delete:
|
|
os.unlink(os.path.join(input_file, item))
|
|
|
|
# Write super_{foo}.img as {foo}.img.
|
|
for src, dst in replace.items():
|
|
assert DoesInputFileContain(input_file, src), \
|
|
'Missing {} in {}; {} cannot be written'.format(src, input_file, dst)
|
|
source_path = os.path.join(input_file, *src.split("/"))
|
|
target_path = os.path.join(input_file, *dst.split("/"))
|
|
os.rename(source_path, target_path)
|
|
|
|
# Write new ab_partitions.txt file
|
|
new_ab_partitions = os.path.join(input_file, AB_PARTITIONS)
|
|
with open(new_ab_partitions, 'w') as f:
|
|
for partition in ab_partitions:
|
|
if (partition in dynamic_partition_list and
|
|
partition not in super_block_devices):
|
|
logger.info("Dropping %s from ab_partitions.txt", partition)
|
|
continue
|
|
f.write(partition + "\n")
|
|
|
|
return input_file
|
|
|
|
|
|
def GetTargetFilesZipForCustomImagesUpdates(input_file, custom_images: dict):
|
|
"""Returns a target-files.zip for custom partitions update.
|
|
|
|
This function modifies ab_partitions list with the desired custom partitions
|
|
and puts the custom images into the target target-files.zip.
|
|
|
|
Args:
|
|
input_file: The input target-files extracted directory
|
|
custom_images: A map of custom partitions and custom images.
|
|
|
|
Returns:
|
|
The extracted dir of a target-files.zip which has renamed the custom images
|
|
in the IMAGES/ to their partition names.
|
|
"""
|
|
for custom_image in custom_images.values():
|
|
if not os.path.exists(os.path.join(input_file, "IMAGES", custom_image)):
|
|
raise ValueError("Specified custom image {} not found in target files {}, available images are {}",
|
|
custom_image, input_file, os.listdir(os.path.join(input_file, "IMAGES")))
|
|
|
|
for custom_partition, custom_image in custom_images.items():
|
|
default_custom_image = '{}.img'.format(custom_partition)
|
|
if default_custom_image != custom_image:
|
|
src = os.path.join(input_file, 'IMAGES', custom_image)
|
|
dst = os.path.join(input_file, 'IMAGES', default_custom_image)
|
|
os.rename(src, dst)
|
|
|
|
return input_file
|
|
|
|
|
|
def GeneratePartitionTimestampFlags(partition_state):
|
|
partition_timestamps = [
|
|
part.partition_name + ":" + part.version
|
|
for part in partition_state]
|
|
return ["--partition_timestamps", ",".join(partition_timestamps)]
|
|
|
|
|
|
def GeneratePartitionTimestampFlagsDowngrade(
|
|
pre_partition_state, post_partition_state):
|
|
assert pre_partition_state is not None
|
|
partition_timestamps = {}
|
|
for part in post_partition_state:
|
|
partition_timestamps[part.partition_name] = part.version
|
|
for part in pre_partition_state:
|
|
if part.partition_name in partition_timestamps:
|
|
partition_timestamps[part.partition_name] = \
|
|
max(part.version, partition_timestamps[part.partition_name])
|
|
return [
|
|
"--partition_timestamps",
|
|
",".join([key + ":" + val for (key, val)
|
|
in partition_timestamps.items()])
|
|
]
|
|
|
|
|
|
def SupportsMainlineGkiUpdates(target_file):
|
|
"""Return True if the build supports MainlineGKIUpdates.
|
|
|
|
This function scans the product.img file in IMAGES/ directory for
|
|
pattern |*/apex/com.android.gki.*.apex|. If there are files
|
|
matching this pattern, conclude that build supports mainline
|
|
GKI and return True
|
|
|
|
Args:
|
|
target_file: Path to a target_file.zip, or an extracted directory
|
|
Return:
|
|
True if thisb uild supports Mainline GKI Updates.
|
|
"""
|
|
if target_file is None:
|
|
return False
|
|
if os.path.isfile(target_file):
|
|
target_file = common.UnzipTemp(target_file, ["IMAGES/product.img"])
|
|
if not os.path.isdir(target_file):
|
|
assert os.path.isdir(target_file), \
|
|
"{} must be a path to zip archive or dir containing extracted"\
|
|
" target_files".format(target_file)
|
|
image_file = os.path.join(target_file, "IMAGES", "product.img")
|
|
|
|
if not os.path.isfile(image_file):
|
|
return False
|
|
|
|
if IsSparseImage(image_file):
|
|
# Unsparse the image
|
|
tmp_img = common.MakeTempFile(suffix=".img")
|
|
subprocess.check_output(["simg2img", image_file, tmp_img])
|
|
image_file = tmp_img
|
|
|
|
cmd = ["debugfs_static", "-R", "ls -p /apex", image_file]
|
|
output = subprocess.check_output(cmd).decode()
|
|
|
|
pattern = re.compile(r"com\.android\.gki\..*\.apex")
|
|
return pattern.search(output) is not None
|
|
|
|
|
|
def ExtractOrCopyTargetFiles(target_file):
|
|
if os.path.isdir(target_file):
|
|
return CopyTargetFilesDir(target_file)
|
|
else:
|
|
return ExtractTargetFiles(target_file)
|
|
|
|
|
|
def ValidateCompressionParam(target_info):
|
|
vabc_compression_param = OPTIONS.vabc_compression_param
|
|
if vabc_compression_param:
|
|
minimum_api_level_required = VABC_COMPRESSION_PARAM_SUPPORT[vabc_compression_param.split(",")[0]]
|
|
if target_info.vendor_api_level < minimum_api_level_required:
|
|
raise ValueError("Specified VABC compression param {} is only supported for API level >= {}, device is on API level {}".format(
|
|
vabc_compression_param, minimum_api_level_required, target_info.vendor_api_level))
|
|
|
|
|
|
def GenerateAbOtaPackage(target_file, output_file, source_file=None):
|
|
"""Generates an Android OTA package that has A/B update payload."""
|
|
# If input target_files are directories, create a copy so that we can modify
|
|
# them directly
|
|
target_info = common.BuildInfo(OPTIONS.info_dict, OPTIONS.oem_dicts)
|
|
if OPTIONS.disable_vabc and target_info.is_release_key:
|
|
raise ValueError("Disabling VABC on release-key builds is not supported.")
|
|
|
|
ValidateCompressionParam(target_info)
|
|
vabc_compression_param = target_info.vabc_compression_param
|
|
|
|
target_file = ExtractOrCopyTargetFiles(target_file)
|
|
if source_file is not None:
|
|
source_file = ExtractOrCopyTargetFiles(source_file)
|
|
# Stage the output zip package for package signing.
|
|
if not OPTIONS.no_signing:
|
|
staging_file = common.MakeTempFile(suffix='.zip')
|
|
else:
|
|
staging_file = output_file
|
|
output_zip = zipfile.ZipFile(staging_file, "w",
|
|
compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True)
|
|
|
|
if source_file is not None:
|
|
source_file = ExtractTargetFiles(source_file)
|
|
assert "ab_partitions" in OPTIONS.source_info_dict, \
|
|
"META/ab_partitions.txt is required for ab_update."
|
|
assert "ab_partitions" in OPTIONS.target_info_dict, \
|
|
"META/ab_partitions.txt is required for ab_update."
|
|
target_info = common.BuildInfo(OPTIONS.target_info_dict, OPTIONS.oem_dicts)
|
|
source_info = common.BuildInfo(OPTIONS.source_info_dict, OPTIONS.oem_dicts)
|
|
# If source supports VABC, delta_generator/update_engine will attempt to
|
|
# use VABC. This dangerous, as the target build won't have snapuserd to
|
|
# serve I/O request when device boots. Therefore, disable VABC if source
|
|
# build doesn't supports it.
|
|
if not source_info.is_vabc or not target_info.is_vabc:
|
|
logger.info("Either source or target does not support VABC, disabling.")
|
|
OPTIONS.disable_vabc = True
|
|
if OPTIONS.vabc_compression_param is None and \
|
|
source_info.vabc_compression_param != target_info.vabc_compression_param:
|
|
logger.info("Source build and target build use different compression methods {} vs {}, default to source builds parameter {}".format(
|
|
source_info.vabc_compression_param, target_info.vabc_compression_param, source_info.vabc_compression_param))
|
|
vabc_compression_param = source_info.vabc_compression_param
|
|
# Virtual AB Cow version 3 is introduced in Android U with improved memory
|
|
# and install time performance. All OTA's with
|
|
# both the source build and target build with VIRTUAL_AB_COW_VERSION = 3
|
|
# can support the new format. Otherwise, fallback on older versions
|
|
if not OPTIONS.vabc_cow_version:
|
|
if not source_info.vabc_cow_version or not target_info.vabc_cow_version:
|
|
logger.info("Source or Target doesn't have VABC_COW_VERSION specified, default to version 2")
|
|
OPTIONS.vabc_cow_version = 2
|
|
elif source_info.vabc_cow_version != target_info.vabc_cow_version:
|
|
logger.info("Source and Target have different cow VABC_COW_VERSION specified, default to minimum version")
|
|
OPTIONS.vabc_cow_version = min(source_info.vabc_cow_version, target_info.vabc_cow_version)
|
|
|
|
# Virtual AB Compression was introduced in Androd S.
|
|
# Later, we backported VABC to Android R. But verity support was not
|
|
# backported, so if VABC is used and we are on Android R, disable
|
|
# verity computation.
|
|
if not OPTIONS.disable_vabc and source_info.is_android_r:
|
|
OPTIONS.disable_verity_computation = True
|
|
OPTIONS.disable_fec_computation = True
|
|
|
|
else:
|
|
assert "ab_partitions" in OPTIONS.info_dict, \
|
|
"META/ab_partitions.txt is required for ab_update."
|
|
source_info = None
|
|
if not OPTIONS.vabc_cow_version:
|
|
if not target_info.vabc_cow_version:
|
|
OPTIONS.vabc_cow_version = 2
|
|
elif target_info.vabc_cow_version >= "3" and target_info.vendor_api_level < 35:
|
|
logger.warning(
|
|
"This full OTA is configured to use VABC cow version"
|
|
" 3 which is supported since"
|
|
" Android API level 35, but device is "
|
|
"launched with {} . If this full OTA is"
|
|
" served to a device running old build, OTA might fail due to "
|
|
"unsupported vabc cow version. For safety, version 2 is used because "
|
|
"it's supported since day 1.".format(
|
|
target_info.vendor_api_level))
|
|
OPTIONS.vabc_cow_version = 2
|
|
if OPTIONS.vabc_compression_param is None and vabc_compression_param:
|
|
minimum_api_level_required = VABC_COMPRESSION_PARAM_SUPPORT[
|
|
vabc_compression_param]
|
|
if target_info.vendor_api_level < minimum_api_level_required:
|
|
logger.warning(
|
|
"This full OTA is configured to use VABC compression algorithm"
|
|
" {}, which is supported since"
|
|
" Android API level {}, but device is "
|
|
"launched with {} . If this full OTA is"
|
|
" served to a device running old build, OTA might fail due to "
|
|
"unsupported compression parameter. For safety, gz is used because "
|
|
"it's supported since day 1.".format(
|
|
vabc_compression_param,
|
|
minimum_api_level_required,
|
|
target_info.vendor_api_level))
|
|
vabc_compression_param = "gz"
|
|
|
|
if OPTIONS.partial == []:
|
|
logger.info(
|
|
"Automatically detecting partial partition list from input target files.")
|
|
OPTIONS.partial = target_info.get(
|
|
"partial_ota_update_partitions_list").split()
|
|
assert OPTIONS.partial, "Input target_file does not have"
|
|
" partial_ota_update_partitions_list defined, failed to auto detect partial"
|
|
" partition list. Please specify list of partitions to update manually via"
|
|
" --partial=a,b,c , or generate a complete OTA by removing the --partial"
|
|
" option"
|
|
OPTIONS.partial.sort()
|
|
if source_info:
|
|
source_partial_list = source_info.get(
|
|
"partial_ota_update_partitions_list").split()
|
|
if source_partial_list:
|
|
source_partial_list.sort()
|
|
if source_partial_list != OPTIONS.partial:
|
|
logger.warning("Source build and target build have different partial partition lists. Source: %s, target: %s, taking the intersection.",
|
|
source_partial_list, OPTIONS.partial)
|
|
OPTIONS.partial = list(
|
|
set(OPTIONS.partial) and set(source_partial_list))
|
|
OPTIONS.partial.sort()
|
|
logger.info("Automatically deduced partial partition list: %s",
|
|
OPTIONS.partial)
|
|
|
|
if target_info.vendor_suppressed_vabc:
|
|
logger.info("Vendor suppressed VABC. Disabling")
|
|
OPTIONS.disable_vabc = True
|
|
|
|
# Both source and target build need to support VABC XOR for us to use it.
|
|
# Source build's update_engine must be able to write XOR ops, and target
|
|
# build's snapuserd must be able to interpret XOR ops.
|
|
if not target_info.is_vabc_xor or OPTIONS.disable_vabc or \
|
|
(source_info is not None and not source_info.is_vabc_xor):
|
|
logger.info("VABC XOR Not supported, disabling")
|
|
OPTIONS.enable_vabc_xor = False
|
|
|
|
if OPTIONS.vabc_compression_param == "none":
|
|
logger.info(
|
|
"VABC Compression algorithm is set to 'none', disabling VABC xor")
|
|
OPTIONS.enable_vabc_xor = False
|
|
|
|
if OPTIONS.enable_vabc_xor:
|
|
api_level = -1
|
|
if source_info is not None:
|
|
api_level = source_info.vendor_api_level
|
|
if api_level == -1:
|
|
api_level = target_info.vendor_api_level
|
|
|
|
# XOR is only supported on T and higher.
|
|
if api_level < 33:
|
|
logger.error("VABC XOR not supported on this vendor, disabling")
|
|
OPTIONS.enable_vabc_xor = False
|
|
|
|
if OPTIONS.vabc_compression_param:
|
|
vabc_compression_param = OPTIONS.vabc_compression_param
|
|
|
|
additional_args = []
|
|
|
|
# Prepare custom images.
|
|
if OPTIONS.custom_images:
|
|
target_file = GetTargetFilesZipForCustomImagesUpdates(
|
|
target_file, OPTIONS.custom_images)
|
|
|
|
if OPTIONS.retrofit_dynamic_partitions:
|
|
target_file = GetTargetFilesZipForRetrofitDynamicPartitions(
|
|
target_file, target_info.get("super_block_devices").strip().split(),
|
|
target_info.get("dynamic_partition_list").strip().split())
|
|
elif OPTIONS.partial:
|
|
target_file = GetTargetFilesZipForPartialUpdates(target_file,
|
|
OPTIONS.partial)
|
|
if vabc_compression_param != target_info.vabc_compression_param:
|
|
target_file = GetTargetFilesZipForCustomVABCCompression(
|
|
target_file, vabc_compression_param)
|
|
if OPTIONS.vabc_cow_version:
|
|
target_file = ModifyTargetFilesDynamicPartitionInfo(target_file, "virtual_ab_cow_version", OPTIONS.vabc_cow_version)
|
|
if OPTIONS.compression_factor:
|
|
target_file = ModifyTargetFilesDynamicPartitionInfo(target_file, "virtual_ab_compression_factor", OPTIONS.compression_factor)
|
|
if OPTIONS.skip_postinstall:
|
|
target_file = GetTargetFilesZipWithoutPostinstallConfig(target_file)
|
|
# Target_file may have been modified, reparse ab_partitions
|
|
target_info.info_dict['ab_partitions'] = common.ReadFromInputFile(target_file,
|
|
AB_PARTITIONS).strip().split("\n")
|
|
|
|
from check_target_files_vintf import CheckVintfIfTrebleEnabled
|
|
CheckVintfIfTrebleEnabled(target_file, target_info)
|
|
|
|
# Allow boot_variable_file to also exist in target-files
|
|
if OPTIONS.boot_variable_file:
|
|
if not os.path.isfile(OPTIONS.boot_variable_file):
|
|
OPTIONS.boot_variable_file = os.path.join(target_file, OPTIONS.boot_variable_file)
|
|
# Metadata to comply with Android OTA package format.
|
|
metadata = GetPackageMetadata(target_info, source_info)
|
|
# Generate payload.
|
|
payload = PayloadGenerator(
|
|
wipe_user_data=OPTIONS.wipe_user_data, minor_version=OPTIONS.force_minor_version, is_partial_update=OPTIONS.partial, spl_downgrade=OPTIONS.spl_downgrade)
|
|
|
|
partition_timestamps_flags = []
|
|
# Enforce a max timestamp this payload can be applied on top of.
|
|
if OPTIONS.downgrade:
|
|
# When generating ota between merged target-files, partition build date can
|
|
# decrease in target, at the same time as ro.build.date.utc increases,
|
|
# so always pick largest value.
|
|
max_timestamp = max(source_info.GetBuildProp("ro.build.date.utc"),
|
|
str(metadata.postcondition.timestamp))
|
|
partition_timestamps_flags = GeneratePartitionTimestampFlagsDowngrade(
|
|
metadata.precondition.partition_state,
|
|
metadata.postcondition.partition_state
|
|
)
|
|
else:
|
|
max_timestamp = str(metadata.postcondition.timestamp)
|
|
partition_timestamps_flags = GeneratePartitionTimestampFlags(
|
|
metadata.postcondition.partition_state)
|
|
|
|
if not ota_utils.IsZucchiniCompatible(source_file, target_file):
|
|
logger.warning(
|
|
"Builds doesn't support zucchini, or source/target don't have compatible zucchini versions. Disabling zucchini.")
|
|
OPTIONS.enable_zucchini = False
|
|
|
|
security_patch_level = target_info.GetBuildProp(
|
|
"ro.build.version.security_patch")
|
|
if OPTIONS.security_patch_level is not None:
|
|
security_patch_level = OPTIONS.security_patch_level
|
|
|
|
additional_args += ["--security_patch_level", security_patch_level]
|
|
|
|
if OPTIONS.max_threads:
|
|
additional_args += ["--max_threads", OPTIONS.max_threads]
|
|
|
|
additional_args += ["--enable_zucchini=" +
|
|
str(OPTIONS.enable_zucchini).lower()]
|
|
if OPTIONS.enable_puffdiff is not None:
|
|
additional_args += ["--enable_puffdiff=" +
|
|
str(OPTIONS.enable_puffdiff).lower()]
|
|
|
|
if not ota_utils.IsLz4diffCompatible(source_file, target_file):
|
|
logger.warning(
|
|
"Source build doesn't support lz4diff, or source/target don't have compatible lz4diff versions. Disabling lz4diff.")
|
|
OPTIONS.enable_lz4diff = False
|
|
|
|
additional_args += ["--enable_lz4diff=" +
|
|
str(OPTIONS.enable_lz4diff).lower()]
|
|
|
|
if source_file and OPTIONS.enable_lz4diff:
|
|
input_tmp = common.UnzipTemp(source_file, ["META/liblz4.so"])
|
|
liblz4_path = os.path.join(input_tmp, "META", "liblz4.so")
|
|
assert os.path.exists(
|
|
liblz4_path), "liblz4.so not found in META/ dir of target file {}".format(liblz4_path)
|
|
logger.info("Enabling lz4diff %s", liblz4_path)
|
|
additional_args += ["--liblz4_path", liblz4_path]
|
|
erofs_compression_param = OPTIONS.target_info_dict.get(
|
|
"erofs_default_compressor")
|
|
assert erofs_compression_param is not None, "'erofs_default_compressor' not found in META/misc_info.txt of target build. This is required to enable lz4diff."
|
|
additional_args += ["--erofs_compression_param", erofs_compression_param]
|
|
|
|
if OPTIONS.disable_vabc:
|
|
additional_args += ["--disable_vabc=true"]
|
|
if OPTIONS.enable_vabc_xor:
|
|
additional_args += ["--enable_vabc_xor=true"]
|
|
if OPTIONS.compressor_types:
|
|
additional_args += ["--compressor_types", OPTIONS.compressor_types]
|
|
additional_args += ["--max_timestamp", max_timestamp]
|
|
|
|
payload.Generate(
|
|
target_file,
|
|
source_file,
|
|
additional_args + partition_timestamps_flags
|
|
)
|
|
|
|
# Sign the payload.
|
|
pw = OPTIONS.key_passwords[OPTIONS.package_key]
|
|
payload_signer = PayloadSigner(
|
|
OPTIONS.package_key, OPTIONS.private_key_suffix,
|
|
pw, OPTIONS.payload_signer)
|
|
payload.Sign(payload_signer)
|
|
|
|
# Write the payload into output zip.
|
|
payload.WriteToZip(output_zip)
|
|
|
|
# Generate and include the secondary payload that installs secondary images
|
|
# (e.g. system_other.img).
|
|
if OPTIONS.include_secondary:
|
|
# We always include a full payload for the secondary slot, even when
|
|
# building an incremental OTA. See the comments for "--include_secondary".
|
|
secondary_target_file = GetTargetFilesZipForSecondaryImages(
|
|
target_file, OPTIONS.skip_postinstall)
|
|
secondary_payload = PayloadGenerator(secondary=True)
|
|
secondary_payload.Generate(secondary_target_file,
|
|
additional_args=["--max_timestamp",
|
|
max_timestamp])
|
|
secondary_payload.Sign(payload_signer)
|
|
secondary_payload.WriteToZip(output_zip)
|
|
|
|
# If dm-verity is supported for the device, copy contents of care_map
|
|
# into A/B OTA package.
|
|
if target_info.get("avb_enable") == "true":
|
|
# Adds care_map if either the protobuf format or the plain text one exists.
|
|
for care_map_name in ["care_map.pb", "care_map.txt"]:
|
|
if not DoesInputFileContain(target_file, "META/" + care_map_name):
|
|
continue
|
|
care_map_data = common.ReadBytesFromInputFile(
|
|
target_file, "META/" + care_map_name)
|
|
# In order to support streaming, care_map needs to be packed as
|
|
# ZIP_STORED.
|
|
common.ZipWriteStr(output_zip, care_map_name, care_map_data,
|
|
compress_type=zipfile.ZIP_STORED)
|
|
# break here to avoid going into else when care map has been handled
|
|
break
|
|
else:
|
|
logger.warning("Cannot find care map file in target_file package")
|
|
|
|
# Add the source apex version for incremental ota updates, and write the
|
|
# result apex info to the ota package.
|
|
ota_apex_info = ota_utils.ConstructOtaApexInfo(target_file, source_file)
|
|
if ota_apex_info is not None:
|
|
common.ZipWriteStr(output_zip, "apex_info.pb", ota_apex_info,
|
|
compress_type=zipfile.ZIP_STORED)
|
|
|
|
# We haven't written the metadata entry yet, which will be handled in
|
|
# FinalizeMetadata().
|
|
common.ZipClose(output_zip)
|
|
|
|
FinalizeMetadata(metadata, staging_file, output_file,
|
|
package_key=OPTIONS.package_key)
|
|
|
|
|
|
def main(argv):
|
|
|
|
def option_handler(o, a):
|
|
if o in ("-i", "--incremental_from"):
|
|
OPTIONS.incremental_source = a
|
|
elif o == "--full_radio":
|
|
OPTIONS.full_radio = True
|
|
elif o == "--full_bootloader":
|
|
OPTIONS.full_bootloader = True
|
|
elif o == "--wipe_user_data":
|
|
OPTIONS.wipe_user_data = True
|
|
elif o == "--downgrade":
|
|
OPTIONS.downgrade = True
|
|
OPTIONS.wipe_user_data = True
|
|
elif o == "--override_timestamp":
|
|
OPTIONS.downgrade = True
|
|
elif o in ("-o", "--oem_settings"):
|
|
OPTIONS.oem_source = a.split(',')
|
|
elif o == "--oem_no_mount":
|
|
OPTIONS.oem_no_mount = True
|
|
elif o in ("-e", "--extra_script"):
|
|
OPTIONS.extra_script = a
|
|
elif o in ("-t", "--worker_threads"):
|
|
if a.isdigit():
|
|
OPTIONS.worker_threads = int(a)
|
|
else:
|
|
raise ValueError("Cannot parse value %r for option %r - only "
|
|
"integers are allowed." % (a, o))
|
|
elif o in ("-2", "--two_step"):
|
|
OPTIONS.two_step = True
|
|
elif o == "--include_secondary":
|
|
OPTIONS.include_secondary = True
|
|
elif o == "--no_signing":
|
|
OPTIONS.no_signing = True
|
|
elif o == "--verify":
|
|
OPTIONS.verify = True
|
|
elif o == "--block":
|
|
OPTIONS.block_based = True
|
|
elif o in ("-b", "--binary"):
|
|
OPTIONS.updater_binary = a
|
|
elif o == "--stash_threshold":
|
|
try:
|
|
OPTIONS.stash_threshold = float(a)
|
|
except ValueError:
|
|
raise ValueError("Cannot parse value %r for option %r - expecting "
|
|
"a float" % (a, o))
|
|
elif o == "--log_diff":
|
|
OPTIONS.log_diff = a
|
|
elif o == "--extracted_input_target_files":
|
|
OPTIONS.extracted_input = a
|
|
elif o == "--skip_postinstall":
|
|
OPTIONS.skip_postinstall = True
|
|
elif o == "--retrofit_dynamic_partitions":
|
|
OPTIONS.retrofit_dynamic_partitions = True
|
|
elif o == "--skip_compatibility_check":
|
|
OPTIONS.skip_compatibility_check = True
|
|
elif o == "--output_metadata_path":
|
|
OPTIONS.output_metadata_path = a
|
|
elif o == "--disable_fec_computation":
|
|
OPTIONS.disable_fec_computation = True
|
|
elif o == "--disable_verity_computation":
|
|
OPTIONS.disable_verity_computation = True
|
|
elif o == "--force_non_ab":
|
|
OPTIONS.force_non_ab = True
|
|
elif o == "--boot_variable_file":
|
|
OPTIONS.boot_variable_file = a
|
|
elif o == "--partial":
|
|
if a:
|
|
partitions = a.split()
|
|
if not partitions:
|
|
raise ValueError("Cannot parse partitions in {}".format(a))
|
|
else:
|
|
partitions = []
|
|
OPTIONS.partial = partitions
|
|
elif o == "--custom_image":
|
|
custom_partition, custom_image = a.split("=")
|
|
OPTIONS.custom_images[custom_partition] = custom_image
|
|
elif o == "--disable_vabc":
|
|
OPTIONS.disable_vabc = True
|
|
elif o == "--spl_downgrade":
|
|
OPTIONS.spl_downgrade = True
|
|
OPTIONS.wipe_user_data = True
|
|
elif o == "--vabc_downgrade":
|
|
OPTIONS.vabc_downgrade = True
|
|
elif o == "--enable_vabc_xor":
|
|
assert a.lower() in ["true", "false"]
|
|
OPTIONS.enable_vabc_xor = a.lower() != "false"
|
|
elif o == "--force_minor_version":
|
|
OPTIONS.force_minor_version = a
|
|
elif o == "--compressor_types":
|
|
OPTIONS.compressor_types = a
|
|
elif o == "--enable_zucchini":
|
|
assert a.lower() in ["true", "false"]
|
|
OPTIONS.enable_zucchini = a.lower() != "false"
|
|
elif o == "--enable_puffdiff":
|
|
assert a.lower() in ["true", "false"]
|
|
OPTIONS.enable_puffdiff = a.lower() != "false"
|
|
elif o == "--enable_lz4diff":
|
|
assert a.lower() in ["true", "false"]
|
|
OPTIONS.enable_lz4diff = a.lower() != "false"
|
|
elif o == "--vabc_compression_param":
|
|
words = a.split(",")
|
|
assert len(words) >= 1 and len(words) <= 2
|
|
OPTIONS.vabc_compression_param = a.lower()
|
|
if len(words) == 2:
|
|
if not words[1].lstrip("-").isdigit():
|
|
raise ValueError("Cannot parse value %r for option $COMPRESSION_LEVEL - only "
|
|
"integers are allowed." % words[1])
|
|
elif o == "--security_patch_level":
|
|
OPTIONS.security_patch_level = a
|
|
elif o in ("--max_threads"):
|
|
if a.isdigit():
|
|
OPTIONS.max_threads = a
|
|
else:
|
|
raise ValueError("Cannot parse value %r for option %r - only "
|
|
"integers are allowed." % (a, o))
|
|
elif o in ("--compression_factor"):
|
|
values = ["4k", "8k", "16k", "32k", "64k", "128k", "256k"]
|
|
if a[:-1].isdigit() and a in values and a.endswith("k"):
|
|
OPTIONS.compression_factor = str(int(a[:-1]) * 1024)
|
|
else:
|
|
raise ValueError("Please specify value from following options: 4k, 8k, 16k, 32k, 64k, 128k", "256k")
|
|
|
|
elif o == "--vabc_cow_version":
|
|
if a.isdigit():
|
|
OPTIONS.vabc_cow_version = a
|
|
else:
|
|
raise ValueError("Cannot parse value %r for option %r - only "
|
|
"integers are allowed." % (a, o))
|
|
else:
|
|
return False
|
|
return True
|
|
|
|
args = common.ParseOptions(argv, __doc__,
|
|
extra_opts="b:k:i:d:e:t:2o:",
|
|
extra_long_opts=[
|
|
"incremental_from=",
|
|
"full_radio",
|
|
"full_bootloader",
|
|
"wipe_user_data",
|
|
"downgrade",
|
|
"override_timestamp",
|
|
"extra_script=",
|
|
"worker_threads=",
|
|
"two_step",
|
|
"include_secondary",
|
|
"no_signing",
|
|
"block",
|
|
"binary=",
|
|
"oem_settings=",
|
|
"oem_no_mount",
|
|
"verify",
|
|
"stash_threshold=",
|
|
"log_diff=",
|
|
"extracted_input_target_files=",
|
|
"skip_postinstall",
|
|
"retrofit_dynamic_partitions",
|
|
"skip_compatibility_check",
|
|
"output_metadata_path=",
|
|
"disable_fec_computation",
|
|
"disable_verity_computation",
|
|
"force_non_ab",
|
|
"boot_variable_file=",
|
|
"partial=",
|
|
"custom_image=",
|
|
"disable_vabc",
|
|
"spl_downgrade",
|
|
"vabc_downgrade",
|
|
"enable_vabc_xor=",
|
|
"force_minor_version=",
|
|
"compressor_types=",
|
|
"enable_zucchini=",
|
|
"enable_puffdiff=",
|
|
"enable_lz4diff=",
|
|
"vabc_compression_param=",
|
|
"security_patch_level=",
|
|
"max_threads=",
|
|
"vabc_cow_version=",
|
|
"compression_factor=",
|
|
], extra_option_handler=[option_handler, payload_signer.signer_options])
|
|
common.InitLogging()
|
|
|
|
if len(args) != 2:
|
|
common.Usage(__doc__)
|
|
sys.exit(1)
|
|
|
|
# Load the build info dicts from the zip directly or the extracted input
|
|
# directory. We don't need to unzip the entire target-files zips, because they
|
|
# won't be needed for A/B OTAs (brillo_update_payload does that on its own).
|
|
# When loading the info dicts, we don't need to provide the second parameter
|
|
# to common.LoadInfoDict(). Specifying the second parameter allows replacing
|
|
# some properties with their actual paths, such as 'selinux_fc',
|
|
# 'ramdisk_dir', which won't be used during OTA generation.
|
|
if OPTIONS.extracted_input is not None:
|
|
OPTIONS.info_dict = common.LoadInfoDict(OPTIONS.extracted_input)
|
|
else:
|
|
OPTIONS.info_dict = common.LoadInfoDict(args[0])
|
|
|
|
if OPTIONS.wipe_user_data:
|
|
if not OPTIONS.vabc_downgrade:
|
|
logger.info("Detected downgrade/datawipe OTA."
|
|
"When wiping userdata, VABC OTA makes the user "
|
|
"wait in recovery mode for merge to finish. Disable VABC by "
|
|
"default. If you really want to do VABC downgrade, pass "
|
|
"--vabc_downgrade")
|
|
OPTIONS.disable_vabc = True
|
|
# We should only allow downgrading incrementals (as opposed to full).
|
|
# Otherwise the device may go back from arbitrary build with this full
|
|
# OTA package.
|
|
if OPTIONS.incremental_source is None and OPTIONS.downgrade:
|
|
raise ValueError("Cannot generate downgradable full OTAs")
|
|
|
|
# TODO(xunchang) for retrofit and partial updates, maybe we should rebuild the
|
|
# target-file and reload the info_dict. So the info will be consistent with
|
|
# the modified target-file.
|
|
|
|
logger.info("--- target info ---")
|
|
common.DumpInfoDict(OPTIONS.info_dict)
|
|
|
|
# Load the source build dict if applicable.
|
|
if OPTIONS.incremental_source is not None:
|
|
OPTIONS.target_info_dict = OPTIONS.info_dict
|
|
OPTIONS.source_info_dict = ParseInfoDict(OPTIONS.incremental_source)
|
|
|
|
logger.info("--- source info ---")
|
|
common.DumpInfoDict(OPTIONS.source_info_dict)
|
|
|
|
if OPTIONS.partial:
|
|
OPTIONS.info_dict['ab_partitions'] = \
|
|
list(
|
|
set(OPTIONS.info_dict['ab_partitions']) & set(OPTIONS.partial)
|
|
)
|
|
if OPTIONS.source_info_dict:
|
|
OPTIONS.source_info_dict['ab_partitions'] = \
|
|
list(
|
|
set(OPTIONS.source_info_dict['ab_partitions']) &
|
|
set(OPTIONS.partial)
|
|
)
|
|
|
|
# Load OEM dicts if provided.
|
|
OPTIONS.oem_dicts = _LoadOemDicts(OPTIONS.oem_source)
|
|
|
|
# Assume retrofitting dynamic partitions when base build does not set
|
|
# use_dynamic_partitions but target build does.
|
|
if (OPTIONS.source_info_dict and
|
|
OPTIONS.source_info_dict.get("use_dynamic_partitions") != "true" and
|
|
OPTIONS.target_info_dict.get("use_dynamic_partitions") == "true"):
|
|
if OPTIONS.target_info_dict.get("dynamic_partition_retrofit") != "true":
|
|
raise common.ExternalError(
|
|
"Expect to generate incremental OTA for retrofitting dynamic "
|
|
"partitions, but dynamic_partition_retrofit is not set in target "
|
|
"build.")
|
|
logger.info("Implicitly generating retrofit incremental OTA.")
|
|
OPTIONS.retrofit_dynamic_partitions = True
|
|
|
|
# Skip postinstall for retrofitting dynamic partitions.
|
|
if OPTIONS.retrofit_dynamic_partitions:
|
|
OPTIONS.skip_postinstall = True
|
|
|
|
ab_update = OPTIONS.info_dict.get("ab_update") == "true"
|
|
allow_non_ab = OPTIONS.info_dict.get("allow_non_ab") == "true"
|
|
if OPTIONS.force_non_ab:
|
|
assert allow_non_ab,\
|
|
"--force_non_ab only allowed on devices that supports non-A/B"
|
|
assert ab_update, "--force_non_ab only allowed on A/B devices"
|
|
|
|
generate_ab = not OPTIONS.force_non_ab and ab_update
|
|
|
|
# Use the default key to sign the package if not specified with package_key.
|
|
# package_keys are needed on ab_updates, so always define them if an
|
|
# A/B update is getting created.
|
|
if not OPTIONS.no_signing or generate_ab:
|
|
if OPTIONS.package_key is None:
|
|
OPTIONS.package_key = OPTIONS.info_dict.get(
|
|
"default_system_dev_certificate",
|
|
"build/make/target/product/security/testkey")
|
|
# Get signing keys
|
|
OPTIONS.key_passwords = common.GetKeyPasswords([OPTIONS.package_key])
|
|
|
|
# Only check for existence of key file if using the default signer.
|
|
# Because the custom signer might not need the key file AT all.
|
|
# b/191704641
|
|
if not OPTIONS.payload_signer:
|
|
private_key_path = OPTIONS.package_key + OPTIONS.private_key_suffix
|
|
if not os.path.exists(private_key_path):
|
|
raise common.ExternalError(
|
|
"Private key {} doesn't exist. Make sure you passed the"
|
|
" correct key path through -k option".format(
|
|
private_key_path)
|
|
)
|
|
signapk_abs_path = os.path.join(
|
|
OPTIONS.search_path, OPTIONS.signapk_path)
|
|
if not os.path.exists(signapk_abs_path):
|
|
raise common.ExternalError(
|
|
"Failed to find sign apk binary {} in search path {}. Make sure the correct search path is passed via -p".format(OPTIONS.signapk_path, OPTIONS.search_path))
|
|
|
|
if OPTIONS.source_info_dict:
|
|
source_build_prop = OPTIONS.source_info_dict["build.prop"]
|
|
target_build_prop = OPTIONS.target_info_dict["build.prop"]
|
|
source_spl = source_build_prop.GetProp(SECURITY_PATCH_LEVEL_PROP_NAME)
|
|
target_spl = target_build_prop.GetProp(SECURITY_PATCH_LEVEL_PROP_NAME)
|
|
is_spl_downgrade = target_spl < source_spl
|
|
if is_spl_downgrade and target_build_prop.GetProp("ro.build.tags") == "release-keys":
|
|
raise common.ExternalError(
|
|
"Target security patch level {} is older than source SPL {} "
|
|
"A locked bootloader will reject SPL downgrade no matter "
|
|
"what(even if data wipe is done), so SPL downgrade on any "
|
|
"release-keys build is not allowed.".format(target_spl, source_spl))
|
|
|
|
logger.info("SPL downgrade on %s",
|
|
target_build_prop.GetProp("ro.build.tags"))
|
|
if is_spl_downgrade and not OPTIONS.spl_downgrade and not OPTIONS.downgrade:
|
|
raise common.ExternalError(
|
|
"Target security patch level {} is older than source SPL {} applying "
|
|
"such OTA will likely cause device fail to boot. Pass --spl_downgrade "
|
|
"to override this check. This script expects security patch level to "
|
|
"be in format yyyy-mm-dd (e.x. 2021-02-05). It's possible to use "
|
|
"separators other than -, so as long as it's used consistenly across "
|
|
"all SPL dates".format(target_spl, source_spl))
|
|
elif not is_spl_downgrade and OPTIONS.spl_downgrade:
|
|
raise ValueError("--spl_downgrade specified but no actual SPL downgrade"
|
|
" detected. Please only pass in this flag if you want a"
|
|
" SPL downgrade. Target SPL: {} Source SPL: {}"
|
|
.format(target_spl, source_spl))
|
|
if generate_ab:
|
|
GenerateAbOtaPackage(
|
|
target_file=args[0],
|
|
output_file=args[1],
|
|
source_file=OPTIONS.incremental_source)
|
|
|
|
else:
|
|
GenerateNonAbOtaPackage(
|
|
target_file=args[0],
|
|
output_file=args[1],
|
|
source_file=OPTIONS.incremental_source)
|
|
|
|
# Post OTA generation works.
|
|
if OPTIONS.incremental_source is not None and OPTIONS.log_diff:
|
|
logger.info("Generating diff logs...")
|
|
logger.info("Unzipping target-files for diffing...")
|
|
target_dir = common.UnzipTemp(args[0], TARGET_DIFFING_UNZIP_PATTERN)
|
|
source_dir = common.UnzipTemp(
|
|
OPTIONS.incremental_source, TARGET_DIFFING_UNZIP_PATTERN)
|
|
|
|
with open(OPTIONS.log_diff, 'w') as out_file:
|
|
target_files_diff.recursiveDiff(
|
|
'', source_dir, target_dir, out_file)
|
|
|
|
logger.info("done.")
|
|
|
|
|
|
if __name__ == '__main__':
|
|
try:
|
|
common.CloseInheritedPipes()
|
|
main(sys.argv[1:])
|
|
finally:
|
|
common.Cleanup()
|