4aae62ee77
Add an `--avb-resolve-rollback-index-location-conflict` option in merge_target_files. When this option is set, the merge tool will resolve conflicting index locations by assigning the smallest unused index location. This is to support merging system and vendor target files from two different targets. In this case, the two target files may have conflict rollback index location because they were built independently. Test: atest releasetools_test Test: validate_target_files *-target_files-*.zip Test: merge_target_files && add_img_to_target_files && img_from_target_files && flash device Bug: 300604688 Change-Id: Ibd18ef2a9f3784157fe17966f5364c3c81c9bd9f
1264 lines
46 KiB
Python
1264 lines
46 KiB
Python
#!/usr/bin/env python
|
|
#
|
|
# Copyright (C) 2014 The Android Open Source Project
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
"""
|
|
Given a target-files zipfile that does not contain images (ie, does
|
|
not have an IMAGES/ top-level subdirectory), produce the images and
|
|
add them to the zipfile.
|
|
|
|
Usage: add_img_to_target_files [flag] target_files
|
|
|
|
-a (--add_missing)
|
|
Build and add missing images to "IMAGES/". If this option is
|
|
not specified, this script will simply exit when "IMAGES/"
|
|
directory exists in the target file.
|
|
|
|
-r (--rebuild_recovery)
|
|
Rebuild the recovery patch and write it to the system image. Only
|
|
meaningful when system image needs to be rebuilt and there're separate
|
|
boot / recovery images.
|
|
|
|
--replace_verity_private_key
|
|
Replace the private key used for verity signing. (same as the option
|
|
in sign_target_files_apks)
|
|
|
|
--replace_verity_public_key
|
|
Replace the certificate (public key) used for verity verification. (same
|
|
as the option in sign_target_files_apks)
|
|
|
|
--is_signing
|
|
Skip building & adding the images for "userdata" and "cache" if we
|
|
are signing the target files.
|
|
|
|
--avb-resolve-rollback-index-location-conflict
|
|
If provided, resolve the conflict AVB rollback index location when
|
|
necessary.
|
|
"""
|
|
|
|
from __future__ import print_function
|
|
|
|
import avbtool
|
|
import datetime
|
|
import logging
|
|
import os
|
|
import shlex
|
|
import shutil
|
|
import stat
|
|
import sys
|
|
import uuid
|
|
import tempfile
|
|
import zipfile
|
|
|
|
import build_image
|
|
import build_super_image
|
|
import common
|
|
import verity_utils
|
|
import ota_metadata_pb2
|
|
import rangelib
|
|
import sparse_img
|
|
from concurrent.futures import ThreadPoolExecutor
|
|
from apex_utils import GetApexInfoFromTargetFiles
|
|
from common import ZipDelete, PARTITIONS_WITH_CARE_MAP, ExternalError, RunAndCheckOutput, IsSparseImage, MakeTempFile, ZipWrite
|
|
from build_image import FIXED_FILE_TIMESTAMP
|
|
|
|
if sys.hexversion < 0x02070000:
|
|
print("Python 2.7 or newer is required.", file=sys.stderr)
|
|
sys.exit(1)
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
OPTIONS = common.OPTIONS
|
|
OPTIONS.add_missing = False
|
|
OPTIONS.rebuild_recovery = False
|
|
OPTIONS.replace_updated_files_list = []
|
|
OPTIONS.is_signing = False
|
|
OPTIONS.avb_resolve_rollback_index_location_conflict = False
|
|
|
|
|
|
def ParseAvbFooter(img_path) -> avbtool.AvbFooter:
|
|
with open(img_path, 'rb') as fp:
|
|
fp.seek(-avbtool.AvbFooter.SIZE, os.SEEK_END)
|
|
data = fp.read(avbtool.AvbFooter.SIZE)
|
|
return avbtool.AvbFooter(data)
|
|
|
|
|
|
def GetCareMap(which, imgname):
|
|
"""Returns the care_map string for the given partition.
|
|
|
|
Args:
|
|
which: The partition name, must be listed in PARTITIONS_WITH_CARE_MAP.
|
|
imgname: The filename of the image.
|
|
|
|
Returns:
|
|
(which, care_map_ranges): care_map_ranges is the raw string of the care_map
|
|
RangeSet; or None.
|
|
"""
|
|
assert which in PARTITIONS_WITH_CARE_MAP
|
|
|
|
is_sparse_img = IsSparseImage(imgname)
|
|
unsparsed_image_size = os.path.getsize(imgname)
|
|
|
|
# A verified image contains original image + hash tree data + FEC data
|
|
# + AVB footer, all concatenated together. The caremap specifies a range
|
|
# of blocks that update_verifier should read on top of dm-verity device
|
|
# to verify correctness of OTA updates. When reading off of dm-verity device,
|
|
# the hashtree and FEC part of image isn't available. So caremap should
|
|
# only contain the original image blocks.
|
|
try:
|
|
avbfooter = None
|
|
if is_sparse_img:
|
|
with tempfile.NamedTemporaryFile() as tmpfile:
|
|
img = sparse_img.SparseImage(imgname)
|
|
unsparsed_image_size = img.total_blocks * img.blocksize
|
|
for data in img.ReadBlocks(img.total_blocks - 1, 1):
|
|
tmpfile.write(data)
|
|
tmpfile.flush()
|
|
avbfooter = ParseAvbFooter(tmpfile.name)
|
|
else:
|
|
avbfooter = ParseAvbFooter(imgname)
|
|
except LookupError as e:
|
|
logger.warning(
|
|
"Failed to parse avbfooter for partition %s image %s, %s", which, imgname, e)
|
|
return None
|
|
|
|
image_size = avbfooter.original_image_size
|
|
assert image_size < unsparsed_image_size, f"AVB footer's original image size {image_size} is larger than or equal to image size on disk {unsparsed_image_size}, this can't happen because a verified image = original image + hash tree data + FEC data + avbfooter."
|
|
assert image_size > 0
|
|
|
|
image_blocks = int(image_size) // 4096 - 1
|
|
# It's OK for image_blocks to be 0, because care map ranges are inclusive.
|
|
# So 0-0 means "just block 0", which is valid.
|
|
assert image_blocks >= 0, "blocks for {} must be non-negative, image size: {}".format(
|
|
which, image_size)
|
|
|
|
# For sparse images, we will only check the blocks that are listed in the care
|
|
# map, i.e. the ones with meaningful data.
|
|
if is_sparse_img:
|
|
simg = sparse_img.SparseImage(imgname)
|
|
care_map_ranges = simg.care_map.intersect(
|
|
rangelib.RangeSet("0-{}".format(image_blocks)))
|
|
|
|
# Otherwise for non-sparse images, we read all the blocks in the filesystem
|
|
# image.
|
|
else:
|
|
care_map_ranges = rangelib.RangeSet("0-{}".format(image_blocks))
|
|
|
|
return [which, care_map_ranges.to_string_raw()]
|
|
|
|
|
|
def AddCareMapForAbOta(output_file, ab_partitions, image_paths):
|
|
"""Generates and adds care_map.pb for a/b partition that has care_map.
|
|
|
|
Args:
|
|
output_file: The output zip file (needs to be already open),
|
|
or file path to write care_map.pb.
|
|
ab_partitions: The list of A/B partitions.
|
|
image_paths: A map from the partition name to the image path.
|
|
"""
|
|
if not output_file:
|
|
raise ExternalError('Expected output_file for AddCareMapForAbOta')
|
|
|
|
care_map_list = []
|
|
for partition in ab_partitions:
|
|
partition = partition.strip()
|
|
if partition not in PARTITIONS_WITH_CARE_MAP:
|
|
continue
|
|
|
|
verity_block_device = "{}_verity_block_device".format(partition)
|
|
avb_hashtree_enable = "avb_{}_hashtree_enable".format(partition)
|
|
if (verity_block_device in OPTIONS.info_dict or
|
|
OPTIONS.info_dict.get(avb_hashtree_enable) == "true"):
|
|
if partition not in image_paths:
|
|
logger.warning('Potential partition with care_map missing from images: %s',
|
|
partition)
|
|
continue
|
|
image_path = image_paths[partition]
|
|
if not os.path.exists(image_path):
|
|
raise ExternalError('Expected image at path {}'.format(image_path))
|
|
|
|
care_map = GetCareMap(partition, image_path)
|
|
if not care_map:
|
|
continue
|
|
care_map_list += care_map
|
|
|
|
# adds fingerprint field to the care_map
|
|
# TODO(xunchang) revisit the fingerprint calculation for care_map.
|
|
partition_props = OPTIONS.info_dict.get(partition + ".build.prop")
|
|
prop_name_list = ["ro.{}.build.fingerprint".format(partition),
|
|
"ro.{}.build.thumbprint".format(partition)]
|
|
|
|
present_props = [x for x in prop_name_list if
|
|
partition_props and partition_props.GetProp(x)]
|
|
if not present_props:
|
|
logger.warning(
|
|
"fingerprint is not present for partition %s", partition)
|
|
property_id, fingerprint = "unknown", "unknown"
|
|
else:
|
|
property_id = present_props[0]
|
|
fingerprint = partition_props.GetProp(property_id)
|
|
care_map_list += [property_id, fingerprint]
|
|
|
|
if not care_map_list:
|
|
return
|
|
|
|
# Converts the list into proto buf message by calling care_map_generator; and
|
|
# writes the result to a temp file.
|
|
temp_care_map_text = MakeTempFile(prefix="caremap_text-",
|
|
suffix=".txt")
|
|
with open(temp_care_map_text, 'w') as text_file:
|
|
text_file.write('\n'.join(care_map_list))
|
|
|
|
temp_care_map = MakeTempFile(prefix="caremap-", suffix=".pb")
|
|
care_map_gen_cmd = ["care_map_generator", temp_care_map_text, temp_care_map]
|
|
RunAndCheckOutput(care_map_gen_cmd)
|
|
|
|
if not isinstance(output_file, zipfile.ZipFile):
|
|
shutil.copy(temp_care_map, output_file)
|
|
return
|
|
# output_file is a zip file
|
|
care_map_path = "META/care_map.pb"
|
|
if care_map_path in output_file.namelist():
|
|
# Copy the temp file into the OPTIONS.input_tmp dir and update the
|
|
# replace_updated_files_list used by add_img_to_target_files
|
|
if not OPTIONS.replace_updated_files_list:
|
|
OPTIONS.replace_updated_files_list = []
|
|
shutil.copy(temp_care_map, os.path.join(OPTIONS.input_tmp, care_map_path))
|
|
OPTIONS.replace_updated_files_list.append(care_map_path)
|
|
else:
|
|
ZipWrite(output_file, temp_care_map, arcname=care_map_path)
|
|
|
|
|
|
class OutputFile(object):
|
|
"""A helper class to write a generated file to the given dir or zip.
|
|
|
|
When generating images, we want the outputs to go into the given zip file, or
|
|
the given dir.
|
|
|
|
Attributes:
|
|
name: The name of the output file, regardless of the final destination.
|
|
"""
|
|
|
|
def __init__(self, output_zip, input_dir, *args):
|
|
# We write the intermediate output file under the given input_dir, even if
|
|
# the final destination is a zip archive.
|
|
self.name = os.path.join(input_dir, *args)
|
|
self._output_zip = output_zip
|
|
if self._output_zip:
|
|
self._zip_name = os.path.join(*args)
|
|
|
|
def Write(self, compress_type=None):
|
|
if self._output_zip:
|
|
common.ZipWrite(self._output_zip, self.name,
|
|
self._zip_name, compress_type=compress_type)
|
|
|
|
|
|
def AddSystem(output_zip, recovery_img=None, boot_img=None):
|
|
"""Turn the contents of SYSTEM into a system image and store it in
|
|
output_zip. Returns the name of the system image file."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "system.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("system.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
def output_sink(fn, data):
|
|
output_file = os.path.join(OPTIONS.input_tmp, "SYSTEM", fn)
|
|
with open(output_file, "wb") as ofile:
|
|
ofile.write(data)
|
|
|
|
if output_zip:
|
|
arc_name = "SYSTEM/" + fn
|
|
if arc_name in output_zip.namelist():
|
|
OPTIONS.replace_updated_files_list.append(arc_name)
|
|
else:
|
|
common.ZipWrite(output_zip, output_file, arc_name)
|
|
|
|
board_uses_vendorimage = OPTIONS.info_dict.get(
|
|
"board_uses_vendorimage") == "true"
|
|
|
|
if (OPTIONS.rebuild_recovery and not board_uses_vendorimage and
|
|
recovery_img is not None and boot_img is not None):
|
|
logger.info("Building new recovery patch on system at system/vendor")
|
|
common.MakeRecoveryPatch(OPTIONS.input_tmp, output_sink, recovery_img,
|
|
boot_img, info_dict=OPTIONS.info_dict)
|
|
|
|
block_list = OutputFile(output_zip, OPTIONS.input_tmp,
|
|
"IMAGES", "system.map")
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "system", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddSystemOther(output_zip):
|
|
"""Turn the contents of SYSTEM_OTHER into a system_other image
|
|
and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "system_other.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("system_other.img already exists; no need to rebuild...")
|
|
return
|
|
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "system_other", img)
|
|
|
|
|
|
def AddVendor(output_zip, recovery_img=None, boot_img=None):
|
|
"""Turn the contents of VENDOR into a vendor image and store in it
|
|
output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "vendor.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("vendor.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
def output_sink(fn, data):
|
|
output_file = os.path.join(OPTIONS.input_tmp, "VENDOR", fn)
|
|
with open(output_file, "wb") as ofile:
|
|
ofile.write(data)
|
|
|
|
if output_zip:
|
|
arc_name = "VENDOR/" + fn
|
|
if arc_name in output_zip.namelist():
|
|
OPTIONS.replace_updated_files_list.append(arc_name)
|
|
else:
|
|
common.ZipWrite(output_zip, output_file, arc_name)
|
|
|
|
board_uses_vendorimage = OPTIONS.info_dict.get(
|
|
"board_uses_vendorimage") == "true"
|
|
|
|
if (OPTIONS.rebuild_recovery and board_uses_vendorimage and
|
|
recovery_img is not None and boot_img is not None):
|
|
logger.info("Building new recovery patch on vendor")
|
|
common.MakeRecoveryPatch(OPTIONS.input_tmp, output_sink, recovery_img,
|
|
boot_img, info_dict=OPTIONS.info_dict)
|
|
|
|
block_list = OutputFile(output_zip, OPTIONS.input_tmp,
|
|
"IMAGES", "vendor.map")
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "vendor", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddProduct(output_zip):
|
|
"""Turn the contents of PRODUCT into a product image and store it in
|
|
output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "product.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("product.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "product.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "product", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddSystemExt(output_zip):
|
|
"""Turn the contents of SYSTEM_EXT into a system_ext image and store it in
|
|
output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES",
|
|
"system_ext.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("system_ext.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "system_ext.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "system_ext", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddOdm(output_zip):
|
|
"""Turn the contents of ODM into an odm image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "odm.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("odm.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "odm.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "odm", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddVendorDlkm(output_zip):
|
|
"""Turn the contents of VENDOR_DLKM into an vendor_dlkm image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "vendor_dlkm.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("vendor_dlkm.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "vendor_dlkm.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "vendor_dlkm", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddOdmDlkm(output_zip):
|
|
"""Turn the contents of OdmDlkm into an odm_dlkm image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "odm_dlkm.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("odm_dlkm.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "odm_dlkm.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "odm_dlkm", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddSystemDlkm(output_zip):
|
|
"""Turn the contents of SystemDlkm into an system_dlkm image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "system_dlkm.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("system_dlkm.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "system_dlkm.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "system_dlkm", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddDtbo(output_zip):
|
|
"""Adds the DTBO image.
|
|
|
|
Uses the image under IMAGES/ if it already exists. Otherwise looks for the
|
|
image under PREBUILT_IMAGES/, signs it as needed, and returns the image name.
|
|
"""
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "dtbo.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("dtbo.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
dtbo_prebuilt_path = os.path.join(
|
|
OPTIONS.input_tmp, "PREBUILT_IMAGES", "dtbo.img")
|
|
assert os.path.exists(dtbo_prebuilt_path)
|
|
shutil.copy(dtbo_prebuilt_path, img.name)
|
|
|
|
# AVB-sign the image as needed.
|
|
if OPTIONS.info_dict.get("avb_enable") == "true":
|
|
# Signing requires +w
|
|
os.chmod(img.name, os.stat(img.name).st_mode | stat.S_IWUSR)
|
|
|
|
avbtool = OPTIONS.info_dict["avb_avbtool"]
|
|
part_size = OPTIONS.info_dict["dtbo_size"]
|
|
# The AVB hash footer will be replaced if already present.
|
|
cmd = [avbtool, "add_hash_footer", "--image", img.name,
|
|
"--partition_size", str(part_size), "--partition_name", "dtbo"]
|
|
common.AppendAVBSigningArgs(cmd, "dtbo")
|
|
args = OPTIONS.info_dict.get("avb_dtbo_add_hash_footer_args")
|
|
if args and args.strip():
|
|
cmd.extend(shlex.split(args))
|
|
common.RunAndCheckOutput(cmd)
|
|
|
|
img.Write()
|
|
return img.name
|
|
|
|
|
|
def AddPvmfw(output_zip):
|
|
"""Adds the pvmfw image.
|
|
|
|
Uses the image under IMAGES/ if it already exists. Otherwise looks for the
|
|
image under PREBUILT_IMAGES/, signs it as needed, and returns the image name.
|
|
"""
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "pvmfw.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("pvmfw.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
pvmfw_prebuilt_path = os.path.join(
|
|
OPTIONS.input_tmp, "PREBUILT_IMAGES", "pvmfw.img")
|
|
assert os.path.exists(pvmfw_prebuilt_path)
|
|
shutil.copy(pvmfw_prebuilt_path, img.name)
|
|
|
|
# AVB-sign the image as needed.
|
|
if OPTIONS.info_dict.get("avb_enable") == "true":
|
|
# Signing requires +w
|
|
os.chmod(img.name, os.stat(img.name).st_mode | stat.S_IWUSR)
|
|
|
|
avbtool = OPTIONS.info_dict["avb_avbtool"]
|
|
part_size = OPTIONS.info_dict["pvmfw_size"]
|
|
# The AVB hash footer will be replaced if already present.
|
|
cmd = [avbtool, "add_hash_footer", "--image", img.name,
|
|
"--partition_size", str(part_size), "--partition_name", "pvmfw"]
|
|
common.AppendAVBSigningArgs(cmd, "pvmfw")
|
|
args = OPTIONS.info_dict.get("avb_pvmfw_add_hash_footer_args")
|
|
if args and args.strip():
|
|
cmd.extend(shlex.split(args))
|
|
common.RunAndCheckOutput(cmd)
|
|
|
|
img.Write()
|
|
return img.name
|
|
|
|
|
|
def AddCustomImages(output_zip, partition_name, image_list):
|
|
"""Adds and signs avb custom images as needed in IMAGES/.
|
|
|
|
Args:
|
|
output_zip: The output zip file (needs to be already open), or None to
|
|
write images to OPTIONS.input_tmp/.
|
|
partition_name: The custom image partition name.
|
|
image_list: The image list of the custom image partition.
|
|
|
|
Uses the image under IMAGES/ if it already exists. Otherwise looks for the
|
|
image under PREBUILT_IMAGES/, signs it as needed, and returns the image name.
|
|
|
|
Raises:
|
|
AssertionError: If image can't be found.
|
|
"""
|
|
|
|
builder = None
|
|
key_path = OPTIONS.info_dict.get("avb_{}_key_path".format(partition_name))
|
|
if key_path is not None:
|
|
algorithm = OPTIONS.info_dict.get("avb_{}_algorithm".format(partition_name))
|
|
extra_args = OPTIONS.info_dict.get(
|
|
"avb_{}_add_hashtree_footer_args".format(partition_name))
|
|
partition_size = OPTIONS.info_dict.get(
|
|
"avb_{}_partition_size".format(partition_name))
|
|
|
|
builder = verity_utils.CreateCustomImageBuilder(
|
|
OPTIONS.info_dict, partition_name, partition_size,
|
|
key_path, algorithm, extra_args)
|
|
|
|
for img_name in image_list:
|
|
custom_image = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", img_name)
|
|
if os.path.exists(custom_image.name):
|
|
continue
|
|
|
|
custom_image_prebuilt_path = os.path.join(
|
|
OPTIONS.input_tmp, "PREBUILT_IMAGES", img_name)
|
|
assert os.path.exists(custom_image_prebuilt_path), \
|
|
"Failed to find %s at %s" % (img_name, custom_image_prebuilt_path)
|
|
|
|
shutil.copy(custom_image_prebuilt_path, custom_image.name)
|
|
|
|
if builder is not None:
|
|
builder.Build(custom_image.name)
|
|
|
|
custom_image.Write()
|
|
|
|
default = os.path.join(OPTIONS.input_tmp, "IMAGES", partition_name + ".img")
|
|
assert os.path.exists(default), \
|
|
"There should be one %s.img" % (partition_name)
|
|
return default
|
|
|
|
|
|
def CreateImage(input_dir, info_dict, what, output_file, block_list=None):
|
|
logger.info("creating %s.img...", what)
|
|
|
|
image_props = build_image.ImagePropFromGlobalDict(info_dict, what)
|
|
image_props["timestamp"] = FIXED_FILE_TIMESTAMP
|
|
|
|
if what == "system":
|
|
fs_config_prefix = ""
|
|
else:
|
|
fs_config_prefix = what + "_"
|
|
|
|
fs_config = os.path.join(
|
|
input_dir, "META/" + fs_config_prefix + "filesystem_config.txt")
|
|
if not os.path.exists(fs_config):
|
|
fs_config = None
|
|
|
|
# Override values loaded from info_dict.
|
|
if fs_config:
|
|
image_props["fs_config"] = fs_config
|
|
if block_list:
|
|
image_props["block_list"] = block_list.name
|
|
|
|
build_image.BuildImage(
|
|
os.path.join(input_dir, what.upper()), image_props, output_file.name)
|
|
|
|
output_file.Write()
|
|
if block_list:
|
|
block_list.Write()
|
|
|
|
# Set the '_image_size' for given image size.
|
|
is_verity_partition = "verity_block_device" in image_props
|
|
verity_supported = (image_props.get("avb_enable") == "true")
|
|
is_avb_enable = image_props.get("avb_hashtree_enable") == "true"
|
|
if verity_supported and (is_verity_partition or is_avb_enable):
|
|
image_size = image_props.get("image_size")
|
|
if image_size:
|
|
image_size_key = what + "_image_size"
|
|
info_dict[image_size_key] = int(image_size)
|
|
|
|
use_dynamic_size = (
|
|
info_dict.get("use_dynamic_partition_size") == "true" and
|
|
what in shlex.split(info_dict.get("dynamic_partition_list", "").strip()))
|
|
if use_dynamic_size:
|
|
info_dict.update(build_image.GlobalDictFromImageProp(image_props, what))
|
|
|
|
|
|
def AddUserdata(output_zip):
|
|
"""Create a userdata image and store it in output_zip.
|
|
|
|
In most case we just create and store an empty userdata.img;
|
|
But the invoker can also request to create userdata.img with real
|
|
data from the target files, by setting "userdata_img_with_data=true"
|
|
in OPTIONS.info_dict.
|
|
"""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "userdata.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("userdata.img already exists; no need to rebuild...")
|
|
return
|
|
|
|
# Skip userdata.img if no size.
|
|
image_props = build_image.ImagePropFromGlobalDict(OPTIONS.info_dict, "data")
|
|
if not image_props.get("partition_size"):
|
|
return
|
|
|
|
logger.info("creating userdata.img...")
|
|
|
|
image_props["timestamp"] = FIXED_FILE_TIMESTAMP
|
|
|
|
if OPTIONS.info_dict.get("userdata_img_with_data") == "true":
|
|
user_dir = os.path.join(OPTIONS.input_tmp, "DATA")
|
|
else:
|
|
user_dir = common.MakeTempDir()
|
|
|
|
build_image.BuildImage(user_dir, image_props, img.name)
|
|
|
|
common.CheckSize(img.name, "userdata.img", OPTIONS.info_dict)
|
|
# Always use compression for useradata image.
|
|
# As it's likely huge and consist of lots of 0s.
|
|
img.Write(zipfile.ZIP_DEFLATED)
|
|
|
|
|
|
def AddVBMeta(output_zip, partitions, name, needed_partitions):
|
|
"""Creates a VBMeta image and stores it in output_zip.
|
|
|
|
It generates the requested VBMeta image. The requested image could be for
|
|
top-level or chained VBMeta image, which is determined based on the name.
|
|
|
|
Args:
|
|
output_zip: The output zip file, which needs to be already open.
|
|
partitions: A dict that's keyed by partition names with image paths as
|
|
values. Only valid partition names are accepted, as partitions listed
|
|
in common.AVB_PARTITIONS and custom partitions listed in
|
|
OPTIONS.info_dict.get("avb_custom_images_partition_list")
|
|
name: Name of the VBMeta partition, e.g. 'vbmeta', 'vbmeta_system'.
|
|
needed_partitions: Partitions whose descriptors should be included into the
|
|
generated VBMeta image.
|
|
|
|
Returns:
|
|
Path to the created image.
|
|
|
|
Raises:
|
|
AssertionError: On invalid input args.
|
|
"""
|
|
assert needed_partitions, "Needed partitions must be specified"
|
|
|
|
img = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "{}.img".format(name))
|
|
if os.path.exists(img.name):
|
|
logger.info("%s.img already exists; not rebuilding...", name)
|
|
return img.name
|
|
|
|
common.BuildVBMeta(img.name, partitions, name, needed_partitions,
|
|
OPTIONS.avb_resolve_rollback_index_location_conflict)
|
|
img.Write()
|
|
return img.name
|
|
|
|
|
|
def AddCache(output_zip):
|
|
"""Create an empty cache image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "cache.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("cache.img already exists; no need to rebuild...")
|
|
return
|
|
|
|
image_props = build_image.ImagePropFromGlobalDict(OPTIONS.info_dict, "cache")
|
|
# The build system has to explicitly request for cache.img.
|
|
if "fs_type" not in image_props:
|
|
return
|
|
|
|
logger.info("creating cache.img...")
|
|
|
|
image_props["timestamp"] = FIXED_FILE_TIMESTAMP
|
|
|
|
user_dir = common.MakeTempDir()
|
|
build_image.BuildImage(user_dir, image_props, img.name)
|
|
|
|
common.CheckSize(img.name, "cache.img", OPTIONS.info_dict)
|
|
img.Write()
|
|
|
|
|
|
def CheckAbOtaImages(output_zip, ab_partitions):
|
|
"""Checks that all the listed A/B partitions have their images available.
|
|
|
|
The images need to be available under IMAGES/ or RADIO/, with the former takes
|
|
a priority.
|
|
|
|
Args:
|
|
output_zip: The output zip file (needs to be already open), or None to
|
|
find images in OPTIONS.input_tmp/.
|
|
ab_partitions: The list of A/B partitions.
|
|
|
|
Raises:
|
|
AssertionError: If it can't find an image.
|
|
"""
|
|
for partition in ab_partitions:
|
|
img_name = partition + ".img"
|
|
|
|
# Assert that the image is present under IMAGES/ now.
|
|
if output_zip:
|
|
# Zip spec says: All slashes MUST be forward slashes.
|
|
images_path = "IMAGES/" + img_name
|
|
radio_path = "RADIO/" + img_name
|
|
available = (images_path in output_zip.namelist() or
|
|
radio_path in output_zip.namelist())
|
|
else:
|
|
images_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name)
|
|
radio_path = os.path.join(OPTIONS.input_tmp, "RADIO", img_name)
|
|
available = os.path.exists(images_path) or os.path.exists(radio_path)
|
|
|
|
assert available, "Failed to find " + img_name
|
|
|
|
|
|
def AddPackRadioImages(output_zip, images):
|
|
"""Copies images listed in META/pack_radioimages.txt from RADIO/ to IMAGES/.
|
|
|
|
Args:
|
|
output_zip: The output zip file (needs to be already open), or None to
|
|
write images to OPTIONS.input_tmp/.
|
|
images: A list of image names.
|
|
|
|
Raises:
|
|
AssertionError: If a listed image can't be found.
|
|
"""
|
|
for image in images:
|
|
img_name = image.strip()
|
|
_, ext = os.path.splitext(img_name)
|
|
if not ext:
|
|
img_name += ".img"
|
|
|
|
prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name)
|
|
if os.path.exists(prebuilt_path):
|
|
logger.info("%s already exists, no need to overwrite...", img_name)
|
|
continue
|
|
|
|
img_radio_path = os.path.join(OPTIONS.input_tmp, "RADIO", img_name)
|
|
assert os.path.exists(img_radio_path), \
|
|
"Failed to find %s at %s" % (img_name, img_radio_path)
|
|
|
|
if output_zip:
|
|
common.ZipWrite(output_zip, img_radio_path, "IMAGES/" + img_name)
|
|
else:
|
|
shutil.copy(img_radio_path, prebuilt_path)
|
|
|
|
|
|
def AddSuperEmpty(output_zip):
|
|
"""Create a super_empty.img and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "super_empty.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("super_empty.img already exists; no need to rebuild...")
|
|
return
|
|
build_super_image.BuildSuperImage(OPTIONS.info_dict, img.name)
|
|
img.Write()
|
|
|
|
|
|
def AddSuperSplit(output_zip):
|
|
"""Create split super_*.img and store it in output_zip."""
|
|
|
|
outdir = os.path.join(OPTIONS.input_tmp, "OTA")
|
|
built = build_super_image.BuildSuperImage(OPTIONS.input_tmp, outdir)
|
|
|
|
if built:
|
|
for dev in OPTIONS.info_dict['super_block_devices'].strip().split():
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "OTA",
|
|
"super_" + dev + ".img")
|
|
img.Write()
|
|
|
|
|
|
def ReplaceUpdatedFiles(zip_filename, files_list):
|
|
"""Updates all the ZIP entries listed in files_list.
|
|
|
|
For now the list includes META/care_map.pb, and the related files under
|
|
SYSTEM/ after rebuilding recovery.
|
|
"""
|
|
common.ZipDelete(zip_filename, files_list)
|
|
output_zip = zipfile.ZipFile(zip_filename, "a",
|
|
compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True)
|
|
for item in files_list:
|
|
file_path = os.path.join(OPTIONS.input_tmp, item)
|
|
assert os.path.exists(file_path)
|
|
common.ZipWrite(output_zip, file_path, arcname=item)
|
|
common.ZipClose(output_zip)
|
|
|
|
|
|
def HasPartition(partition_name):
|
|
"""Determines if the target files archive should build a given partition."""
|
|
|
|
return ((os.path.isdir(
|
|
os.path.join(OPTIONS.input_tmp, partition_name.upper())) and
|
|
OPTIONS.info_dict.get(
|
|
"building_{}_image".format(partition_name)) == "true") or
|
|
os.path.exists(
|
|
os.path.join(OPTIONS.input_tmp, "IMAGES",
|
|
"{}.img".format(partition_name))))
|
|
|
|
|
|
def AddApexInfo(output_zip):
|
|
apex_infos = GetApexInfoFromTargetFiles(OPTIONS.input_tmp, 'system',
|
|
compressed_only=False)
|
|
apex_metadata_proto = ota_metadata_pb2.ApexMetadata()
|
|
apex_metadata_proto.apex_info.extend(apex_infos)
|
|
apex_info_bytes = apex_metadata_proto.SerializeToString()
|
|
|
|
output_file = os.path.join(OPTIONS.input_tmp, "META", "apex_info.pb")
|
|
with open(output_file, "wb") as ofile:
|
|
ofile.write(apex_info_bytes)
|
|
if output_zip:
|
|
arc_name = "META/apex_info.pb"
|
|
if arc_name in output_zip.namelist():
|
|
OPTIONS.replace_updated_files_list.append(arc_name)
|
|
else:
|
|
common.ZipWrite(output_zip, output_file, arc_name)
|
|
|
|
|
|
def AddVbmetaDigest(output_zip):
|
|
"""Write the vbmeta digest to the output dir and zipfile."""
|
|
|
|
# Calculate the vbmeta digest and put the result in to META/
|
|
boot_images = OPTIONS.info_dict.get("boot_images")
|
|
# Disable the digest calculation if the target_file is used as a container
|
|
# for boot images. A boot container might contain boot-5.4.img, boot-5.10.img
|
|
# etc., instead of just a boot.img and will fail in vbmeta digest calculation.
|
|
boot_container = boot_images and (
|
|
len(boot_images.split()) >= 2 or boot_images.split()[0] != 'boot.img')
|
|
if (OPTIONS.info_dict.get("avb_enable") == "true" and not boot_container and
|
|
OPTIONS.info_dict.get("avb_building_vbmeta_image") == "true"):
|
|
avbtool = OPTIONS.info_dict["avb_avbtool"]
|
|
digest = verity_utils.CalculateVbmetaDigest(OPTIONS.input_tmp, avbtool)
|
|
vbmeta_digest_txt = os.path.join(OPTIONS.input_tmp, "META",
|
|
"vbmeta_digest.txt")
|
|
with open(vbmeta_digest_txt, 'w') as f:
|
|
f.write(digest)
|
|
# writes to the output zipfile
|
|
if output_zip:
|
|
arc_name = "META/vbmeta_digest.txt"
|
|
if arc_name in output_zip.namelist():
|
|
OPTIONS.replace_updated_files_list.append(arc_name)
|
|
else:
|
|
common.ZipWriteStr(output_zip, arc_name, digest)
|
|
|
|
|
|
def AddImagesToTargetFiles(filename):
|
|
"""Creates and adds images (boot/recovery/system/...) to a target_files.zip.
|
|
|
|
It works with either a zip file (zip mode), or a directory that contains the
|
|
files to be packed into a target_files.zip (dir mode). The latter is used when
|
|
being called from build/make/core/Makefile.
|
|
|
|
The images will be created under IMAGES/ in the input target_files.zip.
|
|
|
|
Args:
|
|
filename: the target_files.zip, or the zip root directory.
|
|
"""
|
|
if os.path.isdir(filename):
|
|
OPTIONS.input_tmp = os.path.abspath(filename)
|
|
else:
|
|
OPTIONS.input_tmp = common.UnzipTemp(filename)
|
|
|
|
if not OPTIONS.add_missing:
|
|
if os.path.isdir(os.path.join(OPTIONS.input_tmp, "IMAGES")):
|
|
logger.warning("target_files appears to already contain images.")
|
|
sys.exit(1)
|
|
|
|
OPTIONS.info_dict = common.LoadInfoDict(OPTIONS.input_tmp, repacking=True)
|
|
|
|
has_recovery = OPTIONS.info_dict.get("no_recovery") != "true"
|
|
has_boot = OPTIONS.info_dict.get("no_boot") != "true"
|
|
has_init_boot = OPTIONS.info_dict.get("init_boot") == "true"
|
|
has_vendor_boot = OPTIONS.info_dict.get("vendor_boot") == "true"
|
|
has_vendor_kernel_boot = OPTIONS.info_dict.get(
|
|
"vendor_kernel_boot") == "true"
|
|
|
|
# {vendor,odm,product,system_ext,vendor_dlkm,odm_dlkm, system_dlkm, system, system_other}.img
|
|
# can be built from source, or dropped into target_files.zip as a prebuilt blob.
|
|
has_vendor = HasPartition("vendor")
|
|
has_odm = HasPartition("odm")
|
|
has_vendor_dlkm = HasPartition("vendor_dlkm")
|
|
has_odm_dlkm = HasPartition("odm_dlkm")
|
|
has_system_dlkm = HasPartition("system_dlkm")
|
|
has_product = HasPartition("product")
|
|
has_system_ext = HasPartition("system_ext")
|
|
has_system = HasPartition("system")
|
|
has_system_other = HasPartition("system_other")
|
|
has_userdata = OPTIONS.info_dict.get("building_userdata_image") == "true"
|
|
has_cache = OPTIONS.info_dict.get("building_cache_image") == "true"
|
|
|
|
# Set up the output destination. It writes to the given directory for dir
|
|
# mode; otherwise appends to the given ZIP.
|
|
if os.path.isdir(filename):
|
|
output_zip = None
|
|
else:
|
|
output_zip = zipfile.ZipFile(filename, "a",
|
|
compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True)
|
|
|
|
# Always make input_tmp/IMAGES available, since we may stage boot / recovery
|
|
# images there even under zip mode. The directory will be cleaned up as part
|
|
# of OPTIONS.input_tmp.
|
|
images_dir = os.path.join(OPTIONS.input_tmp, "IMAGES")
|
|
if not os.path.isdir(images_dir):
|
|
os.makedirs(images_dir)
|
|
|
|
# A map between partition names and their paths, which could be used when
|
|
# generating AVB vbmeta image.
|
|
partitions = {}
|
|
|
|
def banner(s):
|
|
logger.info("\n\n++++ %s ++++\n\n", s)
|
|
|
|
boot_image = None
|
|
if has_boot:
|
|
banner("boot")
|
|
boot_images = OPTIONS.info_dict.get("boot_images")
|
|
if boot_images is None:
|
|
boot_images = "boot.img"
|
|
for index, b in enumerate(boot_images.split()):
|
|
# common.GetBootableImage() returns the image directly if present.
|
|
boot_image = common.GetBootableImage(
|
|
"IMAGES/" + b, b, OPTIONS.input_tmp, "BOOT")
|
|
# boot.img may be unavailable in some targets (e.g. aosp_arm64).
|
|
if boot_image:
|
|
boot_image_path = os.path.join(OPTIONS.input_tmp, "IMAGES", b)
|
|
# Although multiple boot images can be generated, include the image
|
|
# descriptor of only the first boot image in vbmeta
|
|
if index == 0:
|
|
partitions['boot'] = boot_image_path
|
|
if not os.path.exists(boot_image_path):
|
|
boot_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
boot_image.AddToZip(output_zip)
|
|
|
|
if has_init_boot:
|
|
banner("init_boot")
|
|
init_boot_image = common.GetBootableImage(
|
|
"IMAGES/init_boot.img", "init_boot.img", OPTIONS.input_tmp, "INIT_BOOT",
|
|
dev_nodes=True)
|
|
if init_boot_image:
|
|
partitions['init_boot'] = os.path.join(
|
|
OPTIONS.input_tmp, "IMAGES", "init_boot.img")
|
|
if not os.path.exists(partitions['init_boot']):
|
|
init_boot_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
init_boot_image.AddToZip(output_zip)
|
|
|
|
if has_vendor_boot:
|
|
banner("vendor_boot")
|
|
vendor_boot_image = common.GetVendorBootImage(
|
|
"IMAGES/vendor_boot.img", "vendor_boot.img", OPTIONS.input_tmp,
|
|
"VENDOR_BOOT")
|
|
if vendor_boot_image:
|
|
partitions['vendor_boot'] = os.path.join(OPTIONS.input_tmp, "IMAGES",
|
|
"vendor_boot.img")
|
|
if not os.path.exists(partitions['vendor_boot']):
|
|
vendor_boot_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
vendor_boot_image.AddToZip(output_zip)
|
|
|
|
if has_vendor_kernel_boot:
|
|
banner("vendor_kernel_boot")
|
|
vendor_kernel_boot_image = common.GetVendorKernelBootImage(
|
|
"IMAGES/vendor_kernel_boot.img", "vendor_kernel_boot.img", OPTIONS.input_tmp,
|
|
"VENDOR_KERNEL_BOOT")
|
|
if vendor_kernel_boot_image:
|
|
partitions['vendor_kernel_boot'] = os.path.join(OPTIONS.input_tmp, "IMAGES",
|
|
"vendor_kernel_boot.img")
|
|
if not os.path.exists(partitions['vendor_kernel_boot']):
|
|
vendor_kernel_boot_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
vendor_kernel_boot_image.AddToZip(output_zip)
|
|
|
|
recovery_image = None
|
|
if has_recovery:
|
|
banner("recovery")
|
|
recovery_image = common.GetBootableImage(
|
|
"IMAGES/recovery.img", "recovery.img", OPTIONS.input_tmp, "RECOVERY")
|
|
assert recovery_image, "Failed to create recovery.img."
|
|
partitions['recovery'] = os.path.join(
|
|
OPTIONS.input_tmp, "IMAGES", "recovery.img")
|
|
if not os.path.exists(partitions['recovery']):
|
|
recovery_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
recovery_image.AddToZip(output_zip)
|
|
|
|
banner("recovery (two-step image)")
|
|
# The special recovery.img for two-step package use.
|
|
recovery_two_step_image = common.GetBootableImage(
|
|
"OTA/recovery-two-step.img", "recovery-two-step.img",
|
|
OPTIONS.input_tmp, "RECOVERY", two_step_image=True)
|
|
assert recovery_two_step_image, "Failed to create recovery-two-step.img."
|
|
recovery_two_step_image_path = os.path.join(
|
|
OPTIONS.input_tmp, "OTA", "recovery-two-step.img")
|
|
if not os.path.exists(recovery_two_step_image_path):
|
|
recovery_two_step_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
recovery_two_step_image.AddToZip(output_zip)
|
|
|
|
def add_partition(partition, has_partition, add_func, add_args):
|
|
if has_partition:
|
|
banner(partition)
|
|
partitions[partition] = add_func(output_zip, *add_args)
|
|
|
|
add_partition_calls = (
|
|
("system", has_system, AddSystem, [recovery_image, boot_image]),
|
|
("vendor", has_vendor, AddVendor, [recovery_image, boot_image]),
|
|
("product", has_product, AddProduct, []),
|
|
("system_ext", has_system_ext, AddSystemExt, []),
|
|
("odm", has_odm, AddOdm, []),
|
|
("vendor_dlkm", has_vendor_dlkm, AddVendorDlkm, []),
|
|
("odm_dlkm", has_odm_dlkm, AddOdmDlkm, []),
|
|
("system_dlkm", has_system_dlkm, AddSystemDlkm, []),
|
|
("system_other", has_system_other, AddSystemOther, []),
|
|
)
|
|
# If output_zip exists, each add_partition_calls writes bytes to the same output_zip,
|
|
# which is not thread-safe. So, run them in serial if output_zip exists.
|
|
if output_zip:
|
|
for call in add_partition_calls:
|
|
add_partition(*call)
|
|
else:
|
|
with ThreadPoolExecutor(max_workers=len(add_partition_calls)) as executor:
|
|
for future in [executor.submit(add_partition, *call) for call in add_partition_calls]:
|
|
future.result()
|
|
|
|
AddApexInfo(output_zip)
|
|
|
|
if not OPTIONS.is_signing:
|
|
banner("userdata")
|
|
AddUserdata(output_zip)
|
|
banner("cache")
|
|
AddCache(output_zip)
|
|
|
|
add_partition("dtbo",
|
|
OPTIONS.info_dict.get("has_dtbo") == "true", AddDtbo, [])
|
|
add_partition("pvmfw",
|
|
OPTIONS.info_dict.get("has_pvmfw") == "true", AddPvmfw, [])
|
|
|
|
# Custom images.
|
|
custom_partitions = OPTIONS.info_dict.get(
|
|
"custom_images_partition_list", "").strip().split()
|
|
for partition_name in custom_partitions:
|
|
partition_name = partition_name.strip()
|
|
banner("custom images for " + partition_name)
|
|
image_list = OPTIONS.info_dict.get(
|
|
"{}_image_list".format(partition_name)).split()
|
|
partitions[partition_name] = AddCustomImages(output_zip, partition_name, image_list)
|
|
|
|
avb_custom_partitions = OPTIONS.info_dict.get(
|
|
"avb_custom_images_partition_list", "").strip().split()
|
|
for partition_name in avb_custom_partitions:
|
|
partition_name = partition_name.strip()
|
|
banner("avb custom images for " + partition_name)
|
|
image_list = OPTIONS.info_dict.get(
|
|
"avb_{}_image_list".format(partition_name)).split()
|
|
partitions[partition_name] = AddCustomImages(output_zip, partition_name, image_list)
|
|
|
|
if OPTIONS.info_dict.get("avb_enable") == "true":
|
|
# vbmeta_partitions includes the partitions that should be included into
|
|
# top-level vbmeta.img, which are the ones that are not included in any
|
|
# chained VBMeta image plus the chained VBMeta images themselves.
|
|
# Currently avb_custom_partitions are all chained to VBMeta image.
|
|
vbmeta_partitions = common.AVB_PARTITIONS[:] + tuple(avb_custom_partitions)
|
|
|
|
vbmeta_system = OPTIONS.info_dict.get("avb_vbmeta_system", "").strip()
|
|
if vbmeta_system:
|
|
banner("vbmeta_system")
|
|
partitions["vbmeta_system"] = AddVBMeta(
|
|
output_zip, partitions, "vbmeta_system", vbmeta_system.split())
|
|
vbmeta_partitions = [
|
|
item for item in vbmeta_partitions
|
|
if item not in vbmeta_system.split()]
|
|
vbmeta_partitions.append("vbmeta_system")
|
|
|
|
vbmeta_vendor = OPTIONS.info_dict.get("avb_vbmeta_vendor", "").strip()
|
|
if vbmeta_vendor:
|
|
banner("vbmeta_vendor")
|
|
partitions["vbmeta_vendor"] = AddVBMeta(
|
|
output_zip, partitions, "vbmeta_vendor", vbmeta_vendor.split())
|
|
vbmeta_partitions = [
|
|
item for item in vbmeta_partitions
|
|
if item not in vbmeta_vendor.split()]
|
|
vbmeta_partitions.append("vbmeta_vendor")
|
|
custom_avb_partitions = OPTIONS.info_dict.get(
|
|
"avb_custom_vbmeta_images_partition_list", "").strip().split()
|
|
if custom_avb_partitions:
|
|
for avb_part in custom_avb_partitions:
|
|
partition_name = "vbmeta_" + avb_part
|
|
included_partitions = OPTIONS.info_dict.get(
|
|
"avb_vbmeta_{}".format(avb_part), "").strip().split()
|
|
assert included_partitions, "Custom vbmeta partition {0} missing avb_vbmeta_{0} prop".format(
|
|
avb_part)
|
|
banner(partition_name)
|
|
logger.info("VBMeta partition {} needs {}".format(
|
|
partition_name, included_partitions))
|
|
partitions[partition_name] = AddVBMeta(
|
|
output_zip, partitions, partition_name, included_partitions)
|
|
vbmeta_partitions = [
|
|
item for item in vbmeta_partitions
|
|
if item not in included_partitions]
|
|
vbmeta_partitions.append(partition_name)
|
|
|
|
if OPTIONS.info_dict.get("avb_building_vbmeta_image") == "true":
|
|
banner("vbmeta")
|
|
AddVBMeta(output_zip, partitions, "vbmeta", vbmeta_partitions)
|
|
|
|
if OPTIONS.info_dict.get("use_dynamic_partitions") == "true":
|
|
if OPTIONS.info_dict.get("build_super_empty_partition") == "true":
|
|
banner("super_empty")
|
|
AddSuperEmpty(output_zip)
|
|
|
|
if OPTIONS.info_dict.get("build_super_partition") == "true":
|
|
if OPTIONS.info_dict.get(
|
|
"build_retrofit_dynamic_partitions_ota_package") == "true":
|
|
banner("super split images")
|
|
AddSuperSplit(output_zip)
|
|
|
|
banner("radio")
|
|
ab_partitions_txt = os.path.join(OPTIONS.input_tmp, "META",
|
|
"ab_partitions.txt")
|
|
if os.path.exists(ab_partitions_txt):
|
|
with open(ab_partitions_txt) as f:
|
|
ab_partitions = f.read().splitlines()
|
|
|
|
# For devices using A/B update, make sure we have all the needed images
|
|
# ready under IMAGES/ or RADIO/.
|
|
CheckAbOtaImages(output_zip, ab_partitions)
|
|
|
|
# Generate care_map.pb for ab_partitions, then write this file to
|
|
# target_files package.
|
|
output_care_map = os.path.join(OPTIONS.input_tmp, "META", "care_map.pb")
|
|
AddCareMapForAbOta(output_zip if output_zip else output_care_map,
|
|
ab_partitions, partitions)
|
|
|
|
# Radio images that need to be packed into IMAGES/, and product-img.zip.
|
|
pack_radioimages_txt = os.path.join(
|
|
OPTIONS.input_tmp, "META", "pack_radioimages.txt")
|
|
if os.path.exists(pack_radioimages_txt):
|
|
with open(pack_radioimages_txt) as f:
|
|
AddPackRadioImages(output_zip, f.readlines())
|
|
|
|
AddVbmetaDigest(output_zip)
|
|
|
|
if output_zip:
|
|
common.ZipClose(output_zip)
|
|
if OPTIONS.replace_updated_files_list:
|
|
ReplaceUpdatedFiles(output_zip.filename,
|
|
OPTIONS.replace_updated_files_list)
|
|
|
|
|
|
def OptimizeCompressedEntries(zipfile_path):
|
|
"""Convert files that do not compress well to uncompressed storage
|
|
|
|
EROFS images tend to be compressed already, so compressing them again
|
|
yields little space savings. Leaving them uncompressed will make
|
|
downstream tooling's job easier, and save compute time.
|
|
"""
|
|
if not zipfile.is_zipfile(zipfile_path):
|
|
return
|
|
entries_to_store = []
|
|
with tempfile.TemporaryDirectory() as tmpdir:
|
|
with zipfile.ZipFile(zipfile_path, "r", allowZip64=True) as zfp:
|
|
for zinfo in zfp.filelist:
|
|
if not zinfo.filename.startswith("IMAGES/") and not zinfo.filename.startswith("META"):
|
|
continue
|
|
# Don't try to store userdata.img uncompressed, it's usually huge.
|
|
if zinfo.filename.endswith("userdata.img"):
|
|
continue
|
|
if zinfo.compress_size > zinfo.file_size * 0.80 and zinfo.compress_type != zipfile.ZIP_STORED:
|
|
entries_to_store.append(zinfo)
|
|
zfp.extract(zinfo, tmpdir)
|
|
if len(entries_to_store) == 0:
|
|
return
|
|
# Remove these entries, then re-add them as ZIP_STORED
|
|
ZipDelete(zipfile_path, [entry.filename for entry in entries_to_store])
|
|
with zipfile.ZipFile(zipfile_path, "a", allowZip64=True) as zfp:
|
|
for entry in entries_to_store:
|
|
zfp.write(os.path.join(tmpdir, entry.filename),
|
|
entry.filename, compress_type=zipfile.ZIP_STORED)
|
|
|
|
|
|
def main(argv):
|
|
def option_handler(o, a):
|
|
if o in ("-a", "--add_missing"):
|
|
OPTIONS.add_missing = True
|
|
elif o in ("-r", "--rebuild_recovery",):
|
|
OPTIONS.rebuild_recovery = True
|
|
elif o == "--replace_verity_private_key":
|
|
raise ValueError("--replace_verity_private_key is no longer supported,"
|
|
" please switch to AVB")
|
|
elif o == "--replace_verity_public_key":
|
|
raise ValueError("--replace_verity_public_key is no longer supported,"
|
|
" please switch to AVB")
|
|
elif o == "--is_signing":
|
|
OPTIONS.is_signing = True
|
|
elif o == "--avb_resolve_rollback_index_location_conflict":
|
|
OPTIONS.avb_resolve_rollback_index_location_conflict = True
|
|
else:
|
|
return False
|
|
return True
|
|
|
|
args = common.ParseOptions(
|
|
argv, __doc__, extra_opts="ar",
|
|
extra_long_opts=["add_missing", "rebuild_recovery",
|
|
"replace_verity_public_key=",
|
|
"replace_verity_private_key=",
|
|
"is_signing",
|
|
"avb_resolve_rollback_index_location_conflict"],
|
|
extra_option_handler=option_handler)
|
|
|
|
if len(args) != 1:
|
|
common.Usage(__doc__)
|
|
sys.exit(1)
|
|
|
|
common.InitLogging()
|
|
|
|
AddImagesToTargetFiles(args[0])
|
|
OptimizeCompressedEntries(args[0])
|
|
logger.info("done.")
|
|
|
|
|
|
if __name__ == '__main__':
|
|
try:
|
|
common.CloseInheritedPipes()
|
|
main(sys.argv[1:])
|
|
finally:
|
|
common.Cleanup()
|