4978fa99d1
When using Verified Boot 2.0, releasetools specifies a salt value based on build fingerprint, so that to give idempotent images. However, the change that removed static `ro.build.fingerprint` [1] broke the behavior, as common.LoadInfoDict still relies on fingerprints. Without a fixed salt, the first call to make_recovery_patch.py and the second one (which writes IMAGES/{boot,recovery}.img) will see different images, which leads to install-recovery.sh failure. Note that currently there's a dependency that requires getting bootable images through two separate calls. make_recovery_patch.py has to happen first to get (placeholder) files in the system image. We then generate canned fs_config files, and finally use add_img_to_target_files.py to write the images. This CL adds a quick workaround to force rebuilding the recovery-from-boot patch while calling add_img_to_target_files.py. [1] https://android-review.googlesource.com/c/platform/build/+/892933 Bug: 134123803 Bug: 134525174 Test: TreeHugger Test: Build a non-A/B target that uses AVB. Run validate_target_files.py on the generated target_files.zip. Change-Id: I5859e30be63bfd54398cf41fd2d907f15285f560
970 lines
34 KiB
Python
Executable file
970 lines
34 KiB
Python
Executable file
#!/usr/bin/env python
|
|
#
|
|
# Copyright (C) 2014 The Android Open Source Project
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
"""
|
|
Given a target-files zipfile that does not contain images (ie, does
|
|
not have an IMAGES/ top-level subdirectory), produce the images and
|
|
add them to the zipfile.
|
|
|
|
Usage: add_img_to_target_files [flag] target_files
|
|
|
|
-a (--add_missing)
|
|
Build and add missing images to "IMAGES/". If this option is
|
|
not specified, this script will simply exit when "IMAGES/"
|
|
directory exists in the target file.
|
|
|
|
-r (--rebuild_recovery)
|
|
Rebuild the recovery patch and write it to the system image. Only
|
|
meaningful when system image needs to be rebuilt and there're separate
|
|
boot / recovery images.
|
|
|
|
--replace_verity_private_key
|
|
Replace the private key used for verity signing. (same as the option
|
|
in sign_target_files_apks)
|
|
|
|
--replace_verity_public_key
|
|
Replace the certificate (public key) used for verity verification. (same
|
|
as the option in sign_target_files_apks)
|
|
|
|
--is_signing
|
|
Skip building & adding the images for "userdata" and "cache" if we
|
|
are signing the target files.
|
|
"""
|
|
|
|
from __future__ import print_function
|
|
|
|
import datetime
|
|
import logging
|
|
import os
|
|
import shlex
|
|
import shutil
|
|
import sys
|
|
import uuid
|
|
import zipfile
|
|
|
|
import build_image
|
|
import build_super_image
|
|
import common
|
|
import rangelib
|
|
import sparse_img
|
|
|
|
if sys.hexversion < 0x02070000:
|
|
print("Python 2.7 or newer is required.", file=sys.stderr)
|
|
sys.exit(1)
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
OPTIONS = common.OPTIONS
|
|
OPTIONS.add_missing = False
|
|
OPTIONS.rebuild_recovery = False
|
|
OPTIONS.replace_updated_files_list = []
|
|
OPTIONS.replace_verity_public_key = False
|
|
OPTIONS.replace_verity_private_key = False
|
|
OPTIONS.is_signing = False
|
|
|
|
# Use a fixed timestamp (01/01/2009 00:00:00 UTC) for files when packaging
|
|
# images. (b/24377993, b/80600931)
|
|
FIXED_FILE_TIMESTAMP = int((
|
|
datetime.datetime(2009, 1, 1, 0, 0, 0, 0, None) -
|
|
datetime.datetime.utcfromtimestamp(0)).total_seconds())
|
|
|
|
|
|
class OutputFile(object):
|
|
"""A helper class to write a generated file to the given dir or zip.
|
|
|
|
When generating images, we want the outputs to go into the given zip file, or
|
|
the given dir.
|
|
|
|
Attributes:
|
|
name: The name of the output file, regardless of the final destination.
|
|
"""
|
|
|
|
def __init__(self, output_zip, input_dir, prefix, name):
|
|
# We write the intermediate output file under the given input_dir, even if
|
|
# the final destination is a zip archive.
|
|
self.name = os.path.join(input_dir, prefix, name)
|
|
self._output_zip = output_zip
|
|
if self._output_zip:
|
|
self._zip_name = os.path.join(prefix, name)
|
|
|
|
def Write(self):
|
|
if self._output_zip:
|
|
common.ZipWrite(self._output_zip, self.name, self._zip_name)
|
|
|
|
|
|
def GetCareMap(which, imgname):
|
|
"""Returns the care_map string for the given partition.
|
|
|
|
Args:
|
|
which: The partition name, must be listed in PARTITIONS_WITH_CARE_MAP.
|
|
imgname: The filename of the image.
|
|
|
|
Returns:
|
|
(which, care_map_ranges): care_map_ranges is the raw string of the care_map
|
|
RangeSet; or None.
|
|
"""
|
|
assert which in common.PARTITIONS_WITH_CARE_MAP
|
|
|
|
# which + "_image_size" contains the size that the actual filesystem image
|
|
# resides in, which is all that needs to be verified. The additional blocks in
|
|
# the image file contain verity metadata, by reading which would trigger
|
|
# invalid reads.
|
|
image_size = OPTIONS.info_dict.get(which + "_image_size")
|
|
if not image_size:
|
|
return None
|
|
|
|
image_blocks = int(image_size) / 4096 - 1
|
|
assert image_blocks > 0, "blocks for {} must be positive".format(which)
|
|
|
|
# For sparse images, we will only check the blocks that are listed in the care
|
|
# map, i.e. the ones with meaningful data.
|
|
if "extfs_sparse_flag" in OPTIONS.info_dict:
|
|
simg = sparse_img.SparseImage(imgname)
|
|
care_map_ranges = simg.care_map.intersect(
|
|
rangelib.RangeSet("0-{}".format(image_blocks)))
|
|
|
|
# Otherwise for non-sparse images, we read all the blocks in the filesystem
|
|
# image.
|
|
else:
|
|
care_map_ranges = rangelib.RangeSet("0-{}".format(image_blocks))
|
|
|
|
return [which, care_map_ranges.to_string_raw()]
|
|
|
|
|
|
def AddSystem(output_zip, recovery_img=None, boot_img=None):
|
|
"""Turn the contents of SYSTEM into a system image and store it in
|
|
output_zip. Returns the name of the system image file."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "system.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("system.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
def output_sink(fn, data):
|
|
ofile = open(os.path.join(OPTIONS.input_tmp, "SYSTEM", fn), "w")
|
|
ofile.write(data)
|
|
ofile.close()
|
|
|
|
if output_zip:
|
|
arc_name = "SYSTEM/" + fn
|
|
if arc_name in output_zip.namelist():
|
|
OPTIONS.replace_updated_files_list.append(arc_name)
|
|
else:
|
|
common.ZipWrite(output_zip, ofile.name, arc_name)
|
|
|
|
if (OPTIONS.rebuild_recovery and recovery_img is not None and
|
|
boot_img is not None):
|
|
logger.info("Building new recovery patch")
|
|
common.MakeRecoveryPatch(OPTIONS.input_tmp, output_sink, recovery_img,
|
|
boot_img, info_dict=OPTIONS.info_dict)
|
|
|
|
block_list = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "system.map")
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "system", img,
|
|
block_list=block_list)
|
|
|
|
return img.name
|
|
|
|
|
|
def AddSystemOther(output_zip):
|
|
"""Turn the contents of SYSTEM_OTHER into a system_other image
|
|
and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "system_other.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("system_other.img already exists; no need to rebuild...")
|
|
return
|
|
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "system_other", img)
|
|
|
|
|
|
def AddVendor(output_zip):
|
|
"""Turn the contents of VENDOR into a vendor image and store in it
|
|
output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "vendor.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("vendor.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "vendor.map")
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "vendor", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddProduct(output_zip):
|
|
"""Turn the contents of PRODUCT into a product image and store it in
|
|
output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "product.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("product.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "product.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "product", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddProductServices(output_zip):
|
|
"""Turn the contents of PRODUCT_SERVICES into a product_services image and
|
|
store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES",
|
|
"product_services.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("product_services.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "product_services.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "product_services", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddOdm(output_zip):
|
|
"""Turn the contents of ODM into an odm image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "odm.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("odm.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
block_list = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "odm.map")
|
|
CreateImage(
|
|
OPTIONS.input_tmp, OPTIONS.info_dict, "odm", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddDtbo(output_zip):
|
|
"""Adds the DTBO image.
|
|
|
|
Uses the image under IMAGES/ if it already exists. Otherwise looks for the
|
|
image under PREBUILT_IMAGES/, signs it as needed, and returns the image name.
|
|
"""
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "dtbo.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("dtbo.img already exists; no need to rebuild...")
|
|
return img.name
|
|
|
|
dtbo_prebuilt_path = os.path.join(
|
|
OPTIONS.input_tmp, "PREBUILT_IMAGES", "dtbo.img")
|
|
assert os.path.exists(dtbo_prebuilt_path)
|
|
shutil.copy(dtbo_prebuilt_path, img.name)
|
|
|
|
# AVB-sign the image as needed.
|
|
if OPTIONS.info_dict.get("avb_enable") == "true":
|
|
avbtool = OPTIONS.info_dict["avb_avbtool"]
|
|
part_size = OPTIONS.info_dict["dtbo_size"]
|
|
# The AVB hash footer will be replaced if already present.
|
|
cmd = [avbtool, "add_hash_footer", "--image", img.name,
|
|
"--partition_size", str(part_size), "--partition_name", "dtbo"]
|
|
common.AppendAVBSigningArgs(cmd, "dtbo")
|
|
args = OPTIONS.info_dict.get("avb_dtbo_add_hash_footer_args")
|
|
if args and args.strip():
|
|
cmd.extend(shlex.split(args))
|
|
common.RunAndCheckOutput(cmd)
|
|
|
|
img.Write()
|
|
return img.name
|
|
|
|
|
|
def CreateImage(input_dir, info_dict, what, output_file, block_list=None):
|
|
logger.info("creating " + what + ".img...")
|
|
|
|
image_props = build_image.ImagePropFromGlobalDict(info_dict, what)
|
|
fstab = info_dict["fstab"]
|
|
mount_point = "/" + what
|
|
if fstab and mount_point in fstab:
|
|
image_props["fs_type"] = fstab[mount_point].fs_type
|
|
|
|
image_props["timestamp"] = FIXED_FILE_TIMESTAMP
|
|
|
|
if what == "system":
|
|
fs_config_prefix = ""
|
|
else:
|
|
fs_config_prefix = what + "_"
|
|
|
|
fs_config = os.path.join(
|
|
input_dir, "META/" + fs_config_prefix + "filesystem_config.txt")
|
|
if not os.path.exists(fs_config):
|
|
fs_config = None
|
|
|
|
# Override values loaded from info_dict.
|
|
if fs_config:
|
|
image_props["fs_config"] = fs_config
|
|
if block_list:
|
|
image_props["block_list"] = block_list.name
|
|
|
|
# Use repeatable ext4 FS UUID and hash_seed UUID (based on partition name and
|
|
# build fingerprint).
|
|
uuid_seed = what + "-"
|
|
if "build.prop" in info_dict:
|
|
build_prop = info_dict["build.prop"]
|
|
if "ro.build.fingerprint" in build_prop:
|
|
uuid_seed += build_prop["ro.build.fingerprint"]
|
|
elif "ro.build.thumbprint" in build_prop:
|
|
uuid_seed += build_prop["ro.build.thumbprint"]
|
|
image_props["uuid"] = str(uuid.uuid5(uuid.NAMESPACE_URL, uuid_seed))
|
|
hash_seed = "hash_seed-" + uuid_seed
|
|
image_props["hash_seed"] = str(uuid.uuid5(uuid.NAMESPACE_URL, hash_seed))
|
|
|
|
build_image.BuildImage(
|
|
os.path.join(input_dir, what.upper()), image_props, output_file.name)
|
|
|
|
output_file.Write()
|
|
if block_list:
|
|
block_list.Write()
|
|
|
|
# Set the '_image_size' for given image size.
|
|
is_verity_partition = "verity_block_device" in image_props
|
|
verity_supported = (image_props.get("verity") == "true" or
|
|
image_props.get("avb_enable") == "true")
|
|
is_avb_enable = image_props.get("avb_hashtree_enable") == "true"
|
|
if verity_supported and (is_verity_partition or is_avb_enable):
|
|
image_size = image_props.get("image_size")
|
|
if image_size:
|
|
image_size_key = what + "_image_size"
|
|
info_dict[image_size_key] = int(image_size)
|
|
|
|
use_dynamic_size = (
|
|
info_dict.get("use_dynamic_partition_size") == "true" and
|
|
what in shlex.split(info_dict.get("dynamic_partition_list", "").strip()))
|
|
if use_dynamic_size:
|
|
info_dict.update(build_image.GlobalDictFromImageProp(image_props, what))
|
|
|
|
|
|
def AddUserdata(output_zip):
|
|
"""Create a userdata image and store it in output_zip.
|
|
|
|
In most case we just create and store an empty userdata.img;
|
|
But the invoker can also request to create userdata.img with real
|
|
data from the target files, by setting "userdata_img_with_data=true"
|
|
in OPTIONS.info_dict.
|
|
"""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "userdata.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("userdata.img already exists; no need to rebuild...")
|
|
return
|
|
|
|
# Skip userdata.img if no size.
|
|
image_props = build_image.ImagePropFromGlobalDict(OPTIONS.info_dict, "data")
|
|
if not image_props.get("partition_size"):
|
|
return
|
|
|
|
logger.info("creating userdata.img...")
|
|
|
|
image_props["timestamp"] = FIXED_FILE_TIMESTAMP
|
|
|
|
if OPTIONS.info_dict.get("userdata_img_with_data") == "true":
|
|
user_dir = os.path.join(OPTIONS.input_tmp, "DATA")
|
|
else:
|
|
user_dir = common.MakeTempDir()
|
|
|
|
fstab = OPTIONS.info_dict["fstab"]
|
|
if fstab:
|
|
image_props["fs_type"] = fstab["/data"].fs_type
|
|
build_image.BuildImage(user_dir, image_props, img.name)
|
|
|
|
common.CheckSize(img.name, "userdata.img", OPTIONS.info_dict)
|
|
img.Write()
|
|
|
|
|
|
def AppendVBMetaArgsForPartition(cmd, partition, image):
|
|
"""Appends the VBMeta arguments for partition.
|
|
|
|
It sets up the VBMeta argument by including the partition descriptor from the
|
|
given 'image', or by configuring the partition as a chained partition.
|
|
|
|
Args:
|
|
cmd: A list of command args that will be used to generate the vbmeta image.
|
|
The argument for the partition will be appended to the list.
|
|
partition: The name of the partition (e.g. "system").
|
|
image: The path to the partition image.
|
|
"""
|
|
# Check if chain partition is used.
|
|
key_path = OPTIONS.info_dict.get("avb_" + partition + "_key_path")
|
|
if key_path:
|
|
chained_partition_arg = common.GetAvbChainedPartitionArg(
|
|
partition, OPTIONS.info_dict)
|
|
cmd.extend(["--chain_partition", chained_partition_arg])
|
|
else:
|
|
cmd.extend(["--include_descriptors_from_image", image])
|
|
|
|
|
|
def AddVBMeta(output_zip, partitions, name, needed_partitions):
|
|
"""Creates a VBMeta image and stores it in output_zip.
|
|
|
|
It generates the requested VBMeta image. The requested image could be for
|
|
top-level or chained VBMeta image, which is determined based on the name.
|
|
|
|
Args:
|
|
output_zip: The output zip file, which needs to be already open.
|
|
partitions: A dict that's keyed by partition names with image paths as
|
|
values. Only valid partition names are accepted, as listed in
|
|
common.AVB_PARTITIONS.
|
|
name: Name of the VBMeta partition, e.g. 'vbmeta', 'vbmeta_system'.
|
|
needed_partitions: Partitions whose descriptors should be included into the
|
|
generated VBMeta image.
|
|
|
|
Returns:
|
|
Path to the created image.
|
|
|
|
Raises:
|
|
AssertionError: On invalid input args.
|
|
"""
|
|
assert needed_partitions, "Needed partitions must be specified"
|
|
|
|
img = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "{}.img".format(name))
|
|
if os.path.exists(img.name):
|
|
logger.info("%s.img already exists; not rebuilding...", name)
|
|
return img.name
|
|
|
|
avbtool = OPTIONS.info_dict["avb_avbtool"]
|
|
cmd = [avbtool, "make_vbmeta_image", "--output", img.name]
|
|
common.AppendAVBSigningArgs(cmd, name)
|
|
|
|
for partition, path in partitions.items():
|
|
if partition not in needed_partitions:
|
|
continue
|
|
assert (partition in common.AVB_PARTITIONS or
|
|
partition.startswith('vbmeta_')), \
|
|
'Unknown partition: {}'.format(partition)
|
|
assert os.path.exists(path), \
|
|
'Failed to find {} for {}'.format(path, partition)
|
|
AppendVBMetaArgsForPartition(cmd, partition, path)
|
|
|
|
args = OPTIONS.info_dict.get("avb_{}_args".format(name))
|
|
if args and args.strip():
|
|
split_args = shlex.split(args)
|
|
for index, arg in enumerate(split_args[:-1]):
|
|
# Sanity check that the image file exists. Some images might be defined
|
|
# as a path relative to source tree, which may not be available at the
|
|
# same location when running this script (we have the input target_files
|
|
# zip only). For such cases, we additionally scan other locations (e.g.
|
|
# IMAGES/, RADIO/, etc) before bailing out.
|
|
if arg == '--include_descriptors_from_image':
|
|
image_path = split_args[index + 1]
|
|
if os.path.exists(image_path):
|
|
continue
|
|
found = False
|
|
for dir_name in ['IMAGES', 'RADIO', 'PREBUILT_IMAGES']:
|
|
alt_path = os.path.join(
|
|
OPTIONS.input_tmp, dir_name, os.path.basename(image_path))
|
|
if os.path.exists(alt_path):
|
|
split_args[index + 1] = alt_path
|
|
found = True
|
|
break
|
|
assert found, 'Failed to find {}'.format(image_path)
|
|
cmd.extend(split_args)
|
|
|
|
common.RunAndCheckOutput(cmd)
|
|
img.Write()
|
|
return img.name
|
|
|
|
|
|
def AddPartitionTable(output_zip):
|
|
"""Create a partition table image and store it in output_zip."""
|
|
|
|
img = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "IMAGES", "partition-table.img")
|
|
bpt = OutputFile(
|
|
output_zip, OPTIONS.input_tmp, "META", "partition-table.bpt")
|
|
|
|
# use BPTTOOL from environ, or "bpttool" if empty or not set.
|
|
bpttool = os.getenv("BPTTOOL") or "bpttool"
|
|
cmd = [bpttool, "make_table", "--output_json", bpt.name,
|
|
"--output_gpt", img.name]
|
|
input_files_str = OPTIONS.info_dict["board_bpt_input_files"]
|
|
input_files = input_files_str.split(" ")
|
|
for i in input_files:
|
|
cmd.extend(["--input", i])
|
|
disk_size = OPTIONS.info_dict.get("board_bpt_disk_size")
|
|
if disk_size:
|
|
cmd.extend(["--disk_size", disk_size])
|
|
args = OPTIONS.info_dict.get("board_bpt_make_table_args")
|
|
if args:
|
|
cmd.extend(shlex.split(args))
|
|
common.RunAndCheckOutput(cmd)
|
|
|
|
img.Write()
|
|
bpt.Write()
|
|
|
|
|
|
def AddCache(output_zip):
|
|
"""Create an empty cache image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "cache.img")
|
|
if os.path.exists(img.name):
|
|
logger.info("cache.img already exists; no need to rebuild...")
|
|
return
|
|
|
|
image_props = build_image.ImagePropFromGlobalDict(OPTIONS.info_dict, "cache")
|
|
# The build system has to explicitly request for cache.img.
|
|
if "fs_type" not in image_props:
|
|
return
|
|
|
|
logger.info("creating cache.img...")
|
|
|
|
image_props["timestamp"] = FIXED_FILE_TIMESTAMP
|
|
|
|
user_dir = common.MakeTempDir()
|
|
|
|
fstab = OPTIONS.info_dict["fstab"]
|
|
if fstab:
|
|
image_props["fs_type"] = fstab["/cache"].fs_type
|
|
build_image.BuildImage(user_dir, image_props, img.name)
|
|
|
|
common.CheckSize(img.name, "cache.img", OPTIONS.info_dict)
|
|
img.Write()
|
|
|
|
|
|
def CheckAbOtaImages(output_zip, ab_partitions):
|
|
"""Checks that all the listed A/B partitions have their images available.
|
|
|
|
The images need to be available under IMAGES/ or RADIO/, with the former takes
|
|
a priority.
|
|
|
|
Args:
|
|
output_zip: The output zip file (needs to be already open), or None to
|
|
find images in OPTIONS.input_tmp/.
|
|
ab_partitions: The list of A/B partitions.
|
|
|
|
Raises:
|
|
AssertionError: If it can't find an image.
|
|
"""
|
|
for partition in ab_partitions:
|
|
img_name = partition.strip() + ".img"
|
|
|
|
# Assert that the image is present under IMAGES/ now.
|
|
if output_zip:
|
|
# Zip spec says: All slashes MUST be forward slashes.
|
|
images_path = "IMAGES/" + img_name
|
|
radio_path = "RADIO/" + img_name
|
|
available = (images_path in output_zip.namelist() or
|
|
radio_path in output_zip.namelist())
|
|
else:
|
|
images_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name)
|
|
radio_path = os.path.join(OPTIONS.input_tmp, "RADIO", img_name)
|
|
available = os.path.exists(images_path) or os.path.exists(radio_path)
|
|
|
|
assert available, "Failed to find " + img_name
|
|
|
|
|
|
def AddCareMapForAbOta(output_zip, ab_partitions, image_paths):
|
|
"""Generates and adds care_map.pb for a/b partition that has care_map.
|
|
|
|
Args:
|
|
output_zip: The output zip file (needs to be already open), or None to
|
|
write care_map.pb to OPTIONS.input_tmp/.
|
|
ab_partitions: The list of A/B partitions.
|
|
image_paths: A map from the partition name to the image path.
|
|
"""
|
|
care_map_list = []
|
|
for partition in ab_partitions:
|
|
partition = partition.strip()
|
|
if partition not in common.PARTITIONS_WITH_CARE_MAP:
|
|
continue
|
|
|
|
verity_block_device = "{}_verity_block_device".format(partition)
|
|
avb_hashtree_enable = "avb_{}_hashtree_enable".format(partition)
|
|
if (verity_block_device in OPTIONS.info_dict or
|
|
OPTIONS.info_dict.get(avb_hashtree_enable) == "true"):
|
|
image_path = image_paths[partition]
|
|
assert os.path.exists(image_path)
|
|
|
|
care_map = GetCareMap(partition, image_path)
|
|
if not care_map:
|
|
continue
|
|
care_map_list += care_map
|
|
|
|
# adds fingerprint field to the care_map
|
|
build_props = OPTIONS.info_dict.get(partition + ".build.prop", {})
|
|
prop_name_list = ["ro.{}.build.fingerprint".format(partition),
|
|
"ro.{}.build.thumbprint".format(partition)]
|
|
|
|
present_props = [x for x in prop_name_list if x in build_props]
|
|
if not present_props:
|
|
logger.warning("fingerprint is not present for partition %s", partition)
|
|
property_id, fingerprint = "unknown", "unknown"
|
|
else:
|
|
property_id = present_props[0]
|
|
fingerprint = build_props[property_id]
|
|
care_map_list += [property_id, fingerprint]
|
|
|
|
if not care_map_list:
|
|
return
|
|
|
|
# Converts the list into proto buf message by calling care_map_generator; and
|
|
# writes the result to a temp file.
|
|
temp_care_map_text = common.MakeTempFile(prefix="caremap_text-",
|
|
suffix=".txt")
|
|
with open(temp_care_map_text, 'w') as text_file:
|
|
text_file.write('\n'.join(care_map_list))
|
|
|
|
temp_care_map = common.MakeTempFile(prefix="caremap-", suffix=".pb")
|
|
care_map_gen_cmd = ["care_map_generator", temp_care_map_text, temp_care_map]
|
|
common.RunAndCheckOutput(care_map_gen_cmd)
|
|
|
|
care_map_path = "META/care_map.pb"
|
|
if output_zip and care_map_path not in output_zip.namelist():
|
|
common.ZipWrite(output_zip, temp_care_map, arcname=care_map_path)
|
|
else:
|
|
shutil.copy(temp_care_map, os.path.join(OPTIONS.input_tmp, care_map_path))
|
|
if output_zip:
|
|
OPTIONS.replace_updated_files_list.append(care_map_path)
|
|
|
|
|
|
def AddPackRadioImages(output_zip, images):
|
|
"""Copies images listed in META/pack_radioimages.txt from RADIO/ to IMAGES/.
|
|
|
|
Args:
|
|
output_zip: The output zip file (needs to be already open), or None to
|
|
write images to OPTIONS.input_tmp/.
|
|
images: A list of image names.
|
|
|
|
Raises:
|
|
AssertionError: If a listed image can't be found.
|
|
"""
|
|
for image in images:
|
|
img_name = image.strip()
|
|
_, ext = os.path.splitext(img_name)
|
|
if not ext:
|
|
img_name += ".img"
|
|
|
|
prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name)
|
|
if os.path.exists(prebuilt_path):
|
|
logger.info("%s already exists, no need to overwrite...", img_name)
|
|
continue
|
|
|
|
img_radio_path = os.path.join(OPTIONS.input_tmp, "RADIO", img_name)
|
|
assert os.path.exists(img_radio_path), \
|
|
"Failed to find %s at %s" % (img_name, img_radio_path)
|
|
|
|
if output_zip:
|
|
common.ZipWrite(output_zip, img_radio_path, "IMAGES/" + img_name)
|
|
else:
|
|
shutil.copy(img_radio_path, prebuilt_path)
|
|
|
|
|
|
def AddSuperEmpty(output_zip):
|
|
"""Create a super_empty.img and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "IMAGES", "super_empty.img")
|
|
build_super_image.BuildSuperImage(OPTIONS.info_dict, img.name)
|
|
img.Write()
|
|
|
|
|
|
def AddSuperSplit(output_zip):
|
|
"""Create split super_*.img and store it in output_zip."""
|
|
|
|
outdir = os.path.join(OPTIONS.input_tmp, "OTA")
|
|
built = build_super_image.BuildSuperImage(OPTIONS.input_tmp, outdir)
|
|
|
|
if built:
|
|
for dev in OPTIONS.info_dict['super_block_devices'].strip().split():
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, "OTA",
|
|
"super_" + dev + ".img")
|
|
img.Write()
|
|
|
|
|
|
def ReplaceUpdatedFiles(zip_filename, files_list):
|
|
"""Updates all the ZIP entries listed in files_list.
|
|
|
|
For now the list includes META/care_map.pb, and the related files under
|
|
SYSTEM/ after rebuilding recovery.
|
|
"""
|
|
common.ZipDelete(zip_filename, files_list)
|
|
output_zip = zipfile.ZipFile(zip_filename, "a",
|
|
compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True)
|
|
for item in files_list:
|
|
file_path = os.path.join(OPTIONS.input_tmp, item)
|
|
assert os.path.exists(file_path)
|
|
common.ZipWrite(output_zip, file_path, arcname=item)
|
|
common.ZipClose(output_zip)
|
|
|
|
|
|
def AddImagesToTargetFiles(filename):
|
|
"""Creates and adds images (boot/recovery/system/...) to a target_files.zip.
|
|
|
|
It works with either a zip file (zip mode), or a directory that contains the
|
|
files to be packed into a target_files.zip (dir mode). The latter is used when
|
|
being called from build/make/core/Makefile.
|
|
|
|
The images will be created under IMAGES/ in the input target_files.zip.
|
|
|
|
Args:
|
|
filename: the target_files.zip, or the zip root directory.
|
|
"""
|
|
if os.path.isdir(filename):
|
|
OPTIONS.input_tmp = os.path.abspath(filename)
|
|
else:
|
|
OPTIONS.input_tmp = common.UnzipTemp(filename)
|
|
|
|
if not OPTIONS.add_missing:
|
|
if os.path.isdir(os.path.join(OPTIONS.input_tmp, "IMAGES")):
|
|
logger.warning("target_files appears to already contain images.")
|
|
sys.exit(1)
|
|
|
|
OPTIONS.info_dict = common.LoadInfoDict(OPTIONS.input_tmp, repacking=True)
|
|
|
|
has_recovery = OPTIONS.info_dict.get("no_recovery") != "true"
|
|
has_boot = OPTIONS.info_dict.get("no_boot") != "true"
|
|
|
|
# {vendor,odm,product,product_services}.img are unlike system.img or
|
|
# system_other.img. Because it could be built from source, or dropped into
|
|
# target_files.zip as a prebuilt blob. We consider either of them as
|
|
# {vendor,product,product_services}.img being available, which could be
|
|
# used when generating vbmeta.img for AVB.
|
|
has_vendor = (os.path.isdir(os.path.join(OPTIONS.input_tmp, "VENDOR")) or
|
|
os.path.exists(os.path.join(OPTIONS.input_tmp, "IMAGES",
|
|
"vendor.img")))
|
|
has_odm = (os.path.isdir(os.path.join(OPTIONS.input_tmp, "ODM")) or
|
|
os.path.exists(os.path.join(OPTIONS.input_tmp, "IMAGES",
|
|
"odm.img")))
|
|
has_product = (os.path.isdir(os.path.join(OPTIONS.input_tmp, "PRODUCT")) or
|
|
os.path.exists(os.path.join(OPTIONS.input_tmp, "IMAGES",
|
|
"product.img")))
|
|
has_product_services = (os.path.isdir(os.path.join(OPTIONS.input_tmp,
|
|
"PRODUCT_SERVICES")) or
|
|
os.path.exists(os.path.join(OPTIONS.input_tmp,
|
|
"IMAGES",
|
|
"product_services.img")))
|
|
has_system = os.path.isdir(os.path.join(OPTIONS.input_tmp, "SYSTEM"))
|
|
has_system_other = os.path.isdir(os.path.join(OPTIONS.input_tmp,
|
|
"SYSTEM_OTHER"))
|
|
|
|
# Set up the output destination. It writes to the given directory for dir
|
|
# mode; otherwise appends to the given ZIP.
|
|
if os.path.isdir(filename):
|
|
output_zip = None
|
|
else:
|
|
output_zip = zipfile.ZipFile(filename, "a",
|
|
compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True)
|
|
|
|
# Always make input_tmp/IMAGES available, since we may stage boot / recovery
|
|
# images there even under zip mode. The directory will be cleaned up as part
|
|
# of OPTIONS.input_tmp.
|
|
images_dir = os.path.join(OPTIONS.input_tmp, "IMAGES")
|
|
if not os.path.isdir(images_dir):
|
|
os.makedirs(images_dir)
|
|
|
|
# A map between partition names and their paths, which could be used when
|
|
# generating AVB vbmeta image.
|
|
partitions = dict()
|
|
|
|
def banner(s):
|
|
logger.info("\n\n++++ " + s + " ++++\n\n")
|
|
|
|
boot_image = None
|
|
if has_boot:
|
|
banner("boot")
|
|
# common.GetBootableImage() returns the image directly if present.
|
|
boot_image = common.GetBootableImage(
|
|
"IMAGES/boot.img", "boot.img", OPTIONS.input_tmp, "BOOT")
|
|
# boot.img may be unavailable in some targets (e.g. aosp_arm64).
|
|
if boot_image:
|
|
partitions['boot'] = os.path.join(OPTIONS.input_tmp, "IMAGES", "boot.img")
|
|
if not os.path.exists(partitions['boot']):
|
|
boot_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
boot_image.AddToZip(output_zip)
|
|
|
|
recovery_image = None
|
|
if has_recovery:
|
|
banner("recovery")
|
|
recovery_image = common.GetBootableImage(
|
|
"IMAGES/recovery.img", "recovery.img", OPTIONS.input_tmp, "RECOVERY")
|
|
assert recovery_image, "Failed to create recovery.img."
|
|
partitions['recovery'] = os.path.join(
|
|
OPTIONS.input_tmp, "IMAGES", "recovery.img")
|
|
if not os.path.exists(partitions['recovery']):
|
|
recovery_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
recovery_image.AddToZip(output_zip)
|
|
|
|
banner("recovery (two-step image)")
|
|
# The special recovery.img for two-step package use.
|
|
recovery_two_step_image = common.GetBootableImage(
|
|
"IMAGES/recovery-two-step.img", "recovery-two-step.img",
|
|
OPTIONS.input_tmp, "RECOVERY", two_step_image=True)
|
|
assert recovery_two_step_image, "Failed to create recovery-two-step.img."
|
|
recovery_two_step_image_path = os.path.join(
|
|
OPTIONS.input_tmp, "IMAGES", "recovery-two-step.img")
|
|
if not os.path.exists(recovery_two_step_image_path):
|
|
recovery_two_step_image.WriteToDir(OPTIONS.input_tmp)
|
|
if output_zip:
|
|
recovery_two_step_image.AddToZip(output_zip)
|
|
|
|
if has_system:
|
|
banner("system")
|
|
partitions['system'] = AddSystem(
|
|
output_zip, recovery_img=recovery_image, boot_img=boot_image)
|
|
|
|
if has_vendor:
|
|
banner("vendor")
|
|
partitions['vendor'] = AddVendor(output_zip)
|
|
|
|
if has_product:
|
|
banner("product")
|
|
partitions['product'] = AddProduct(output_zip)
|
|
|
|
if has_product_services:
|
|
banner("product_services")
|
|
partitions['product_services'] = AddProductServices(output_zip)
|
|
|
|
if has_odm:
|
|
banner("odm")
|
|
partitions['odm'] = AddOdm(output_zip)
|
|
|
|
if has_system_other:
|
|
banner("system_other")
|
|
AddSystemOther(output_zip)
|
|
|
|
if not OPTIONS.is_signing:
|
|
banner("userdata")
|
|
AddUserdata(output_zip)
|
|
banner("cache")
|
|
AddCache(output_zip)
|
|
|
|
if OPTIONS.info_dict.get("board_bpt_enable") == "true":
|
|
banner("partition-table")
|
|
AddPartitionTable(output_zip)
|
|
|
|
if OPTIONS.info_dict.get("has_dtbo") == "true":
|
|
banner("dtbo")
|
|
partitions['dtbo'] = AddDtbo(output_zip)
|
|
|
|
if OPTIONS.info_dict.get("avb_enable") == "true":
|
|
# vbmeta_partitions includes the partitions that should be included into
|
|
# top-level vbmeta.img, which are the ones that are not included in any
|
|
# chained VBMeta image plus the chained VBMeta images themselves.
|
|
vbmeta_partitions = common.AVB_PARTITIONS[:]
|
|
|
|
vbmeta_system = OPTIONS.info_dict.get("avb_vbmeta_system", "").strip()
|
|
if vbmeta_system:
|
|
banner("vbmeta_system")
|
|
partitions["vbmeta_system"] = AddVBMeta(
|
|
output_zip, partitions, "vbmeta_system", vbmeta_system.split())
|
|
vbmeta_partitions = [
|
|
item for item in vbmeta_partitions
|
|
if item not in vbmeta_system.split()]
|
|
vbmeta_partitions.append("vbmeta_system")
|
|
|
|
vbmeta_vendor = OPTIONS.info_dict.get("avb_vbmeta_vendor", "").strip()
|
|
if vbmeta_vendor:
|
|
banner("vbmeta_vendor")
|
|
partitions["vbmeta_vendor"] = AddVBMeta(
|
|
output_zip, partitions, "vbmeta_vendor", vbmeta_vendor.split())
|
|
vbmeta_partitions = [
|
|
item for item in vbmeta_partitions
|
|
if item not in vbmeta_vendor.split()]
|
|
vbmeta_partitions.append("vbmeta_vendor")
|
|
|
|
banner("vbmeta")
|
|
AddVBMeta(output_zip, partitions, "vbmeta", vbmeta_partitions)
|
|
|
|
if OPTIONS.info_dict.get("build_super_partition") == "true":
|
|
banner("super_empty")
|
|
AddSuperEmpty(output_zip)
|
|
|
|
if OPTIONS.info_dict.get(
|
|
"build_retrofit_dynamic_partitions_ota_package") == "true":
|
|
banner("super split images")
|
|
AddSuperSplit(output_zip)
|
|
|
|
banner("radio")
|
|
ab_partitions_txt = os.path.join(OPTIONS.input_tmp, "META",
|
|
"ab_partitions.txt")
|
|
if os.path.exists(ab_partitions_txt):
|
|
with open(ab_partitions_txt, 'r') as f:
|
|
ab_partitions = f.readlines()
|
|
|
|
# For devices using A/B update, make sure we have all the needed images
|
|
# ready under IMAGES/ or RADIO/.
|
|
CheckAbOtaImages(output_zip, ab_partitions)
|
|
|
|
# Generate care_map.pb for ab_partitions, then write this file to
|
|
# target_files package.
|
|
AddCareMapForAbOta(output_zip, ab_partitions, partitions)
|
|
|
|
# Radio images that need to be packed into IMAGES/, and product-img.zip.
|
|
pack_radioimages_txt = os.path.join(
|
|
OPTIONS.input_tmp, "META", "pack_radioimages.txt")
|
|
if os.path.exists(pack_radioimages_txt):
|
|
with open(pack_radioimages_txt, 'r') as f:
|
|
AddPackRadioImages(output_zip, f.readlines())
|
|
|
|
if output_zip:
|
|
common.ZipClose(output_zip)
|
|
if OPTIONS.replace_updated_files_list:
|
|
ReplaceUpdatedFiles(output_zip.filename,
|
|
OPTIONS.replace_updated_files_list)
|
|
|
|
|
|
def main(argv):
|
|
def option_handler(o, a):
|
|
if o in ("-a", "--add_missing"):
|
|
OPTIONS.add_missing = True
|
|
elif o in ("-r", "--rebuild_recovery",):
|
|
OPTIONS.rebuild_recovery = True
|
|
elif o == "--replace_verity_private_key":
|
|
OPTIONS.replace_verity_private_key = (True, a)
|
|
elif o == "--replace_verity_public_key":
|
|
OPTIONS.replace_verity_public_key = (True, a)
|
|
elif o == "--is_signing":
|
|
OPTIONS.is_signing = True
|
|
else:
|
|
return False
|
|
return True
|
|
|
|
args = common.ParseOptions(
|
|
argv, __doc__, extra_opts="ar",
|
|
extra_long_opts=["add_missing", "rebuild_recovery",
|
|
"replace_verity_public_key=",
|
|
"replace_verity_private_key=",
|
|
"is_signing"],
|
|
extra_option_handler=option_handler)
|
|
|
|
if len(args) != 1:
|
|
common.Usage(__doc__)
|
|
sys.exit(1)
|
|
|
|
common.InitLogging()
|
|
|
|
AddImagesToTargetFiles(args[0])
|
|
logger.info("done.")
|
|
|
|
if __name__ == '__main__':
|
|
try:
|
|
common.CloseInheritedPipes()
|
|
main(sys.argv[1:])
|
|
except common.ExternalError:
|
|
logger.exception("\n ERROR:\n")
|
|
sys.exit(1)
|
|
finally:
|
|
common.Cleanup()
|