2b6dfd6689
We want the generated images being identical for the same source files.
Currently the generated ext4 image (either from make_ext4fs or mke2fs)
is reproducible, but the AVB footer added by avbtool contain changes
because of the random salt being used.
This CL changes the avbtool invocation to specify "--salt <hexstring>"
(already supported by avbtool) to use reproducible salt that's computed
based on fingerprints (or thumbprints if applicable).
Bug: 67023482
Test: Regenerate images from the same source as follows:
Use a target_files.zip from an AVB-enabled target.
$ zip -d target_files.zip IMAGES/\*
$ ./build/make/tools/releasetools/add_img_to_target_files.py \
-v target_files.zip
Repeat the above commands and compare the generated images.
Change-Id: Id9db17ae0132ca3a820b4be5a5ef06ca3fef71ed
(cherry picked from commit 8f05cca1d9
)
767 lines
27 KiB
Python
Executable file
767 lines
27 KiB
Python
Executable file
#!/usr/bin/env python
|
|
#
|
|
# Copyright (C) 2014 The Android Open Source Project
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
"""
|
|
Given a target-files zipfile that does not contain images (ie, does
|
|
not have an IMAGES/ top-level subdirectory), produce the images and
|
|
add them to the zipfile.
|
|
|
|
Usage: add_img_to_target_files [flag] target_files
|
|
|
|
-a (--add_missing)
|
|
Build and add missing images to "IMAGES/". If this option is
|
|
not specified, this script will simply exit when "IMAGES/"
|
|
directory exists in the target file.
|
|
|
|
-r (--rebuild_recovery)
|
|
Rebuild the recovery patch and write it to the system image. Only
|
|
meaningful when system image needs to be rebuilt.
|
|
|
|
--replace_verity_private_key
|
|
Replace the private key used for verity signing. (same as the option
|
|
in sign_target_files_apks)
|
|
|
|
--replace_verity_public_key
|
|
Replace the certificate (public key) used for verity verification. (same
|
|
as the option in sign_target_files_apks)
|
|
|
|
--is_signing
|
|
Skip building & adding the images for "userdata" and "cache" if we
|
|
are signing the target files.
|
|
"""
|
|
|
|
from __future__ import print_function
|
|
|
|
import sys
|
|
|
|
if sys.hexversion < 0x02070000:
|
|
print("Python 2.7 or newer is required.", file=sys.stderr)
|
|
sys.exit(1)
|
|
|
|
import datetime
|
|
import errno
|
|
import hashlib
|
|
import os
|
|
import shlex
|
|
import shutil
|
|
import subprocess
|
|
import tempfile
|
|
import zipfile
|
|
|
|
import build_image
|
|
import common
|
|
import rangelib
|
|
import sparse_img
|
|
|
|
OPTIONS = common.OPTIONS
|
|
|
|
OPTIONS.add_missing = False
|
|
OPTIONS.rebuild_recovery = False
|
|
OPTIONS.replace_updated_files_list = []
|
|
OPTIONS.replace_verity_public_key = False
|
|
OPTIONS.replace_verity_private_key = False
|
|
OPTIONS.is_signing = False
|
|
|
|
|
|
class OutputFile(object):
|
|
def __init__(self, output_zip, input_dir, prefix, name):
|
|
self._output_zip = output_zip
|
|
self.input_name = os.path.join(input_dir, prefix, name)
|
|
|
|
if self._output_zip:
|
|
self._zip_name = os.path.join(prefix, name)
|
|
|
|
root, suffix = os.path.splitext(name)
|
|
self.name = common.MakeTempFile(prefix=root + '-', suffix=suffix)
|
|
else:
|
|
self.name = self.input_name
|
|
|
|
def Write(self):
|
|
if self._output_zip:
|
|
common.ZipWrite(self._output_zip, self.name, self._zip_name)
|
|
|
|
|
|
def GetCareMap(which, imgname):
|
|
"""Generate care_map of system (or vendor) partition"""
|
|
|
|
assert which in ("system", "vendor")
|
|
|
|
simg = sparse_img.SparseImage(imgname)
|
|
care_map_list = [which]
|
|
|
|
care_map_ranges = simg.care_map
|
|
key = which + "_adjusted_partition_size"
|
|
adjusted_blocks = OPTIONS.info_dict.get(key)
|
|
if adjusted_blocks:
|
|
assert adjusted_blocks > 0, "blocks should be positive for " + which
|
|
care_map_ranges = care_map_ranges.intersect(rangelib.RangeSet(
|
|
"0-%d" % (adjusted_blocks,)))
|
|
|
|
care_map_list.append(care_map_ranges.to_string_raw())
|
|
return care_map_list
|
|
|
|
|
|
def AddSystem(output_zip, prefix="IMAGES/", recovery_img=None, boot_img=None):
|
|
"""Turn the contents of SYSTEM into a system image and store it in
|
|
output_zip. Returns the name of the system image file."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "system.img")
|
|
if os.path.exists(img.input_name):
|
|
print("system.img already exists in %s, no need to rebuild..." % (prefix,))
|
|
return img.input_name
|
|
|
|
def output_sink(fn, data):
|
|
ofile = open(os.path.join(OPTIONS.input_tmp, "SYSTEM", fn), "w")
|
|
ofile.write(data)
|
|
ofile.close()
|
|
|
|
arc_name = "SYSTEM/" + fn
|
|
if arc_name in output_zip.namelist():
|
|
OPTIONS.replace_updated_files_list.append(arc_name)
|
|
else:
|
|
common.ZipWrite(output_zip, ofile.name, arc_name)
|
|
|
|
if OPTIONS.rebuild_recovery:
|
|
print("Building new recovery patch")
|
|
common.MakeRecoveryPatch(OPTIONS.input_tmp, output_sink, recovery_img,
|
|
boot_img, info_dict=OPTIONS.info_dict)
|
|
|
|
block_list = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "system.map")
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "system", img,
|
|
block_list=block_list)
|
|
|
|
return img.name
|
|
|
|
|
|
def AddSystemOther(output_zip, prefix="IMAGES/"):
|
|
"""Turn the contents of SYSTEM_OTHER into a system_other image
|
|
and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "system_other.img")
|
|
if os.path.exists(img.input_name):
|
|
print("system_other.img already exists in %s, no need to rebuild..." % (
|
|
prefix,))
|
|
return
|
|
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "system_other", img)
|
|
|
|
|
|
def AddVendor(output_zip, prefix="IMAGES/"):
|
|
"""Turn the contents of VENDOR into a vendor image and store in it
|
|
output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "vendor.img")
|
|
if os.path.exists(img.input_name):
|
|
print("vendor.img already exists in %s, no need to rebuild..." % (prefix,))
|
|
return img.input_name
|
|
|
|
block_list = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "vendor.map")
|
|
CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "vendor", img,
|
|
block_list=block_list)
|
|
return img.name
|
|
|
|
|
|
def AddDtbo(output_zip, prefix="IMAGES/"):
|
|
"""Adds the DTBO image.
|
|
|
|
Uses the image under prefix if it already exists. Otherwise looks for the
|
|
image under PREBUILT_IMAGES/, signs it as needed, and returns the image name.
|
|
"""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "dtbo.img")
|
|
if os.path.exists(img.input_name):
|
|
print("dtbo.img already exists in %s, no need to rebuild..." % (prefix,))
|
|
return img.input_name
|
|
|
|
dtbo_prebuilt_path = os.path.join(
|
|
OPTIONS.input_tmp, "PREBUILT_IMAGES", "dtbo.img")
|
|
assert os.path.exists(dtbo_prebuilt_path)
|
|
shutil.copy(dtbo_prebuilt_path, img.name)
|
|
|
|
# AVB-sign the image as needed.
|
|
if OPTIONS.info_dict.get("avb_enable") == "true":
|
|
avbtool = os.getenv('AVBTOOL') or OPTIONS.info_dict["avb_avbtool"]
|
|
part_size = OPTIONS.info_dict["dtbo_size"]
|
|
# The AVB hash footer will be replaced if already present.
|
|
cmd = [avbtool, "add_hash_footer", "--image", img.name,
|
|
"--partition_size", str(part_size), "--partition_name", "dtbo"]
|
|
common.AppendAVBSigningArgs(cmd, "dtbo")
|
|
args = OPTIONS.info_dict.get("avb_dtbo_add_hash_footer_args")
|
|
if args and args.strip():
|
|
cmd.extend(shlex.split(args))
|
|
p = common.Run(cmd, stdout=subprocess.PIPE)
|
|
p.communicate()
|
|
assert p.returncode == 0, \
|
|
"avbtool add_hash_footer of %s failed" % (img.name,)
|
|
|
|
img.Write()
|
|
return img.name
|
|
|
|
|
|
def CreateImage(input_dir, info_dict, what, output_file, block_list=None):
|
|
print("creating " + what + ".img...")
|
|
|
|
# The name of the directory it is making an image out of matters to
|
|
# mkyaffs2image. It wants "system" but we have a directory named
|
|
# "SYSTEM", so create a symlink.
|
|
temp_dir = tempfile.mkdtemp()
|
|
OPTIONS.tempfiles.append(temp_dir)
|
|
try:
|
|
os.symlink(os.path.join(input_dir, what.upper()),
|
|
os.path.join(temp_dir, what))
|
|
except OSError as e:
|
|
# bogus error on my mac version?
|
|
# File "./build/tools/releasetools/img_from_target_files"
|
|
# os.path.join(OPTIONS.input_tmp, "system"))
|
|
# OSError: [Errno 17] File exists
|
|
if e.errno == errno.EEXIST:
|
|
pass
|
|
|
|
image_props = build_image.ImagePropFromGlobalDict(info_dict, what)
|
|
fstab = info_dict["fstab"]
|
|
mount_point = "/" + what
|
|
if fstab and mount_point in fstab:
|
|
image_props["fs_type"] = fstab[mount_point].fs_type
|
|
|
|
# Use a fixed timestamp (01/01/2009) when packaging the image.
|
|
# Bug: 24377993
|
|
epoch = datetime.datetime.fromtimestamp(0)
|
|
timestamp = (datetime.datetime(2009, 1, 1) - epoch).total_seconds()
|
|
image_props["timestamp"] = int(timestamp)
|
|
|
|
if what == "system":
|
|
fs_config_prefix = ""
|
|
else:
|
|
fs_config_prefix = what + "_"
|
|
|
|
fs_config = os.path.join(
|
|
input_dir, "META/" + fs_config_prefix + "filesystem_config.txt")
|
|
if not os.path.exists(fs_config):
|
|
fs_config = None
|
|
|
|
# Override values loaded from info_dict.
|
|
if fs_config:
|
|
image_props["fs_config"] = fs_config
|
|
if block_list:
|
|
image_props["block_list"] = block_list.name
|
|
|
|
succ = build_image.BuildImage(os.path.join(temp_dir, what),
|
|
image_props, output_file.name)
|
|
assert succ, "build " + what + ".img image failed"
|
|
|
|
output_file.Write()
|
|
if block_list:
|
|
block_list.Write()
|
|
|
|
# Set the 'adjusted_partition_size' that excludes the verity blocks of the
|
|
# given image. When avb is enabled, this size is the max image size returned
|
|
# by the avb tool.
|
|
is_verity_partition = "verity_block_device" in image_props
|
|
verity_supported = (image_props.get("verity") == "true" or
|
|
image_props.get("avb_enable") == "true")
|
|
is_avb_enable = image_props.get("avb_hashtree_enable") == "true"
|
|
if verity_supported and (is_verity_partition or is_avb_enable):
|
|
adjusted_blocks_value = image_props.get("partition_size")
|
|
if adjusted_blocks_value:
|
|
adjusted_blocks_key = what + "_adjusted_partition_size"
|
|
info_dict[adjusted_blocks_key] = int(adjusted_blocks_value)/4096 - 1
|
|
|
|
|
|
def AddUserdata(output_zip, prefix="IMAGES/"):
|
|
"""Create a userdata image and store it in output_zip.
|
|
|
|
In most case we just create and store an empty userdata.img;
|
|
But the invoker can also request to create userdata.img with real
|
|
data from the target files, by setting "userdata_img_with_data=true"
|
|
in OPTIONS.info_dict.
|
|
"""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "userdata.img")
|
|
if os.path.exists(img.input_name):
|
|
print("userdata.img already exists in %s, no need to rebuild..." % (
|
|
prefix,))
|
|
return
|
|
|
|
# Skip userdata.img if no size.
|
|
image_props = build_image.ImagePropFromGlobalDict(OPTIONS.info_dict, "data")
|
|
if not image_props.get("partition_size"):
|
|
return
|
|
|
|
print("creating userdata.img...")
|
|
|
|
# Use a fixed timestamp (01/01/2009) when packaging the image.
|
|
# Bug: 24377993
|
|
epoch = datetime.datetime.fromtimestamp(0)
|
|
timestamp = (datetime.datetime(2009, 1, 1) - epoch).total_seconds()
|
|
image_props["timestamp"] = int(timestamp)
|
|
|
|
# The name of the directory it is making an image out of matters to
|
|
# mkyaffs2image. So we create a temp dir, and within it we create an
|
|
# empty dir named "data", or a symlink to the DATA dir,
|
|
# and build the image from that.
|
|
temp_dir = tempfile.mkdtemp()
|
|
OPTIONS.tempfiles.append(temp_dir)
|
|
user_dir = os.path.join(temp_dir, "data")
|
|
empty = (OPTIONS.info_dict.get("userdata_img_with_data") != "true")
|
|
if empty:
|
|
# Create an empty dir.
|
|
os.mkdir(user_dir)
|
|
else:
|
|
# Symlink to the DATA dir.
|
|
os.symlink(os.path.join(OPTIONS.input_tmp, "DATA"),
|
|
user_dir)
|
|
|
|
fstab = OPTIONS.info_dict["fstab"]
|
|
if fstab:
|
|
image_props["fs_type"] = fstab["/data"].fs_type
|
|
succ = build_image.BuildImage(user_dir, image_props, img.name)
|
|
assert succ, "build userdata.img image failed"
|
|
|
|
common.CheckSize(img.name, "userdata.img", OPTIONS.info_dict)
|
|
img.Write()
|
|
|
|
|
|
def AppendVBMetaArgsForPartition(cmd, partition, img_path, public_key_dir):
|
|
if not img_path:
|
|
return
|
|
|
|
# Check if chain partition is used.
|
|
key_path = OPTIONS.info_dict.get("avb_" + partition + "_key_path")
|
|
if key_path:
|
|
# extract public key in AVB format to be included in vbmeta.img
|
|
avbtool = os.getenv('AVBTOOL') or OPTIONS.info_dict["avb_avbtool"]
|
|
public_key_path = os.path.join(public_key_dir, "%s.avbpubkey" % partition)
|
|
p = common.Run([avbtool, "extract_public_key", "--key", key_path,
|
|
"--output", public_key_path],
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
p.communicate()
|
|
assert p.returncode == 0, \
|
|
"avbtool extract_public_key fail for partition: %r" % partition
|
|
|
|
rollback_index_location = OPTIONS.info_dict[
|
|
"avb_" + partition + "_rollback_index_location"]
|
|
cmd.extend(["--chain_partition", "%s:%s:%s" % (
|
|
partition, rollback_index_location, public_key_path)])
|
|
else:
|
|
cmd.extend(["--include_descriptors_from_image", img_path])
|
|
|
|
|
|
def AddVBMeta(output_zip, boot_img_path, system_img_path, vendor_img_path,
|
|
dtbo_img_path, prefix="IMAGES/"):
|
|
"""Create a VBMeta image and store it in output_zip."""
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "vbmeta.img")
|
|
avbtool = os.getenv('AVBTOOL') or OPTIONS.info_dict["avb_avbtool"]
|
|
cmd = [avbtool, "make_vbmeta_image", "--output", img.name]
|
|
common.AppendAVBSigningArgs(cmd, "vbmeta")
|
|
|
|
public_key_dir = tempfile.mkdtemp(prefix="avbpubkey-")
|
|
OPTIONS.tempfiles.append(public_key_dir)
|
|
|
|
AppendVBMetaArgsForPartition(cmd, "boot", boot_img_path, public_key_dir)
|
|
AppendVBMetaArgsForPartition(cmd, "system", system_img_path, public_key_dir)
|
|
AppendVBMetaArgsForPartition(cmd, "vendor", vendor_img_path, public_key_dir)
|
|
AppendVBMetaArgsForPartition(cmd, "dtbo", dtbo_img_path, public_key_dir)
|
|
|
|
args = OPTIONS.info_dict.get("avb_vbmeta_args")
|
|
if args and args.strip():
|
|
split_args = shlex.split(args)
|
|
for index, arg in enumerate(split_args[:-1]):
|
|
# Sanity check that the image file exists. Some images might be defined
|
|
# as a path relative to source tree, which may not be available at the
|
|
# same location when running this script (we have the input target_files
|
|
# zip only). For such cases, we additionally scan other locations (e.g.
|
|
# IMAGES/, RADIO/, etc) before bailing out.
|
|
if arg == '--include_descriptors_from_image':
|
|
image_path = split_args[index + 1]
|
|
if os.path.exists(image_path):
|
|
continue
|
|
found = False
|
|
for dir in ['IMAGES', 'RADIO', 'VENDOR_IMAGES', 'PREBUILT_IMAGES']:
|
|
alt_path = os.path.join(
|
|
OPTIONS.input_tmp, dir, os.path.basename(image_path))
|
|
if os.path.exists(alt_path):
|
|
split_args[index + 1] = alt_path
|
|
found = True
|
|
break
|
|
assert found, 'failed to find %s' % (image_path,)
|
|
cmd.extend(split_args)
|
|
|
|
p = common.Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
p.communicate()
|
|
assert p.returncode == 0, "avbtool make_vbmeta_image failed"
|
|
img.Write()
|
|
|
|
|
|
def AddPartitionTable(output_zip, prefix="IMAGES/"):
|
|
"""Create a partition table image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "partition-table.img")
|
|
bpt = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "partition-table.bpt")
|
|
|
|
# use BPTTOOL from environ, or "bpttool" if empty or not set.
|
|
bpttool = os.getenv("BPTTOOL") or "bpttool"
|
|
cmd = [bpttool, "make_table", "--output_json", bpt.name,
|
|
"--output_gpt", img.name]
|
|
input_files_str = OPTIONS.info_dict["board_bpt_input_files"]
|
|
input_files = input_files_str.split(" ")
|
|
for i in input_files:
|
|
cmd.extend(["--input", i])
|
|
disk_size = OPTIONS.info_dict.get("board_bpt_disk_size")
|
|
if disk_size:
|
|
cmd.extend(["--disk_size", disk_size])
|
|
args = OPTIONS.info_dict.get("board_bpt_make_table_args")
|
|
if args:
|
|
cmd.extend(shlex.split(args))
|
|
|
|
p = common.Run(cmd, stdout=subprocess.PIPE)
|
|
p.communicate()
|
|
assert p.returncode == 0, "bpttool make_table failed"
|
|
|
|
img.Write()
|
|
bpt.Write()
|
|
|
|
|
|
def AddCache(output_zip, prefix="IMAGES/"):
|
|
"""Create an empty cache image and store it in output_zip."""
|
|
|
|
img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "cache.img")
|
|
if os.path.exists(img.input_name):
|
|
print("cache.img already exists in %s, no need to rebuild..." % (prefix,))
|
|
return
|
|
|
|
image_props = build_image.ImagePropFromGlobalDict(OPTIONS.info_dict, "cache")
|
|
# The build system has to explicitly request for cache.img.
|
|
if "fs_type" not in image_props:
|
|
return
|
|
|
|
print("creating cache.img...")
|
|
|
|
# Use a fixed timestamp (01/01/2009) when packaging the image.
|
|
# Bug: 24377993
|
|
epoch = datetime.datetime.fromtimestamp(0)
|
|
timestamp = (datetime.datetime(2009, 1, 1) - epoch).total_seconds()
|
|
image_props["timestamp"] = int(timestamp)
|
|
|
|
# The name of the directory it is making an image out of matters to
|
|
# mkyaffs2image. So we create a temp dir, and within it we create an
|
|
# empty dir named "cache", and build the image from that.
|
|
temp_dir = tempfile.mkdtemp()
|
|
OPTIONS.tempfiles.append(temp_dir)
|
|
user_dir = os.path.join(temp_dir, "cache")
|
|
os.mkdir(user_dir)
|
|
|
|
fstab = OPTIONS.info_dict["fstab"]
|
|
if fstab:
|
|
image_props["fs_type"] = fstab["/cache"].fs_type
|
|
succ = build_image.BuildImage(user_dir, image_props, img.name)
|
|
assert succ, "build cache.img image failed"
|
|
|
|
common.CheckSize(img.name, "cache.img", OPTIONS.info_dict)
|
|
img.Write()
|
|
|
|
|
|
def ReplaceUpdatedFiles(zip_filename, files_list):
|
|
"""Update all the zip entries listed in the files_list.
|
|
|
|
For now the list includes META/care_map.txt, and the related files under
|
|
SYSTEM/ after rebuilding recovery.
|
|
"""
|
|
|
|
cmd = ["zip", "-d", zip_filename] + files_list
|
|
p = common.Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
|
p.communicate()
|
|
|
|
output_zip = zipfile.ZipFile(zip_filename, "a",
|
|
compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True)
|
|
for item in files_list:
|
|
file_path = os.path.join(OPTIONS.input_tmp, item)
|
|
assert os.path.exists(file_path)
|
|
common.ZipWrite(output_zip, file_path, arcname=item)
|
|
common.ZipClose(output_zip)
|
|
|
|
|
|
def AddImagesToTargetFiles(filename):
|
|
if os.path.isdir(filename):
|
|
OPTIONS.input_tmp = os.path.abspath(filename)
|
|
input_zip = None
|
|
else:
|
|
OPTIONS.input_tmp, input_zip = common.UnzipTemp(filename)
|
|
|
|
if not OPTIONS.add_missing:
|
|
if os.path.isdir(os.path.join(OPTIONS.input_tmp, "IMAGES")):
|
|
print("target_files appears to already contain images.")
|
|
sys.exit(1)
|
|
|
|
# vendor.img is unlike system.img or system_other.img. Because it could be
|
|
# built from source, or dropped into target_files.zip as a prebuilt blob. We
|
|
# consider either of them as vendor.img being available, which could be used
|
|
# when generating vbmeta.img for AVB.
|
|
has_vendor = (os.path.isdir(os.path.join(OPTIONS.input_tmp, "VENDOR")) or
|
|
os.path.exists(os.path.join(OPTIONS.input_tmp, "IMAGES",
|
|
"vendor.img")))
|
|
has_system_other = os.path.isdir(os.path.join(OPTIONS.input_tmp,
|
|
"SYSTEM_OTHER"))
|
|
|
|
if input_zip:
|
|
OPTIONS.info_dict = common.LoadInfoDict(input_zip, OPTIONS.input_tmp)
|
|
|
|
common.ZipClose(input_zip)
|
|
output_zip = zipfile.ZipFile(filename, "a",
|
|
compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True)
|
|
else:
|
|
OPTIONS.info_dict = common.LoadInfoDict(filename, filename)
|
|
output_zip = None
|
|
images_dir = os.path.join(OPTIONS.input_tmp, "IMAGES")
|
|
if not os.path.isdir(images_dir):
|
|
os.makedirs(images_dir)
|
|
images_dir = None
|
|
|
|
has_recovery = (OPTIONS.info_dict.get("no_recovery") != "true")
|
|
|
|
if OPTIONS.info_dict.get("avb_enable") == "true":
|
|
fp = None
|
|
if "build.prop" in OPTIONS.info_dict:
|
|
build_prop = OPTIONS.info_dict["build.prop"]
|
|
if "ro.build.fingerprint" in build_prop:
|
|
fp = build_prop["ro.build.fingerprint"]
|
|
elif "ro.build.thumbprint" in build_prop:
|
|
fp = build_prop["ro.build.thumbprint"]
|
|
if fp:
|
|
OPTIONS.info_dict["avb_salt"] = hashlib.sha256(fp).hexdigest()
|
|
|
|
def banner(s):
|
|
print("\n\n++++ " + s + " ++++\n\n")
|
|
|
|
prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", "boot.img")
|
|
boot_image = None
|
|
if os.path.exists(prebuilt_path):
|
|
banner("boot")
|
|
print("boot.img already exists in IMAGES/, no need to rebuild...")
|
|
if OPTIONS.rebuild_recovery:
|
|
boot_image = common.GetBootableImage(
|
|
"IMAGES/boot.img", "boot.img", OPTIONS.input_tmp, "BOOT")
|
|
else:
|
|
banner("boot")
|
|
boot_image = common.GetBootableImage(
|
|
"IMAGES/boot.img", "boot.img", OPTIONS.input_tmp, "BOOT")
|
|
if boot_image:
|
|
if output_zip:
|
|
boot_image.AddToZip(output_zip)
|
|
else:
|
|
boot_image.WriteToDir(OPTIONS.input_tmp)
|
|
|
|
recovery_image = None
|
|
if has_recovery:
|
|
banner("recovery")
|
|
prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", "recovery.img")
|
|
if os.path.exists(prebuilt_path):
|
|
print("recovery.img already exists in IMAGES/, no need to rebuild...")
|
|
if OPTIONS.rebuild_recovery:
|
|
recovery_image = common.GetBootableImage(
|
|
"IMAGES/recovery.img", "recovery.img", OPTIONS.input_tmp,
|
|
"RECOVERY")
|
|
else:
|
|
recovery_image = common.GetBootableImage(
|
|
"IMAGES/recovery.img", "recovery.img", OPTIONS.input_tmp, "RECOVERY")
|
|
if recovery_image:
|
|
if output_zip:
|
|
recovery_image.AddToZip(output_zip)
|
|
else:
|
|
recovery_image.WriteToDir(OPTIONS.input_tmp)
|
|
|
|
banner("recovery (two-step image)")
|
|
# The special recovery.img for two-step package use.
|
|
recovery_two_step_image = common.GetBootableImage(
|
|
"IMAGES/recovery-two-step.img", "recovery-two-step.img",
|
|
OPTIONS.input_tmp, "RECOVERY", two_step_image=True)
|
|
if recovery_two_step_image:
|
|
if output_zip:
|
|
recovery_two_step_image.AddToZip(output_zip)
|
|
else:
|
|
recovery_two_step_image.WriteToDir(OPTIONS.input_tmp)
|
|
|
|
banner("system")
|
|
system_img_path = AddSystem(
|
|
output_zip, recovery_img=recovery_image, boot_img=boot_image)
|
|
vendor_img_path = None
|
|
if has_vendor:
|
|
banner("vendor")
|
|
vendor_img_path = AddVendor(output_zip)
|
|
if has_system_other:
|
|
banner("system_other")
|
|
AddSystemOther(output_zip)
|
|
if not OPTIONS.is_signing:
|
|
banner("userdata")
|
|
AddUserdata(output_zip)
|
|
banner("cache")
|
|
AddCache(output_zip)
|
|
|
|
if OPTIONS.info_dict.get("board_bpt_enable") == "true":
|
|
banner("partition-table")
|
|
AddPartitionTable(output_zip)
|
|
|
|
dtbo_img_path = None
|
|
if OPTIONS.info_dict.get("has_dtbo") == "true":
|
|
banner("dtbo")
|
|
dtbo_img_path = AddDtbo(output_zip)
|
|
|
|
if OPTIONS.info_dict.get("avb_enable") == "true":
|
|
banner("vbmeta")
|
|
boot_contents = boot_image.WriteToTemp()
|
|
AddVBMeta(output_zip, boot_contents.name, system_img_path,
|
|
vendor_img_path, dtbo_img_path)
|
|
|
|
# For devices using A/B update, copy over images from RADIO/ and/or
|
|
# VENDOR_IMAGES/ to IMAGES/ and make sure we have all the needed
|
|
# images ready under IMAGES/. All images should have '.img' as extension.
|
|
banner("radio")
|
|
ab_partitions = os.path.join(OPTIONS.input_tmp, "META", "ab_partitions.txt")
|
|
if os.path.exists(ab_partitions):
|
|
with open(ab_partitions, 'r') as f:
|
|
lines = f.readlines()
|
|
# For devices using A/B update, generate care_map for system and vendor
|
|
# partitions (if present), then write this file to target_files package.
|
|
care_map_list = []
|
|
for line in lines:
|
|
if line.strip() == "system" and (
|
|
"system_verity_block_device" in OPTIONS.info_dict or
|
|
OPTIONS.info_dict.get("avb_system_hashtree_enable") == "true"):
|
|
assert os.path.exists(system_img_path)
|
|
care_map_list += GetCareMap("system", system_img_path)
|
|
if line.strip() == "vendor" and (
|
|
"vendor_verity_block_device" in OPTIONS.info_dict or
|
|
OPTIONS.info_dict.get("avb_vendor_hashtree_enable") == "true"):
|
|
assert os.path.exists(vendor_img_path)
|
|
care_map_list += GetCareMap("vendor", vendor_img_path)
|
|
|
|
img_name = line.strip() + ".img"
|
|
prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name)
|
|
if os.path.exists(prebuilt_path):
|
|
print("%s already exists, no need to overwrite..." % (img_name,))
|
|
continue
|
|
|
|
img_radio_path = os.path.join(OPTIONS.input_tmp, "RADIO", img_name)
|
|
img_vendor_dir = os.path.join(
|
|
OPTIONS.input_tmp, "VENDOR_IMAGES")
|
|
if os.path.exists(img_radio_path):
|
|
if output_zip:
|
|
common.ZipWrite(output_zip, img_radio_path,
|
|
os.path.join("IMAGES", img_name))
|
|
else:
|
|
shutil.copy(img_radio_path, prebuilt_path)
|
|
else:
|
|
for root, _, files in os.walk(img_vendor_dir):
|
|
if img_name in files:
|
|
if output_zip:
|
|
common.ZipWrite(output_zip, os.path.join(root, img_name),
|
|
os.path.join("IMAGES", img_name))
|
|
else:
|
|
shutil.copy(os.path.join(root, img_name), prebuilt_path)
|
|
break
|
|
|
|
if output_zip:
|
|
# Zip spec says: All slashes MUST be forward slashes.
|
|
img_path = 'IMAGES/' + img_name
|
|
assert img_path in output_zip.namelist(), "cannot find " + img_name
|
|
else:
|
|
img_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name)
|
|
assert os.path.exists(img_path), "cannot find " + img_name
|
|
|
|
if care_map_list:
|
|
care_map_path = "META/care_map.txt"
|
|
if output_zip and care_map_path not in output_zip.namelist():
|
|
common.ZipWriteStr(output_zip, care_map_path, '\n'.join(care_map_list))
|
|
else:
|
|
with open(os.path.join(OPTIONS.input_tmp, care_map_path), 'w') as fp:
|
|
fp.write('\n'.join(care_map_list))
|
|
if output_zip:
|
|
OPTIONS.replace_updated_files_list.append(care_map_path)
|
|
|
|
# Radio images that need to be packed into IMAGES/, and product-img.zip.
|
|
pack_radioimages = os.path.join(
|
|
OPTIONS.input_tmp, "META", "pack_radioimages.txt")
|
|
if os.path.exists(pack_radioimages):
|
|
with open(pack_radioimages, 'r') as f:
|
|
lines = f.readlines()
|
|
for line in lines:
|
|
img_name = line.strip()
|
|
_, ext = os.path.splitext(img_name)
|
|
if not ext:
|
|
img_name += ".img"
|
|
prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name)
|
|
if os.path.exists(prebuilt_path):
|
|
print("%s already exists, no need to overwrite..." % (img_name,))
|
|
continue
|
|
|
|
img_radio_path = os.path.join(OPTIONS.input_tmp, "RADIO", img_name)
|
|
assert os.path.exists(img_radio_path), \
|
|
"Failed to find %s at %s" % (img_name, img_radio_path)
|
|
if output_zip:
|
|
common.ZipWrite(output_zip, img_radio_path,
|
|
os.path.join("IMAGES", img_name))
|
|
else:
|
|
shutil.copy(img_radio_path, prebuilt_path)
|
|
|
|
if output_zip:
|
|
common.ZipClose(output_zip)
|
|
if OPTIONS.replace_updated_files_list:
|
|
ReplaceUpdatedFiles(output_zip.filename,
|
|
OPTIONS.replace_updated_files_list)
|
|
|
|
|
|
def main(argv):
|
|
def option_handler(o, a):
|
|
if o in ("-a", "--add_missing"):
|
|
OPTIONS.add_missing = True
|
|
elif o in ("-r", "--rebuild_recovery",):
|
|
OPTIONS.rebuild_recovery = True
|
|
elif o == "--replace_verity_private_key":
|
|
OPTIONS.replace_verity_private_key = (True, a)
|
|
elif o == "--replace_verity_public_key":
|
|
OPTIONS.replace_verity_public_key = (True, a)
|
|
elif o == "--is_signing":
|
|
OPTIONS.is_signing = True
|
|
else:
|
|
return False
|
|
return True
|
|
|
|
args = common.ParseOptions(
|
|
argv, __doc__, extra_opts="ar",
|
|
extra_long_opts=["add_missing", "rebuild_recovery",
|
|
"replace_verity_public_key=",
|
|
"replace_verity_private_key=",
|
|
"is_signing"],
|
|
extra_option_handler=option_handler)
|
|
|
|
|
|
if len(args) != 1:
|
|
common.Usage(__doc__)
|
|
sys.exit(1)
|
|
|
|
AddImagesToTargetFiles(args[0])
|
|
print("done.")
|
|
|
|
if __name__ == '__main__':
|
|
try:
|
|
common.CloseInheritedPipes()
|
|
main(sys.argv[1:])
|
|
except common.ExternalError as e:
|
|
print("\n ERROR: %s\n" % (e,))
|
|
sys.exit(1)
|
|
finally:
|
|
common.Cleanup()
|