9fd5886e23
If an OTA contains compressed APEX inside it, then the device will need to allocate space on /data partition for their decompression. In order to calculate how much space the OTA process needs to allocate, the process needs more information about the APEX contained inside the OTA. In this CL, we are adding functionality to the OTA generation script that allows us to gather information about the APEX stored inside the target-file zip. However, we did not integrate the new functionality with the ota_from_target_files.py scrip yet. That will be done on follow up CL. Bug: 172911822 Test: atest releasetools_py3_test Change-Id: I2ac42018f628c2c21527b3e086be1f4e7e7247ad
255 lines
8.3 KiB
Python
Executable file
255 lines
8.3 KiB
Python
Executable file
#!/usr/bin/env python
|
|
#
|
|
# Copyright (C) 2018 The Android Open Source Project
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
"""
|
|
Utils for running unittests.
|
|
"""
|
|
|
|
import logging
|
|
import os
|
|
import os.path
|
|
import re
|
|
import struct
|
|
import sys
|
|
import unittest
|
|
import zipfile
|
|
|
|
import common
|
|
|
|
# Some test runner doesn't like outputs from stderr.
|
|
logging.basicConfig(stream=sys.stdout)
|
|
|
|
# Use ANDROID_BUILD_TOP as an indicator to tell if the needed tools (e.g.
|
|
# avbtool, mke2fs) are available while running the tests, unless
|
|
# FORCE_RUN_RELEASETOOLS is set to '1'. Not having the required vars means we
|
|
# can't run the tests that require external tools.
|
|
EXTERNAL_TOOLS_UNAVAILABLE = (
|
|
not os.environ.get('ANDROID_BUILD_TOP') and
|
|
os.environ.get('FORCE_RUN_RELEASETOOLS') != '1')
|
|
|
|
|
|
def SkipIfExternalToolsUnavailable():
|
|
"""Decorator function that allows skipping tests per tools availability."""
|
|
if EXTERNAL_TOOLS_UNAVAILABLE:
|
|
return unittest.skip('External tools unavailable')
|
|
return lambda func: func
|
|
|
|
|
|
def get_testdata_dir():
|
|
"""Returns the testdata dir, in relative to the script dir."""
|
|
# The script dir is the one we want, which could be different from pwd.
|
|
current_dir = os.path.dirname(os.path.realpath(__file__))
|
|
return os.path.join(current_dir, 'testdata')
|
|
|
|
def get_current_dir():
|
|
"""Returns the current dir, relative to the script dir."""
|
|
# The script dir is the one we want, which could be different from pwd.
|
|
current_dir = os.path.dirname(os.path.realpath(__file__))
|
|
return current_dir
|
|
|
|
def get_search_path():
|
|
"""Returns the search path that has 'framework/signapk.jar' under."""
|
|
|
|
def signapk_exists(path):
|
|
signapk_path = os.path.realpath(
|
|
os.path.join(path, 'framework', 'signapk.jar'))
|
|
return os.path.exists(signapk_path)
|
|
|
|
# Try with ANDROID_BUILD_TOP first.
|
|
full_path = os.path.realpath(os.path.join(
|
|
os.environ.get('ANDROID_BUILD_TOP', ''), 'out', 'host', 'linux-x86'))
|
|
if signapk_exists(full_path):
|
|
return full_path
|
|
|
|
# Otherwise try going with relative pathes.
|
|
current_dir = os.path.dirname(os.path.realpath(__file__))
|
|
for path in (
|
|
# In relative to 'build/make/tools/releasetools' in the Android source.
|
|
['..'] * 4 + ['out', 'host', 'linux-x86'],
|
|
# Or running the script unpacked from otatools.zip.
|
|
['..']):
|
|
full_path = os.path.realpath(os.path.join(current_dir, *path))
|
|
if signapk_exists(full_path):
|
|
return full_path
|
|
return None
|
|
|
|
|
|
def construct_sparse_image(chunks):
|
|
"""Returns a sparse image file constructed from the given chunks.
|
|
|
|
From system/core/libsparse/sparse_format.h.
|
|
typedef struct sparse_header {
|
|
__le32 magic; // 0xed26ff3a
|
|
__le16 major_version; // (0x1) - reject images with higher major versions
|
|
__le16 minor_version; // (0x0) - allow images with higer minor versions
|
|
__le16 file_hdr_sz; // 28 bytes for first revision of the file format
|
|
__le16 chunk_hdr_sz; // 12 bytes for first revision of the file format
|
|
__le32 blk_sz; // block size in bytes, must be a multiple of 4 (4096)
|
|
__le32 total_blks; // total blocks in the non-sparse output image
|
|
__le32 total_chunks; // total chunks in the sparse input image
|
|
__le32 image_checksum; // CRC32 checksum of the original data, counting
|
|
// "don't care" as 0. Standard 802.3 polynomial,
|
|
// use a Public Domain table implementation
|
|
} sparse_header_t;
|
|
|
|
typedef struct chunk_header {
|
|
__le16 chunk_type; // 0xCAC1 -> raw; 0xCAC2 -> fill;
|
|
// 0xCAC3 -> don't care
|
|
__le16 reserved1;
|
|
__le32 chunk_sz; // in blocks in output image
|
|
__le32 total_sz; // in bytes of chunk input file including chunk header
|
|
// and data
|
|
} chunk_header_t;
|
|
|
|
Args:
|
|
chunks: A list of chunks to be written. Each entry should be a tuple of
|
|
(chunk_type, block_number).
|
|
|
|
Returns:
|
|
Filename of the created sparse image.
|
|
"""
|
|
SPARSE_HEADER_MAGIC = 0xED26FF3A
|
|
SPARSE_HEADER_FORMAT = "<I4H4I"
|
|
CHUNK_HEADER_FORMAT = "<2H2I"
|
|
|
|
sparse_image = common.MakeTempFile(prefix='sparse-', suffix='.img')
|
|
with open(sparse_image, 'wb') as fp:
|
|
fp.write(struct.pack(
|
|
SPARSE_HEADER_FORMAT, SPARSE_HEADER_MAGIC, 1, 0, 28, 12, 4096,
|
|
sum(chunk[1] for chunk in chunks),
|
|
len(chunks), 0))
|
|
|
|
for chunk in chunks:
|
|
data_size = 0
|
|
if chunk[0] == 0xCAC1:
|
|
data_size = 4096 * chunk[1]
|
|
elif chunk[0] == 0xCAC2:
|
|
data_size = 4
|
|
elif chunk[0] == 0xCAC3:
|
|
pass
|
|
else:
|
|
assert False, "Unsupported chunk type: {}".format(chunk[0])
|
|
|
|
fp.write(struct.pack(
|
|
CHUNK_HEADER_FORMAT, chunk[0], 0, chunk[1], data_size + 12))
|
|
if data_size != 0:
|
|
fp.write(os.urandom(data_size))
|
|
|
|
return sparse_image
|
|
|
|
|
|
class MockScriptWriter(object):
|
|
"""A class that mocks edify_generator.EdifyGenerator.
|
|
|
|
It simply pushes the incoming arguments onto script stack, which is to assert
|
|
the calls to EdifyGenerator functions.
|
|
"""
|
|
|
|
def __init__(self, enable_comments=False):
|
|
self.lines = []
|
|
self.enable_comments = enable_comments
|
|
|
|
def Mount(self, *args):
|
|
self.lines.append(('Mount',) + args)
|
|
|
|
def AssertDevice(self, *args):
|
|
self.lines.append(('AssertDevice',) + args)
|
|
|
|
def AssertOemProperty(self, *args):
|
|
self.lines.append(('AssertOemProperty',) + args)
|
|
|
|
def AssertFingerprintOrThumbprint(self, *args):
|
|
self.lines.append(('AssertFingerprintOrThumbprint',) + args)
|
|
|
|
def AssertSomeFingerprint(self, *args):
|
|
self.lines.append(('AssertSomeFingerprint',) + args)
|
|
|
|
def AssertSomeThumbprint(self, *args):
|
|
self.lines.append(('AssertSomeThumbprint',) + args)
|
|
|
|
def Comment(self, comment):
|
|
if not self.enable_comments:
|
|
return
|
|
self.lines.append('# {}'.format(comment))
|
|
|
|
def AppendExtra(self, extra):
|
|
self.lines.append(extra)
|
|
|
|
def __str__(self):
|
|
return '\n'.join(self.lines)
|
|
|
|
|
|
class ReleaseToolsTestCase(unittest.TestCase):
|
|
"""A common base class for all the releasetools unittests."""
|
|
|
|
def tearDown(self):
|
|
common.Cleanup()
|
|
|
|
class PropertyFilesTestCase(ReleaseToolsTestCase):
|
|
|
|
@staticmethod
|
|
def construct_zip_package(entries):
|
|
zip_file = common.MakeTempFile(suffix='.zip')
|
|
with zipfile.ZipFile(zip_file, 'w', allowZip64=True) as zip_fp:
|
|
for entry in entries:
|
|
zip_fp.writestr(
|
|
entry,
|
|
entry.replace('.', '-').upper(),
|
|
zipfile.ZIP_STORED)
|
|
return zip_file
|
|
|
|
@staticmethod
|
|
def _parse_property_files_string(data):
|
|
result = {}
|
|
for token in data.split(','):
|
|
name, info = token.split(':', 1)
|
|
result[name] = info
|
|
return result
|
|
|
|
def setUp(self):
|
|
common.OPTIONS.no_signing = False
|
|
|
|
def _verify_entries(self, input_file, tokens, entries):
|
|
for entry in entries:
|
|
offset, size = map(int, tokens[entry].split(':'))
|
|
with open(input_file, 'rb') as input_fp:
|
|
input_fp.seek(offset)
|
|
if entry == 'metadata':
|
|
expected = b'META-INF/COM/ANDROID/METADATA'
|
|
elif entry == 'metadata.pb':
|
|
expected = b'META-INF/COM/ANDROID/METADATA-PB'
|
|
else:
|
|
expected = entry.replace('.', '-').upper().encode()
|
|
self.assertEqual(expected, input_fp.read(size))
|
|
|
|
|
|
if __name__ == '__main__':
|
|
# We only want to run tests from the top level directory. Unfortunately the
|
|
# pattern option of unittest.discover, internally using fnmatch, doesn't
|
|
# provide a good API to filter the test files based on directory. So we do an
|
|
# os walk and load them manually.
|
|
test_modules = []
|
|
base_path = os.path.dirname(os.path.realpath(__file__))
|
|
for dirpath, _, files in os.walk(base_path):
|
|
for fn in files:
|
|
if dirpath == base_path and re.match('test_.*\\.py$', fn):
|
|
test_modules.append(fn[:-3])
|
|
|
|
test_suite = unittest.TestLoader().loadTestsFromNames(test_modules)
|
|
|
|
# atest needs a verbosity level of >= 2 to correctly parse the result.
|
|
unittest.TextTestRunner(verbosity=2).run(test_suite)
|