c1a1ec30df
Python 2 and 3 behave differently when calling ZipFile.writestr() with
zinfo.external_attr being 0. Python 3 uses `0o600 << 16` as the value
for such a case (since
18ee29d0b8
),
which seems to make more sense. Otherwise the entry will end up with
0o000 as the permission bits. This CL updates common.ZipWriteStr to
follow the logic in Python 3, in order to get consistent behavior
between using the two versions.
Bug: 131631303
Test: `python -m unittest test_common.CommonZipTest`
Test: `python3 -m unittest test_common.CommonZipTest`
Change-Id: If8429855d922ef1ad76591f703215a0ce5089f0f
1351 lines
55 KiB
Python
1351 lines
55 KiB
Python
#
|
|
# Copyright (C) 2015 The Android Open Source Project
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
import copy
|
|
import os
|
|
import subprocess
|
|
import tempfile
|
|
import time
|
|
import zipfile
|
|
from hashlib import sha1
|
|
|
|
import common
|
|
import test_utils
|
|
import validate_target_files
|
|
from rangelib import RangeSet
|
|
|
|
from blockimgdiff import EmptyImage, DataImage
|
|
|
|
KiB = 1024
|
|
MiB = 1024 * KiB
|
|
GiB = 1024 * MiB
|
|
|
|
|
|
def get_2gb_string():
|
|
size = int(2 * GiB + 1)
|
|
block_size = 4 * KiB
|
|
step_size = 4 * MiB
|
|
# Generate a long string with holes, e.g. 'xyz\x00abc\x00...'.
|
|
for _ in range(0, size, step_size):
|
|
yield os.urandom(block_size)
|
|
yield b'\0' * (step_size - block_size)
|
|
|
|
|
|
class CommonZipTest(test_utils.ReleaseToolsTestCase):
|
|
|
|
def _verify(self, zip_file, zip_file_name, arcname, expected_hash,
|
|
test_file_name=None, expected_stat=None, expected_mode=0o644,
|
|
expected_compress_type=zipfile.ZIP_STORED):
|
|
# Verify the stat if present.
|
|
if test_file_name is not None:
|
|
new_stat = os.stat(test_file_name)
|
|
self.assertEqual(int(expected_stat.st_mode), int(new_stat.st_mode))
|
|
self.assertEqual(int(expected_stat.st_mtime), int(new_stat.st_mtime))
|
|
|
|
# Reopen the zip file to verify.
|
|
zip_file = zipfile.ZipFile(zip_file_name, "r")
|
|
|
|
# Verify the timestamp.
|
|
info = zip_file.getinfo(arcname)
|
|
self.assertEqual(info.date_time, (2009, 1, 1, 0, 0, 0))
|
|
|
|
# Verify the file mode.
|
|
mode = (info.external_attr >> 16) & 0o777
|
|
self.assertEqual(mode, expected_mode)
|
|
|
|
# Verify the compress type.
|
|
self.assertEqual(info.compress_type, expected_compress_type)
|
|
|
|
# Verify the zip contents.
|
|
entry = zip_file.open(arcname)
|
|
sha1_hash = sha1()
|
|
for chunk in iter(lambda: entry.read(4 * MiB), b''):
|
|
sha1_hash.update(chunk)
|
|
self.assertEqual(expected_hash, sha1_hash.hexdigest())
|
|
self.assertIsNone(zip_file.testzip())
|
|
|
|
def _test_ZipWrite(self, contents, extra_zipwrite_args=None):
|
|
extra_zipwrite_args = dict(extra_zipwrite_args or {})
|
|
|
|
test_file = tempfile.NamedTemporaryFile(delete=False)
|
|
test_file_name = test_file.name
|
|
|
|
zip_file = tempfile.NamedTemporaryFile(delete=False)
|
|
zip_file_name = zip_file.name
|
|
|
|
# File names within an archive strip the leading slash.
|
|
arcname = extra_zipwrite_args.get("arcname", test_file_name)
|
|
if arcname[0] == "/":
|
|
arcname = arcname[1:]
|
|
|
|
zip_file.close()
|
|
zip_file = zipfile.ZipFile(zip_file_name, "w")
|
|
|
|
try:
|
|
sha1_hash = sha1()
|
|
for data in contents:
|
|
sha1_hash.update(bytes(data))
|
|
test_file.write(bytes(data))
|
|
test_file.close()
|
|
|
|
expected_stat = os.stat(test_file_name)
|
|
expected_mode = extra_zipwrite_args.get("perms", 0o644)
|
|
expected_compress_type = extra_zipwrite_args.get("compress_type",
|
|
zipfile.ZIP_STORED)
|
|
time.sleep(5) # Make sure the atime/mtime will change measurably.
|
|
|
|
common.ZipWrite(zip_file, test_file_name, **extra_zipwrite_args)
|
|
common.ZipClose(zip_file)
|
|
|
|
self._verify(zip_file, zip_file_name, arcname, sha1_hash.hexdigest(),
|
|
test_file_name, expected_stat, expected_mode,
|
|
expected_compress_type)
|
|
finally:
|
|
os.remove(test_file_name)
|
|
os.remove(zip_file_name)
|
|
|
|
def _test_ZipWriteStr(self, zinfo_or_arcname, contents, extra_args=None):
|
|
extra_args = dict(extra_args or {})
|
|
|
|
zip_file = tempfile.NamedTemporaryFile(delete=False)
|
|
zip_file_name = zip_file.name
|
|
zip_file.close()
|
|
|
|
zip_file = zipfile.ZipFile(zip_file_name, "w")
|
|
|
|
try:
|
|
expected_compress_type = extra_args.get("compress_type",
|
|
zipfile.ZIP_STORED)
|
|
time.sleep(5) # Make sure the atime/mtime will change measurably.
|
|
|
|
if not isinstance(zinfo_or_arcname, zipfile.ZipInfo):
|
|
arcname = zinfo_or_arcname
|
|
expected_mode = extra_args.get("perms", 0o644)
|
|
else:
|
|
arcname = zinfo_or_arcname.filename
|
|
if zinfo_or_arcname.external_attr:
|
|
zinfo_perms = zinfo_or_arcname.external_attr >> 16
|
|
else:
|
|
zinfo_perms = 0o600
|
|
expected_mode = extra_args.get("perms", zinfo_perms)
|
|
|
|
common.ZipWriteStr(zip_file, zinfo_or_arcname, contents, **extra_args)
|
|
common.ZipClose(zip_file)
|
|
|
|
self._verify(zip_file, zip_file_name, arcname, sha1(contents).hexdigest(),
|
|
expected_mode=expected_mode,
|
|
expected_compress_type=expected_compress_type)
|
|
finally:
|
|
os.remove(zip_file_name)
|
|
|
|
def _test_ZipWriteStr_large_file(self, large, small, extra_args=None):
|
|
extra_args = dict(extra_args or {})
|
|
|
|
zip_file = tempfile.NamedTemporaryFile(delete=False)
|
|
zip_file_name = zip_file.name
|
|
|
|
test_file = tempfile.NamedTemporaryFile(delete=False)
|
|
test_file_name = test_file.name
|
|
|
|
arcname_large = test_file_name
|
|
arcname_small = "bar"
|
|
|
|
# File names within an archive strip the leading slash.
|
|
if arcname_large[0] == "/":
|
|
arcname_large = arcname_large[1:]
|
|
|
|
zip_file.close()
|
|
zip_file = zipfile.ZipFile(zip_file_name, "w")
|
|
|
|
try:
|
|
sha1_hash = sha1()
|
|
for data in large:
|
|
sha1_hash.update(data)
|
|
test_file.write(data)
|
|
test_file.close()
|
|
|
|
expected_stat = os.stat(test_file_name)
|
|
expected_mode = 0o644
|
|
expected_compress_type = extra_args.get("compress_type",
|
|
zipfile.ZIP_STORED)
|
|
time.sleep(5) # Make sure the atime/mtime will change measurably.
|
|
|
|
common.ZipWrite(zip_file, test_file_name, **extra_args)
|
|
common.ZipWriteStr(zip_file, arcname_small, small, **extra_args)
|
|
common.ZipClose(zip_file)
|
|
|
|
# Verify the contents written by ZipWrite().
|
|
self._verify(zip_file, zip_file_name, arcname_large,
|
|
sha1_hash.hexdigest(), test_file_name, expected_stat,
|
|
expected_mode, expected_compress_type)
|
|
|
|
# Verify the contents written by ZipWriteStr().
|
|
self._verify(zip_file, zip_file_name, arcname_small,
|
|
sha1(small).hexdigest(),
|
|
expected_compress_type=expected_compress_type)
|
|
finally:
|
|
os.remove(zip_file_name)
|
|
os.remove(test_file_name)
|
|
|
|
def _test_reset_ZIP64_LIMIT(self, func, *args):
|
|
default_limit = (1 << 31) - 1
|
|
self.assertEqual(default_limit, zipfile.ZIP64_LIMIT)
|
|
func(*args)
|
|
self.assertEqual(default_limit, zipfile.ZIP64_LIMIT)
|
|
|
|
def test_ZipWrite(self):
|
|
file_contents = os.urandom(1024)
|
|
self._test_ZipWrite(file_contents)
|
|
|
|
def test_ZipWrite_with_opts(self):
|
|
file_contents = os.urandom(1024)
|
|
self._test_ZipWrite(file_contents, {
|
|
"arcname": "foobar",
|
|
"perms": 0o777,
|
|
"compress_type": zipfile.ZIP_DEFLATED,
|
|
})
|
|
self._test_ZipWrite(file_contents, {
|
|
"arcname": "foobar",
|
|
"perms": 0o700,
|
|
"compress_type": zipfile.ZIP_STORED,
|
|
})
|
|
|
|
def test_ZipWrite_large_file(self):
|
|
file_contents = get_2gb_string()
|
|
self._test_ZipWrite(file_contents, {
|
|
"compress_type": zipfile.ZIP_DEFLATED,
|
|
})
|
|
|
|
def test_ZipWrite_resets_ZIP64_LIMIT(self):
|
|
self._test_reset_ZIP64_LIMIT(self._test_ZipWrite, "")
|
|
|
|
def test_ZipWriteStr(self):
|
|
random_string = os.urandom(1024)
|
|
# Passing arcname
|
|
self._test_ZipWriteStr("foo", random_string)
|
|
|
|
# Passing zinfo
|
|
zinfo = zipfile.ZipInfo(filename="foo")
|
|
self._test_ZipWriteStr(zinfo, random_string)
|
|
|
|
# Timestamp in the zinfo should be overwritten.
|
|
zinfo.date_time = (2015, 3, 1, 15, 30, 0)
|
|
self._test_ZipWriteStr(zinfo, random_string)
|
|
|
|
def test_ZipWriteStr_with_opts(self):
|
|
random_string = os.urandom(1024)
|
|
# Passing arcname
|
|
self._test_ZipWriteStr("foo", random_string, {
|
|
"perms": 0o700,
|
|
"compress_type": zipfile.ZIP_DEFLATED,
|
|
})
|
|
self._test_ZipWriteStr("bar", random_string, {
|
|
"compress_type": zipfile.ZIP_STORED,
|
|
})
|
|
|
|
# Passing zinfo
|
|
zinfo = zipfile.ZipInfo(filename="foo")
|
|
self._test_ZipWriteStr(zinfo, random_string, {
|
|
"compress_type": zipfile.ZIP_DEFLATED,
|
|
})
|
|
self._test_ZipWriteStr(zinfo, random_string, {
|
|
"perms": 0o600,
|
|
"compress_type": zipfile.ZIP_STORED,
|
|
})
|
|
self._test_ZipWriteStr(zinfo, random_string, {
|
|
"perms": 0o000,
|
|
"compress_type": zipfile.ZIP_STORED,
|
|
})
|
|
|
|
def test_ZipWriteStr_large_file(self):
|
|
# zipfile.writestr() doesn't work when the str size is over 2GiB even with
|
|
# the workaround. We will only test the case of writing a string into a
|
|
# large archive.
|
|
long_string = get_2gb_string()
|
|
short_string = os.urandom(1024)
|
|
self._test_ZipWriteStr_large_file(long_string, short_string, {
|
|
"compress_type": zipfile.ZIP_DEFLATED,
|
|
})
|
|
|
|
def test_ZipWriteStr_resets_ZIP64_LIMIT(self):
|
|
self._test_reset_ZIP64_LIMIT(self._test_ZipWriteStr, 'foo', b'')
|
|
zinfo = zipfile.ZipInfo(filename="foo")
|
|
self._test_reset_ZIP64_LIMIT(self._test_ZipWriteStr, zinfo, b'')
|
|
|
|
def test_bug21309935(self):
|
|
zip_file = tempfile.NamedTemporaryFile(delete=False)
|
|
zip_file_name = zip_file.name
|
|
zip_file.close()
|
|
|
|
try:
|
|
random_string = os.urandom(1024)
|
|
zip_file = zipfile.ZipFile(zip_file_name, "w")
|
|
# Default perms should be 0o644 when passing the filename.
|
|
common.ZipWriteStr(zip_file, "foo", random_string)
|
|
# Honor the specified perms.
|
|
common.ZipWriteStr(zip_file, "bar", random_string, perms=0o755)
|
|
# The perms in zinfo should be untouched.
|
|
zinfo = zipfile.ZipInfo(filename="baz")
|
|
zinfo.external_attr = 0o740 << 16
|
|
common.ZipWriteStr(zip_file, zinfo, random_string)
|
|
# Explicitly specified perms has the priority.
|
|
zinfo = zipfile.ZipInfo(filename="qux")
|
|
zinfo.external_attr = 0o700 << 16
|
|
common.ZipWriteStr(zip_file, zinfo, random_string, perms=0o400)
|
|
common.ZipClose(zip_file)
|
|
|
|
self._verify(zip_file, zip_file_name, "foo",
|
|
sha1(random_string).hexdigest(),
|
|
expected_mode=0o644)
|
|
self._verify(zip_file, zip_file_name, "bar",
|
|
sha1(random_string).hexdigest(),
|
|
expected_mode=0o755)
|
|
self._verify(zip_file, zip_file_name, "baz",
|
|
sha1(random_string).hexdigest(),
|
|
expected_mode=0o740)
|
|
self._verify(zip_file, zip_file_name, "qux",
|
|
sha1(random_string).hexdigest(),
|
|
expected_mode=0o400)
|
|
finally:
|
|
os.remove(zip_file_name)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_ZipDelete(self):
|
|
zip_file = tempfile.NamedTemporaryFile(delete=False, suffix='.zip')
|
|
output_zip = zipfile.ZipFile(zip_file.name, 'w',
|
|
compression=zipfile.ZIP_DEFLATED)
|
|
with tempfile.NamedTemporaryFile() as entry_file:
|
|
entry_file.write(os.urandom(1024))
|
|
common.ZipWrite(output_zip, entry_file.name, arcname='Test1')
|
|
common.ZipWrite(output_zip, entry_file.name, arcname='Test2')
|
|
common.ZipWrite(output_zip, entry_file.name, arcname='Test3')
|
|
common.ZipClose(output_zip)
|
|
zip_file.close()
|
|
|
|
try:
|
|
common.ZipDelete(zip_file.name, 'Test2')
|
|
with zipfile.ZipFile(zip_file.name, 'r') as check_zip:
|
|
entries = check_zip.namelist()
|
|
self.assertTrue('Test1' in entries)
|
|
self.assertFalse('Test2' in entries)
|
|
self.assertTrue('Test3' in entries)
|
|
|
|
self.assertRaises(
|
|
common.ExternalError, common.ZipDelete, zip_file.name, 'Test2')
|
|
with zipfile.ZipFile(zip_file.name, 'r') as check_zip:
|
|
entries = check_zip.namelist()
|
|
self.assertTrue('Test1' in entries)
|
|
self.assertFalse('Test2' in entries)
|
|
self.assertTrue('Test3' in entries)
|
|
|
|
common.ZipDelete(zip_file.name, ['Test3'])
|
|
with zipfile.ZipFile(zip_file.name, 'r') as check_zip:
|
|
entries = check_zip.namelist()
|
|
self.assertTrue('Test1' in entries)
|
|
self.assertFalse('Test2' in entries)
|
|
self.assertFalse('Test3' in entries)
|
|
|
|
common.ZipDelete(zip_file.name, ['Test1', 'Test2'])
|
|
with zipfile.ZipFile(zip_file.name, 'r') as check_zip:
|
|
entries = check_zip.namelist()
|
|
self.assertFalse('Test1' in entries)
|
|
self.assertFalse('Test2' in entries)
|
|
self.assertFalse('Test3' in entries)
|
|
finally:
|
|
os.remove(zip_file.name)
|
|
|
|
@staticmethod
|
|
def _test_UnzipTemp_createZipFile():
|
|
zip_file = common.MakeTempFile(suffix='.zip')
|
|
output_zip = zipfile.ZipFile(
|
|
zip_file, 'w', compression=zipfile.ZIP_DEFLATED)
|
|
contents = os.urandom(1024)
|
|
with tempfile.NamedTemporaryFile() as entry_file:
|
|
entry_file.write(contents)
|
|
common.ZipWrite(output_zip, entry_file.name, arcname='Test1')
|
|
common.ZipWrite(output_zip, entry_file.name, arcname='Test2')
|
|
common.ZipWrite(output_zip, entry_file.name, arcname='Foo3')
|
|
common.ZipWrite(output_zip, entry_file.name, arcname='Bar4')
|
|
common.ZipWrite(output_zip, entry_file.name, arcname='Dir5/Baz5')
|
|
common.ZipClose(output_zip)
|
|
common.ZipClose(output_zip)
|
|
return zip_file
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_UnzipTemp(self):
|
|
zip_file = self._test_UnzipTemp_createZipFile()
|
|
unzipped_dir = common.UnzipTemp(zip_file)
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test2')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Foo3')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Bar4')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5')))
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_UnzipTemp_withPatterns(self):
|
|
zip_file = self._test_UnzipTemp_createZipFile()
|
|
|
|
unzipped_dir = common.UnzipTemp(zip_file, ['Test1'])
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5')))
|
|
|
|
unzipped_dir = common.UnzipTemp(zip_file, ['Test1', 'Foo3'])
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Foo3')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5')))
|
|
|
|
unzipped_dir = common.UnzipTemp(zip_file, ['Test*', 'Foo3*'])
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test2')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Foo3')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5')))
|
|
|
|
unzipped_dir = common.UnzipTemp(zip_file, ['*Test1', '*Baz*'])
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5')))
|
|
|
|
def test_UnzipTemp_withEmptyPatterns(self):
|
|
zip_file = self._test_UnzipTemp_createZipFile()
|
|
unzipped_dir = common.UnzipTemp(zip_file, [])
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test1')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5')))
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_UnzipTemp_withPartiallyMatchingPatterns(self):
|
|
zip_file = self._test_UnzipTemp_createZipFile()
|
|
unzipped_dir = common.UnzipTemp(zip_file, ['Test*', 'Nonexistent*'])
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1')))
|
|
self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test2')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5')))
|
|
|
|
def test_UnzipTemp_withNoMatchingPatterns(self):
|
|
zip_file = self._test_UnzipTemp_createZipFile()
|
|
unzipped_dir = common.UnzipTemp(zip_file, ['Foo4', 'Nonexistent*'])
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test1')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4')))
|
|
self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5')))
|
|
|
|
|
|
class CommonApkUtilsTest(test_utils.ReleaseToolsTestCase):
|
|
"""Tests the APK utils related functions."""
|
|
|
|
APKCERTS_TXT1 = (
|
|
'name="RecoveryLocalizer.apk" certificate="certs/devkey.x509.pem"'
|
|
' private_key="certs/devkey.pk8"\n'
|
|
'name="Settings.apk"'
|
|
' certificate="build/make/target/product/security/platform.x509.pem"'
|
|
' private_key="build/make/target/product/security/platform.pk8"\n'
|
|
'name="TV.apk" certificate="PRESIGNED" private_key=""\n'
|
|
)
|
|
|
|
APKCERTS_CERTMAP1 = {
|
|
'RecoveryLocalizer.apk' : 'certs/devkey',
|
|
'Settings.apk' : 'build/make/target/product/security/platform',
|
|
'TV.apk' : 'PRESIGNED',
|
|
}
|
|
|
|
APKCERTS_TXT2 = (
|
|
'name="Compressed1.apk" certificate="certs/compressed1.x509.pem"'
|
|
' private_key="certs/compressed1.pk8" compressed="gz"\n'
|
|
'name="Compressed2a.apk" certificate="certs/compressed2.x509.pem"'
|
|
' private_key="certs/compressed2.pk8" compressed="gz"\n'
|
|
'name="Compressed2b.apk" certificate="certs/compressed2.x509.pem"'
|
|
' private_key="certs/compressed2.pk8" compressed="gz"\n'
|
|
'name="Compressed3.apk" certificate="certs/compressed3.x509.pem"'
|
|
' private_key="certs/compressed3.pk8" compressed="gz"\n'
|
|
)
|
|
|
|
APKCERTS_CERTMAP2 = {
|
|
'Compressed1.apk' : 'certs/compressed1',
|
|
'Compressed2a.apk' : 'certs/compressed2',
|
|
'Compressed2b.apk' : 'certs/compressed2',
|
|
'Compressed3.apk' : 'certs/compressed3',
|
|
}
|
|
|
|
APKCERTS_TXT3 = (
|
|
'name="Compressed4.apk" certificate="certs/compressed4.x509.pem"'
|
|
' private_key="certs/compressed4.pk8" compressed="xz"\n'
|
|
)
|
|
|
|
APKCERTS_CERTMAP3 = {
|
|
'Compressed4.apk' : 'certs/compressed4',
|
|
}
|
|
|
|
def setUp(self):
|
|
self.testdata_dir = test_utils.get_testdata_dir()
|
|
|
|
@staticmethod
|
|
def _write_apkcerts_txt(apkcerts_txt, additional=None):
|
|
if additional is None:
|
|
additional = []
|
|
target_files = common.MakeTempFile(suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
target_files_zip.writestr('META/apkcerts.txt', apkcerts_txt)
|
|
for entry in additional:
|
|
target_files_zip.writestr(entry, '')
|
|
return target_files
|
|
|
|
def test_ReadApkCerts_NoncompressedApks(self):
|
|
target_files = self._write_apkcerts_txt(self.APKCERTS_TXT1)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
certmap, ext = common.ReadApkCerts(input_zip)
|
|
|
|
self.assertDictEqual(self.APKCERTS_CERTMAP1, certmap)
|
|
self.assertIsNone(ext)
|
|
|
|
def test_ReadApkCerts_CompressedApks(self):
|
|
# We have "installed" Compressed1.apk.gz only. Note that Compressed3.apk is
|
|
# not stored in '.gz' format, so it shouldn't be considered as installed.
|
|
target_files = self._write_apkcerts_txt(
|
|
self.APKCERTS_TXT2,
|
|
['Compressed1.apk.gz', 'Compressed3.apk'])
|
|
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
certmap, ext = common.ReadApkCerts(input_zip)
|
|
|
|
self.assertDictEqual(self.APKCERTS_CERTMAP2, certmap)
|
|
self.assertEqual('.gz', ext)
|
|
|
|
# Alternative case with '.xz'.
|
|
target_files = self._write_apkcerts_txt(
|
|
self.APKCERTS_TXT3, ['Compressed4.apk.xz'])
|
|
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
certmap, ext = common.ReadApkCerts(input_zip)
|
|
|
|
self.assertDictEqual(self.APKCERTS_CERTMAP3, certmap)
|
|
self.assertEqual('.xz', ext)
|
|
|
|
def test_ReadApkCerts_CompressedAndNoncompressedApks(self):
|
|
target_files = self._write_apkcerts_txt(
|
|
self.APKCERTS_TXT1 + self.APKCERTS_TXT2,
|
|
['Compressed1.apk.gz', 'Compressed3.apk'])
|
|
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
certmap, ext = common.ReadApkCerts(input_zip)
|
|
|
|
certmap_merged = self.APKCERTS_CERTMAP1.copy()
|
|
certmap_merged.update(self.APKCERTS_CERTMAP2)
|
|
self.assertDictEqual(certmap_merged, certmap)
|
|
self.assertEqual('.gz', ext)
|
|
|
|
def test_ReadApkCerts_MultipleCompressionMethods(self):
|
|
target_files = self._write_apkcerts_txt(
|
|
self.APKCERTS_TXT2 + self.APKCERTS_TXT3,
|
|
['Compressed1.apk.gz', 'Compressed4.apk.xz'])
|
|
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
self.assertRaises(ValueError, common.ReadApkCerts, input_zip)
|
|
|
|
def test_ReadApkCerts_MismatchingKeys(self):
|
|
malformed_apkcerts_txt = (
|
|
'name="App1.apk" certificate="certs/cert1.x509.pem"'
|
|
' private_key="certs/cert2.pk8"\n'
|
|
)
|
|
target_files = self._write_apkcerts_txt(malformed_apkcerts_txt)
|
|
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
self.assertRaises(ValueError, common.ReadApkCerts, input_zip)
|
|
|
|
def test_ExtractPublicKey(self):
|
|
cert = os.path.join(self.testdata_dir, 'testkey.x509.pem')
|
|
pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem')
|
|
with open(pubkey, 'rb') as pubkey_fp:
|
|
self.assertEqual(pubkey_fp.read(), common.ExtractPublicKey(cert))
|
|
|
|
def test_ExtractPublicKey_invalidInput(self):
|
|
wrong_input = os.path.join(self.testdata_dir, 'testkey.pk8')
|
|
self.assertRaises(AssertionError, common.ExtractPublicKey, wrong_input)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_ExtractAvbPublicKey(self):
|
|
privkey = os.path.join(self.testdata_dir, 'testkey.key')
|
|
pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem')
|
|
with open(common.ExtractAvbPublicKey(privkey)) as privkey_fp, \
|
|
open(common.ExtractAvbPublicKey(pubkey)) as pubkey_fp:
|
|
self.assertEqual(privkey_fp.read(), pubkey_fp.read())
|
|
|
|
def test_ParseCertificate(self):
|
|
cert = os.path.join(self.testdata_dir, 'testkey.x509.pem')
|
|
|
|
cmd = ['openssl', 'x509', '-in', cert, '-outform', 'DER']
|
|
proc = common.Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
expected, _ = proc.communicate()
|
|
self.assertEqual(0, proc.returncode)
|
|
|
|
with open(cert) as cert_fp:
|
|
actual = common.ParseCertificate(cert_fp.read())
|
|
self.assertEqual(expected, actual)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetMinSdkVersion(self):
|
|
test_app = os.path.join(self.testdata_dir, 'TestApp.apk')
|
|
self.assertEqual('24', common.GetMinSdkVersion(test_app))
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetMinSdkVersion_invalidInput(self):
|
|
self.assertRaises(
|
|
common.ExternalError, common.GetMinSdkVersion, 'does-not-exist.apk')
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetMinSdkVersionInt(self):
|
|
test_app = os.path.join(self.testdata_dir, 'TestApp.apk')
|
|
self.assertEqual(24, common.GetMinSdkVersionInt(test_app, {}))
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetMinSdkVersionInt_invalidInput(self):
|
|
self.assertRaises(
|
|
common.ExternalError, common.GetMinSdkVersionInt, 'does-not-exist.apk',
|
|
{})
|
|
|
|
|
|
class CommonUtilsTest(test_utils.ReleaseToolsTestCase):
|
|
|
|
def setUp(self):
|
|
self.testdata_dir = test_utils.get_testdata_dir()
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetSparseImage_emptyBlockMapFile(self):
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
target_files_zip.write(
|
|
test_utils.construct_sparse_image([
|
|
(0xCAC1, 6),
|
|
(0xCAC3, 3),
|
|
(0xCAC1, 4)]),
|
|
arcname='IMAGES/system.img')
|
|
target_files_zip.writestr('IMAGES/system.map', '')
|
|
target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 8))
|
|
target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3))
|
|
|
|
tempdir = common.UnzipTemp(target_files)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
sparse_image = common.GetSparseImage('system', tempdir, input_zip, False)
|
|
|
|
self.assertDictEqual(
|
|
{
|
|
'__COPY': RangeSet("0"),
|
|
'__NONZERO-0': RangeSet("1-5 9-12"),
|
|
},
|
|
sparse_image.file_map)
|
|
|
|
def test_GetSparseImage_missingImageFile(self):
|
|
self.assertRaises(
|
|
AssertionError, common.GetSparseImage, 'system2', self.testdata_dir,
|
|
None, False)
|
|
self.assertRaises(
|
|
AssertionError, common.GetSparseImage, 'unknown', self.testdata_dir,
|
|
None, False)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetSparseImage_missingBlockMapFile(self):
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
target_files_zip.write(
|
|
test_utils.construct_sparse_image([
|
|
(0xCAC1, 6),
|
|
(0xCAC3, 3),
|
|
(0xCAC1, 4)]),
|
|
arcname='IMAGES/system.img')
|
|
target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 8))
|
|
target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3))
|
|
|
|
tempdir = common.UnzipTemp(target_files)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
self.assertRaises(
|
|
AssertionError, common.GetSparseImage, 'system', tempdir, input_zip,
|
|
False)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetSparseImage_sharedBlocks_notAllowed(self):
|
|
"""Tests the case of having overlapping blocks but disallowed."""
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
target_files_zip.write(
|
|
test_utils.construct_sparse_image([(0xCAC2, 16)]),
|
|
arcname='IMAGES/system.img')
|
|
# Block 10 is shared between two files.
|
|
target_files_zip.writestr(
|
|
'IMAGES/system.map',
|
|
'\n'.join([
|
|
'/system/file1 1-5 9-10',
|
|
'/system/file2 10-12']))
|
|
target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7))
|
|
target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3))
|
|
|
|
tempdir = common.UnzipTemp(target_files)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
self.assertRaises(
|
|
AssertionError, common.GetSparseImage, 'system', tempdir, input_zip,
|
|
False)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetSparseImage_sharedBlocks_allowed(self):
|
|
"""Tests the case for target using BOARD_EXT4_SHARE_DUP_BLOCKS := true."""
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
# Construct an image with a care_map of "0-5 9-12".
|
|
target_files_zip.write(
|
|
test_utils.construct_sparse_image([(0xCAC2, 16)]),
|
|
arcname='IMAGES/system.img')
|
|
# Block 10 is shared between two files.
|
|
target_files_zip.writestr(
|
|
'IMAGES/system.map',
|
|
'\n'.join([
|
|
'/system/file1 1-5 9-10',
|
|
'/system/file2 10-12']))
|
|
target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7))
|
|
target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3))
|
|
|
|
tempdir = common.UnzipTemp(target_files)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
sparse_image = common.GetSparseImage('system', tempdir, input_zip, True)
|
|
|
|
self.assertDictEqual(
|
|
{
|
|
'__COPY': RangeSet("0"),
|
|
'__NONZERO-0': RangeSet("6-8 13-15"),
|
|
'/system/file1': RangeSet("1-5 9-10"),
|
|
'/system/file2': RangeSet("11-12"),
|
|
},
|
|
sparse_image.file_map)
|
|
|
|
# '/system/file2' should be marked with 'uses_shared_blocks', but not with
|
|
# 'incomplete'.
|
|
self.assertTrue(
|
|
sparse_image.file_map['/system/file2'].extra['uses_shared_blocks'])
|
|
self.assertNotIn(
|
|
'incomplete', sparse_image.file_map['/system/file2'].extra)
|
|
|
|
# All other entries should look normal without any tags.
|
|
self.assertFalse(sparse_image.file_map['__COPY'].extra)
|
|
self.assertFalse(sparse_image.file_map['__NONZERO-0'].extra)
|
|
self.assertFalse(sparse_image.file_map['/system/file1'].extra)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetSparseImage_incompleteRanges(self):
|
|
"""Tests the case of ext4 images with holes."""
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
target_files_zip.write(
|
|
test_utils.construct_sparse_image([(0xCAC2, 16)]),
|
|
arcname='IMAGES/system.img')
|
|
target_files_zip.writestr(
|
|
'IMAGES/system.map',
|
|
'\n'.join([
|
|
'/system/file1 1-5 9-10',
|
|
'/system/file2 11-12']))
|
|
target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7))
|
|
# '/system/file2' has less blocks listed (2) than actual (3).
|
|
target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3))
|
|
|
|
tempdir = common.UnzipTemp(target_files)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
sparse_image = common.GetSparseImage('system', tempdir, input_zip, False)
|
|
|
|
self.assertFalse(sparse_image.file_map['/system/file1'].extra)
|
|
self.assertTrue(sparse_image.file_map['/system/file2'].extra['incomplete'])
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetSparseImage_systemRootImage_filenameWithExtraLeadingSlash(self):
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
target_files_zip.write(
|
|
test_utils.construct_sparse_image([(0xCAC2, 16)]),
|
|
arcname='IMAGES/system.img')
|
|
target_files_zip.writestr(
|
|
'IMAGES/system.map',
|
|
'\n'.join([
|
|
'//system/file1 1-5 9-10',
|
|
'//system/file2 11-12',
|
|
'/system/app/file3 13-15']))
|
|
target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7))
|
|
# '/system/file2' has less blocks listed (2) than actual (3).
|
|
target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3))
|
|
# '/system/app/file3' has less blocks listed (3) than actual (4).
|
|
target_files_zip.writestr('SYSTEM/app/file3', os.urandom(4096 * 4))
|
|
|
|
tempdir = common.UnzipTemp(target_files)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
sparse_image = common.GetSparseImage('system', tempdir, input_zip, False)
|
|
|
|
self.assertFalse(sparse_image.file_map['//system/file1'].extra)
|
|
self.assertTrue(sparse_image.file_map['//system/file2'].extra['incomplete'])
|
|
self.assertTrue(
|
|
sparse_image.file_map['/system/app/file3'].extra['incomplete'])
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetSparseImage_systemRootImage_nonSystemFiles(self):
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
target_files_zip.write(
|
|
test_utils.construct_sparse_image([(0xCAC2, 16)]),
|
|
arcname='IMAGES/system.img')
|
|
target_files_zip.writestr(
|
|
'IMAGES/system.map',
|
|
'\n'.join([
|
|
'//system/file1 1-5 9-10',
|
|
'//init.rc 13-15']))
|
|
target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7))
|
|
# '/init.rc' has less blocks listed (3) than actual (4).
|
|
target_files_zip.writestr('ROOT/init.rc', os.urandom(4096 * 4))
|
|
|
|
tempdir = common.UnzipTemp(target_files)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
sparse_image = common.GetSparseImage('system', tempdir, input_zip, False)
|
|
|
|
self.assertFalse(sparse_image.file_map['//system/file1'].extra)
|
|
self.assertTrue(sparse_image.file_map['//init.rc'].extra['incomplete'])
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetSparseImage_fileNotFound(self):
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
target_files_zip.write(
|
|
test_utils.construct_sparse_image([(0xCAC2, 16)]),
|
|
arcname='IMAGES/system.img')
|
|
target_files_zip.writestr(
|
|
'IMAGES/system.map',
|
|
'\n'.join([
|
|
'//system/file1 1-5 9-10',
|
|
'//system/file2 11-12']))
|
|
target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7))
|
|
|
|
tempdir = common.UnzipTemp(target_files)
|
|
with zipfile.ZipFile(target_files, 'r') as input_zip:
|
|
self.assertRaises(
|
|
AssertionError, common.GetSparseImage, 'system', tempdir, input_zip,
|
|
False)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetAvbChainedPartitionArg(self):
|
|
pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem')
|
|
info_dict = {
|
|
'avb_avbtool': 'avbtool',
|
|
'avb_system_key_path': pubkey,
|
|
'avb_system_rollback_index_location': 2,
|
|
}
|
|
args = common.GetAvbChainedPartitionArg('system', info_dict).split(':')
|
|
self.assertEqual(3, len(args))
|
|
self.assertEqual('system', args[0])
|
|
self.assertEqual('2', args[1])
|
|
self.assertTrue(os.path.exists(args[2]))
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetAvbChainedPartitionArg_withPrivateKey(self):
|
|
key = os.path.join(self.testdata_dir, 'testkey.key')
|
|
info_dict = {
|
|
'avb_avbtool': 'avbtool',
|
|
'avb_product_key_path': key,
|
|
'avb_product_rollback_index_location': 2,
|
|
}
|
|
args = common.GetAvbChainedPartitionArg('product', info_dict).split(':')
|
|
self.assertEqual(3, len(args))
|
|
self.assertEqual('product', args[0])
|
|
self.assertEqual('2', args[1])
|
|
self.assertTrue(os.path.exists(args[2]))
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetAvbChainedPartitionArg_withSpecifiedKey(self):
|
|
info_dict = {
|
|
'avb_avbtool': 'avbtool',
|
|
'avb_system_key_path': 'does-not-exist',
|
|
'avb_system_rollback_index_location': 2,
|
|
}
|
|
pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem')
|
|
args = common.GetAvbChainedPartitionArg(
|
|
'system', info_dict, pubkey).split(':')
|
|
self.assertEqual(3, len(args))
|
|
self.assertEqual('system', args[0])
|
|
self.assertEqual('2', args[1])
|
|
self.assertTrue(os.path.exists(args[2]))
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_GetAvbChainedPartitionArg_invalidKey(self):
|
|
pubkey = os.path.join(self.testdata_dir, 'testkey_with_passwd.x509.pem')
|
|
info_dict = {
|
|
'avb_avbtool': 'avbtool',
|
|
'avb_system_key_path': pubkey,
|
|
'avb_system_rollback_index_location': 2,
|
|
}
|
|
self.assertRaises(
|
|
common.ExternalError, common.GetAvbChainedPartitionArg, 'system',
|
|
info_dict)
|
|
|
|
INFO_DICT_DEFAULT = {
|
|
'recovery_api_version': 3,
|
|
'fstab_version': 2,
|
|
'system_root_image': 'true',
|
|
'no_recovery' : 'true',
|
|
'recovery_as_boot': 'true',
|
|
}
|
|
|
|
@staticmethod
|
|
def _test_LoadInfoDict_createTargetFiles(info_dict, fstab_path):
|
|
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
|
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
|
info_values = ''.join(
|
|
['{}={}\n'.format(k, v) for k, v in sorted(info_dict.iteritems())])
|
|
common.ZipWriteStr(target_files_zip, 'META/misc_info.txt', info_values)
|
|
|
|
FSTAB_TEMPLATE = "/dev/block/system {} ext4 ro,barrier=1 defaults"
|
|
if info_dict.get('system_root_image') == 'true':
|
|
fstab_values = FSTAB_TEMPLATE.format('/')
|
|
else:
|
|
fstab_values = FSTAB_TEMPLATE.format('/system')
|
|
common.ZipWriteStr(target_files_zip, fstab_path, fstab_values)
|
|
|
|
common.ZipWriteStr(
|
|
target_files_zip, 'META/file_contexts', 'file-contexts')
|
|
return target_files
|
|
|
|
def test_LoadInfoDict(self):
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
self.INFO_DICT_DEFAULT,
|
|
'BOOT/RAMDISK/system/etc/recovery.fstab')
|
|
with zipfile.ZipFile(target_files, 'r') as target_files_zip:
|
|
loaded_dict = common.LoadInfoDict(target_files_zip)
|
|
self.assertEqual(3, loaded_dict['recovery_api_version'])
|
|
self.assertEqual(2, loaded_dict['fstab_version'])
|
|
self.assertIn('/', loaded_dict['fstab'])
|
|
self.assertIn('/system', loaded_dict['fstab'])
|
|
|
|
def test_LoadInfoDict_legacyRecoveryFstabPath(self):
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
self.INFO_DICT_DEFAULT,
|
|
'BOOT/RAMDISK/etc/recovery.fstab')
|
|
with zipfile.ZipFile(target_files, 'r') as target_files_zip:
|
|
loaded_dict = common.LoadInfoDict(target_files_zip)
|
|
self.assertEqual(3, loaded_dict['recovery_api_version'])
|
|
self.assertEqual(2, loaded_dict['fstab_version'])
|
|
self.assertIn('/', loaded_dict['fstab'])
|
|
self.assertIn('/system', loaded_dict['fstab'])
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_LoadInfoDict_dirInput(self):
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
self.INFO_DICT_DEFAULT,
|
|
'BOOT/RAMDISK/system/etc/recovery.fstab')
|
|
unzipped = common.UnzipTemp(target_files)
|
|
loaded_dict = common.LoadInfoDict(unzipped)
|
|
self.assertEqual(3, loaded_dict['recovery_api_version'])
|
|
self.assertEqual(2, loaded_dict['fstab_version'])
|
|
self.assertIn('/', loaded_dict['fstab'])
|
|
self.assertIn('/system', loaded_dict['fstab'])
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_LoadInfoDict_dirInput_legacyRecoveryFstabPath(self):
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
self.INFO_DICT_DEFAULT,
|
|
'BOOT/RAMDISK/system/etc/recovery.fstab')
|
|
unzipped = common.UnzipTemp(target_files)
|
|
loaded_dict = common.LoadInfoDict(unzipped)
|
|
self.assertEqual(3, loaded_dict['recovery_api_version'])
|
|
self.assertEqual(2, loaded_dict['fstab_version'])
|
|
self.assertIn('/', loaded_dict['fstab'])
|
|
self.assertIn('/system', loaded_dict['fstab'])
|
|
|
|
def test_LoadInfoDict_systemRootImageFalse(self):
|
|
# Devices not using system-as-root nor recovery-as-boot. Non-A/B devices
|
|
# launched prior to P will likely have this config.
|
|
info_dict = copy.copy(self.INFO_DICT_DEFAULT)
|
|
del info_dict['no_recovery']
|
|
del info_dict['system_root_image']
|
|
del info_dict['recovery_as_boot']
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
info_dict,
|
|
'RECOVERY/RAMDISK/system/etc/recovery.fstab')
|
|
with zipfile.ZipFile(target_files, 'r') as target_files_zip:
|
|
loaded_dict = common.LoadInfoDict(target_files_zip)
|
|
self.assertEqual(3, loaded_dict['recovery_api_version'])
|
|
self.assertEqual(2, loaded_dict['fstab_version'])
|
|
self.assertNotIn('/', loaded_dict['fstab'])
|
|
self.assertIn('/system', loaded_dict['fstab'])
|
|
|
|
def test_LoadInfoDict_recoveryAsBootFalse(self):
|
|
# Devices using system-as-root, but with standalone recovery image. Non-A/B
|
|
# devices launched since P will likely have this config.
|
|
info_dict = copy.copy(self.INFO_DICT_DEFAULT)
|
|
del info_dict['no_recovery']
|
|
del info_dict['recovery_as_boot']
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
info_dict,
|
|
'RECOVERY/RAMDISK/system/etc/recovery.fstab')
|
|
with zipfile.ZipFile(target_files, 'r') as target_files_zip:
|
|
loaded_dict = common.LoadInfoDict(target_files_zip)
|
|
self.assertEqual(3, loaded_dict['recovery_api_version'])
|
|
self.assertEqual(2, loaded_dict['fstab_version'])
|
|
self.assertIn('/', loaded_dict['fstab'])
|
|
self.assertIn('/system', loaded_dict['fstab'])
|
|
|
|
def test_LoadInfoDict_noRecoveryTrue(self):
|
|
# Device doesn't have a recovery partition at all.
|
|
info_dict = copy.copy(self.INFO_DICT_DEFAULT)
|
|
del info_dict['recovery_as_boot']
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
info_dict,
|
|
'RECOVERY/RAMDISK/system/etc/recovery.fstab')
|
|
with zipfile.ZipFile(target_files, 'r') as target_files_zip:
|
|
loaded_dict = common.LoadInfoDict(target_files_zip)
|
|
self.assertEqual(3, loaded_dict['recovery_api_version'])
|
|
self.assertEqual(2, loaded_dict['fstab_version'])
|
|
self.assertIsNone(loaded_dict['fstab'])
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_LoadInfoDict_missingMetaMiscInfoTxt(self):
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
self.INFO_DICT_DEFAULT,
|
|
'BOOT/RAMDISK/system/etc/recovery.fstab')
|
|
common.ZipDelete(target_files, 'META/misc_info.txt')
|
|
with zipfile.ZipFile(target_files, 'r') as target_files_zip:
|
|
self.assertRaises(ValueError, common.LoadInfoDict, target_files_zip)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_LoadInfoDict_repacking(self):
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
self.INFO_DICT_DEFAULT,
|
|
'BOOT/RAMDISK/system/etc/recovery.fstab')
|
|
unzipped = common.UnzipTemp(target_files)
|
|
loaded_dict = common.LoadInfoDict(unzipped, True)
|
|
self.assertEqual(3, loaded_dict['recovery_api_version'])
|
|
self.assertEqual(2, loaded_dict['fstab_version'])
|
|
self.assertIn('/', loaded_dict['fstab'])
|
|
self.assertIn('/system', loaded_dict['fstab'])
|
|
self.assertEqual(
|
|
os.path.join(unzipped, 'ROOT'), loaded_dict['root_dir'])
|
|
self.assertEqual(
|
|
os.path.join(unzipped, 'META', 'root_filesystem_config.txt'),
|
|
loaded_dict['root_fs_config'])
|
|
|
|
def test_LoadInfoDict_repackingWithZipFileInput(self):
|
|
target_files = self._test_LoadInfoDict_createTargetFiles(
|
|
self.INFO_DICT_DEFAULT,
|
|
'BOOT/RAMDISK/system/etc/recovery.fstab')
|
|
with zipfile.ZipFile(target_files, 'r') as target_files_zip:
|
|
self.assertRaises(
|
|
AssertionError, common.LoadInfoDict, target_files_zip, True)
|
|
|
|
|
|
class InstallRecoveryScriptFormatTest(test_utils.ReleaseToolsTestCase):
|
|
"""Checks the format of install-recovery.sh.
|
|
|
|
Its format should match between common.py and validate_target_files.py.
|
|
"""
|
|
|
|
def setUp(self):
|
|
self._tempdir = common.MakeTempDir()
|
|
# Create a dummy dict that contains the fstab info for boot&recovery.
|
|
self._info = {"fstab" : {}}
|
|
dummy_fstab = [
|
|
"/dev/soc.0/by-name/boot /boot emmc defaults defaults",
|
|
"/dev/soc.0/by-name/recovery /recovery emmc defaults defaults"]
|
|
self._info["fstab"] = common.LoadRecoveryFSTab("\n".join, 2, dummy_fstab)
|
|
# Construct the gzipped recovery.img and boot.img
|
|
self.recovery_data = bytearray([
|
|
0x1f, 0x8b, 0x08, 0x00, 0x81, 0x11, 0x02, 0x5a, 0x00, 0x03, 0x2b, 0x4a,
|
|
0x4d, 0xce, 0x2f, 0x4b, 0x2d, 0xaa, 0x04, 0x00, 0xc9, 0x93, 0x43, 0xf3,
|
|
0x08, 0x00, 0x00, 0x00
|
|
])
|
|
# echo -n "boot" | gzip -f | hd
|
|
self.boot_data = bytearray([
|
|
0x1f, 0x8b, 0x08, 0x00, 0x8c, 0x12, 0x02, 0x5a, 0x00, 0x03, 0x4b, 0xca,
|
|
0xcf, 0x2f, 0x01, 0x00, 0xc4, 0xae, 0xed, 0x46, 0x04, 0x00, 0x00, 0x00
|
|
])
|
|
|
|
def _out_tmp_sink(self, name, data, prefix="SYSTEM"):
|
|
loc = os.path.join(self._tempdir, prefix, name)
|
|
if not os.path.exists(os.path.dirname(loc)):
|
|
os.makedirs(os.path.dirname(loc))
|
|
with open(loc, "w+") as f:
|
|
f.write(data)
|
|
|
|
def test_full_recovery(self):
|
|
recovery_image = common.File("recovery.img", self.recovery_data)
|
|
boot_image = common.File("boot.img", self.boot_data)
|
|
self._info["full_recovery_image"] = "true"
|
|
|
|
common.MakeRecoveryPatch(self._tempdir, self._out_tmp_sink,
|
|
recovery_image, boot_image, self._info)
|
|
validate_target_files.ValidateInstallRecoveryScript(self._tempdir,
|
|
self._info)
|
|
|
|
@test_utils.SkipIfExternalToolsUnavailable()
|
|
def test_recovery_from_boot(self):
|
|
recovery_image = common.File("recovery.img", self.recovery_data)
|
|
self._out_tmp_sink("recovery.img", recovery_image.data, "IMAGES")
|
|
boot_image = common.File("boot.img", self.boot_data)
|
|
self._out_tmp_sink("boot.img", boot_image.data, "IMAGES")
|
|
|
|
common.MakeRecoveryPatch(self._tempdir, self._out_tmp_sink,
|
|
recovery_image, boot_image, self._info)
|
|
validate_target_files.ValidateInstallRecoveryScript(self._tempdir,
|
|
self._info)
|
|
# Validate 'recovery-from-boot' with bonus argument.
|
|
self._out_tmp_sink("etc/recovery-resource.dat", "bonus", "SYSTEM")
|
|
common.MakeRecoveryPatch(self._tempdir, self._out_tmp_sink,
|
|
recovery_image, boot_image, self._info)
|
|
validate_target_files.ValidateInstallRecoveryScript(self._tempdir,
|
|
self._info)
|
|
|
|
|
|
class MockScriptWriter(object):
|
|
"""A class that mocks edify_generator.EdifyGenerator.
|
|
"""
|
|
def __init__(self, enable_comments=False):
|
|
self.lines = []
|
|
self.enable_comments = enable_comments
|
|
def Comment(self, comment):
|
|
if self.enable_comments:
|
|
self.lines.append("# {}".format(comment))
|
|
def AppendExtra(self, extra):
|
|
self.lines.append(extra)
|
|
def __str__(self):
|
|
return "\n".join(self.lines)
|
|
|
|
|
|
class MockBlockDifference(object):
|
|
def __init__(self, partition, tgt, src=None):
|
|
self.partition = partition
|
|
self.tgt = tgt
|
|
self.src = src
|
|
def WriteScript(self, script, _, progress=None,
|
|
write_verify_script=False):
|
|
if progress:
|
|
script.AppendExtra("progress({})".format(progress))
|
|
script.AppendExtra("patch({});".format(self.partition))
|
|
if write_verify_script:
|
|
self.WritePostInstallVerifyScript(script)
|
|
def WritePostInstallVerifyScript(self, script):
|
|
script.AppendExtra("verify({});".format(self.partition))
|
|
|
|
|
|
class FakeSparseImage(object):
|
|
def __init__(self, size):
|
|
self.blocksize = 4096
|
|
self.total_blocks = size // 4096
|
|
assert size % 4096 == 0, "{} is not a multiple of 4096".format(size)
|
|
|
|
|
|
class DynamicPartitionsDifferenceTest(test_utils.ReleaseToolsTestCase):
|
|
@staticmethod
|
|
def get_op_list(output_path):
|
|
with zipfile.ZipFile(output_path) as output_zip:
|
|
with output_zip.open("dynamic_partitions_op_list") as op_list:
|
|
return [line.strip() for line in op_list.readlines()
|
|
if not line.startswith("#")]
|
|
|
|
def setUp(self):
|
|
self.script = MockScriptWriter()
|
|
self.output_path = common.MakeTempFile(suffix='.zip')
|
|
|
|
def test_full(self):
|
|
target_info = common.LoadDictionaryFromLines("""
|
|
dynamic_partition_list=system vendor
|
|
super_partition_groups=group_foo
|
|
super_group_foo_group_size={group_size}
|
|
super_group_foo_partition_list=system vendor
|
|
""".format(group_size=4 * GiB).split("\n"))
|
|
block_diffs = [MockBlockDifference("system", FakeSparseImage(3 * GiB)),
|
|
MockBlockDifference("vendor", FakeSparseImage(1 * GiB))]
|
|
|
|
dp_diff = common.DynamicPartitionsDifference(target_info, block_diffs)
|
|
with zipfile.ZipFile(self.output_path, 'w') as output_zip:
|
|
dp_diff.WriteScript(self.script, output_zip, write_verify_script=True)
|
|
|
|
self.assertEqual(str(self.script).strip(), """
|
|
assert(update_dynamic_partitions(package_extract_file("dynamic_partitions_op_list")));
|
|
patch(system);
|
|
verify(system);
|
|
unmap_partition("system");
|
|
patch(vendor);
|
|
verify(vendor);
|
|
unmap_partition("vendor");
|
|
""".strip())
|
|
|
|
lines = self.get_op_list(self.output_path)
|
|
|
|
remove_all_groups = lines.index("remove_all_groups")
|
|
add_group = lines.index("add_group group_foo 4294967296")
|
|
add_vendor = lines.index("add vendor group_foo")
|
|
add_system = lines.index("add system group_foo")
|
|
resize_vendor = lines.index("resize vendor 1073741824")
|
|
resize_system = lines.index("resize system 3221225472")
|
|
|
|
self.assertLess(remove_all_groups, add_group,
|
|
"Should add groups after removing all groups")
|
|
self.assertLess(add_group, min(add_vendor, add_system),
|
|
"Should add partitions after adding group")
|
|
self.assertLess(add_system, resize_system,
|
|
"Should resize system after adding it")
|
|
self.assertLess(add_vendor, resize_vendor,
|
|
"Should resize vendor after adding it")
|
|
|
|
def test_inc_groups(self):
|
|
source_info = common.LoadDictionaryFromLines("""
|
|
super_partition_groups=group_foo group_bar group_baz
|
|
super_group_foo_group_size={group_foo_size}
|
|
super_group_bar_group_size={group_bar_size}
|
|
""".format(group_foo_size=4 * GiB, group_bar_size=3 * GiB).split("\n"))
|
|
target_info = common.LoadDictionaryFromLines("""
|
|
super_partition_groups=group_foo group_baz group_qux
|
|
super_group_foo_group_size={group_foo_size}
|
|
super_group_baz_group_size={group_baz_size}
|
|
super_group_qux_group_size={group_qux_size}
|
|
""".format(group_foo_size=3 * GiB, group_baz_size=4 * GiB,
|
|
group_qux_size=1 * GiB).split("\n"))
|
|
|
|
dp_diff = common.DynamicPartitionsDifference(target_info,
|
|
block_diffs=[],
|
|
source_info_dict=source_info)
|
|
with zipfile.ZipFile(self.output_path, 'w') as output_zip:
|
|
dp_diff.WriteScript(self.script, output_zip, write_verify_script=True)
|
|
|
|
lines = self.get_op_list(self.output_path)
|
|
|
|
removed = lines.index("remove_group group_bar")
|
|
shrunk = lines.index("resize_group group_foo 3221225472")
|
|
grown = lines.index("resize_group group_baz 4294967296")
|
|
added = lines.index("add_group group_qux 1073741824")
|
|
|
|
self.assertLess(max(removed, shrunk),
|
|
min(grown, added),
|
|
"ops that remove / shrink partitions must precede ops that "
|
|
"grow / add partitions")
|
|
|
|
def test_incremental(self):
|
|
source_info = common.LoadDictionaryFromLines("""
|
|
dynamic_partition_list=system vendor product product_services
|
|
super_partition_groups=group_foo
|
|
super_group_foo_group_size={group_foo_size}
|
|
super_group_foo_partition_list=system vendor product product_services
|
|
""".format(group_foo_size=4 * GiB).split("\n"))
|
|
target_info = common.LoadDictionaryFromLines("""
|
|
dynamic_partition_list=system vendor product odm
|
|
super_partition_groups=group_foo group_bar
|
|
super_group_foo_group_size={group_foo_size}
|
|
super_group_foo_partition_list=system vendor odm
|
|
super_group_bar_group_size={group_bar_size}
|
|
super_group_bar_partition_list=product
|
|
""".format(group_foo_size=3 * GiB, group_bar_size=1 * GiB).split("\n"))
|
|
|
|
block_diffs = [MockBlockDifference("system", FakeSparseImage(1536 * MiB),
|
|
src=FakeSparseImage(1024 * MiB)),
|
|
MockBlockDifference("vendor", FakeSparseImage(512 * MiB),
|
|
src=FakeSparseImage(1024 * MiB)),
|
|
MockBlockDifference("product", FakeSparseImage(1024 * MiB),
|
|
src=FakeSparseImage(1024 * MiB)),
|
|
MockBlockDifference("product_services", None,
|
|
src=FakeSparseImage(1024 * MiB)),
|
|
MockBlockDifference("odm", FakeSparseImage(1024 * MiB),
|
|
src=None)]
|
|
|
|
dp_diff = common.DynamicPartitionsDifference(target_info, block_diffs,
|
|
source_info_dict=source_info)
|
|
with zipfile.ZipFile(self.output_path, 'w') as output_zip:
|
|
dp_diff.WriteScript(self.script, output_zip, write_verify_script=True)
|
|
|
|
metadata_idx = self.script.lines.index(
|
|
'assert(update_dynamic_partitions(package_extract_file('
|
|
'"dynamic_partitions_op_list")));')
|
|
self.assertLess(self.script.lines.index('patch(vendor);'), metadata_idx)
|
|
self.assertLess(metadata_idx, self.script.lines.index('verify(vendor);'))
|
|
for p in ("product", "system", "odm"):
|
|
patch_idx = self.script.lines.index("patch({});".format(p))
|
|
verify_idx = self.script.lines.index("verify({});".format(p))
|
|
self.assertLess(metadata_idx, patch_idx,
|
|
"Should patch {} after updating metadata".format(p))
|
|
self.assertLess(patch_idx, verify_idx,
|
|
"Should verify {} after patching".format(p))
|
|
|
|
self.assertNotIn("patch(product_services);", self.script.lines)
|
|
|
|
lines = self.get_op_list(self.output_path)
|
|
|
|
remove = lines.index("remove product_services")
|
|
move_product_out = lines.index("move product default")
|
|
shrink = lines.index("resize vendor 536870912")
|
|
shrink_group = lines.index("resize_group group_foo 3221225472")
|
|
add_group_bar = lines.index("add_group group_bar 1073741824")
|
|
add_odm = lines.index("add odm group_foo")
|
|
grow_existing = lines.index("resize system 1610612736")
|
|
grow_added = lines.index("resize odm 1073741824")
|
|
move_product_in = lines.index("move product group_bar")
|
|
|
|
max_idx_move_partition_out_foo = max(remove, move_product_out, shrink)
|
|
min_idx_move_partition_in_foo = min(add_odm, grow_existing, grow_added)
|
|
|
|
self.assertLess(max_idx_move_partition_out_foo, shrink_group,
|
|
"Must shrink group after partitions inside group are shrunk"
|
|
" / removed")
|
|
|
|
self.assertLess(add_group_bar, move_product_in,
|
|
"Must add partitions to group after group is added")
|
|
|
|
self.assertLess(max_idx_move_partition_out_foo,
|
|
min_idx_move_partition_in_foo,
|
|
"Must shrink partitions / remove partitions from group"
|
|
"before adding / moving partitions into group")
|
|
|
|
def test_remove_partition(self):
|
|
source_info = common.LoadDictionaryFromLines("""
|
|
blockimgdiff_versions=3,4
|
|
use_dynamic_partitions=true
|
|
dynamic_partition_list=foo
|
|
super_partition_groups=group_foo
|
|
super_group_foo_group_size={group_foo_size}
|
|
super_group_foo_partition_list=foo
|
|
""".format(group_foo_size=4 * GiB).split("\n"))
|
|
target_info = common.LoadDictionaryFromLines("""
|
|
blockimgdiff_versions=3,4
|
|
use_dynamic_partitions=true
|
|
super_partition_groups=group_foo
|
|
super_group_foo_group_size={group_foo_size}
|
|
""".format(group_foo_size=4 * GiB).split("\n"))
|
|
|
|
common.OPTIONS.info_dict = target_info
|
|
common.OPTIONS.target_info_dict = target_info
|
|
common.OPTIONS.source_info_dict = source_info
|
|
common.OPTIONS.cache_size = 4 * 4096
|
|
|
|
block_diffs = [common.BlockDifference("foo", EmptyImage(),
|
|
src=DataImage("source", pad=True))]
|
|
|
|
dp_diff = common.DynamicPartitionsDifference(target_info, block_diffs,
|
|
source_info_dict=source_info)
|
|
with zipfile.ZipFile(self.output_path, 'w') as output_zip:
|
|
dp_diff.WriteScript(self.script, output_zip, write_verify_script=True)
|
|
|
|
self.assertNotIn("block_image_update", str(self.script),
|
|
"Removed partition should not be patched.")
|
|
|
|
lines = self.get_op_list(self.output_path)
|
|
self.assertEqual(lines, ["remove foo"])
|