# # Copyright (C) 2015 The Android Open Source Project # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import copy import os import subprocess import tempfile import time import zipfile from hashlib import sha1 import common import test_utils import validate_target_files from images import EmptyImage, DataImage from rangelib import RangeSet KiB = 1024 MiB = 1024 * KiB GiB = 1024 * MiB def get_2gb_string(): size = int(2 * GiB + 1) block_size = 4 * KiB step_size = 4 * MiB # Generate a long string with holes, e.g. 'xyz\x00abc\x00...'. for _ in range(0, size, step_size): yield os.urandom(block_size) yield b'\0' * (step_size - block_size) class CommonZipTest(test_utils.ReleaseToolsTestCase): def _verify(self, zip_file, zip_file_name, arcname, expected_hash, test_file_name=None, expected_stat=None, expected_mode=0o644, expected_compress_type=zipfile.ZIP_STORED): # Verify the stat if present. if test_file_name is not None: new_stat = os.stat(test_file_name) self.assertEqual(int(expected_stat.st_mode), int(new_stat.st_mode)) self.assertEqual(int(expected_stat.st_mtime), int(new_stat.st_mtime)) # Reopen the zip file to verify. zip_file = zipfile.ZipFile(zip_file_name, "r") # Verify the timestamp. info = zip_file.getinfo(arcname) self.assertEqual(info.date_time, (2009, 1, 1, 0, 0, 0)) # Verify the file mode. mode = (info.external_attr >> 16) & 0o777 self.assertEqual(mode, expected_mode) # Verify the compress type. self.assertEqual(info.compress_type, expected_compress_type) # Verify the zip contents. entry = zip_file.open(arcname) sha1_hash = sha1() for chunk in iter(lambda: entry.read(4 * MiB), b''): sha1_hash.update(chunk) self.assertEqual(expected_hash, sha1_hash.hexdigest()) self.assertIsNone(zip_file.testzip()) def _test_ZipWrite(self, contents, extra_zipwrite_args=None): extra_zipwrite_args = dict(extra_zipwrite_args or {}) test_file = tempfile.NamedTemporaryFile(delete=False) test_file_name = test_file.name zip_file = tempfile.NamedTemporaryFile(delete=False) zip_file_name = zip_file.name # File names within an archive strip the leading slash. arcname = extra_zipwrite_args.get("arcname", test_file_name) if arcname[0] == "/": arcname = arcname[1:] zip_file.close() zip_file = zipfile.ZipFile(zip_file_name, "w") try: sha1_hash = sha1() for data in contents: sha1_hash.update(bytes(data)) test_file.write(bytes(data)) test_file.close() expected_stat = os.stat(test_file_name) expected_mode = extra_zipwrite_args.get("perms", 0o644) expected_compress_type = extra_zipwrite_args.get("compress_type", zipfile.ZIP_STORED) time.sleep(5) # Make sure the atime/mtime will change measurably. common.ZipWrite(zip_file, test_file_name, **extra_zipwrite_args) common.ZipClose(zip_file) self._verify(zip_file, zip_file_name, arcname, sha1_hash.hexdigest(), test_file_name, expected_stat, expected_mode, expected_compress_type) finally: os.remove(test_file_name) os.remove(zip_file_name) def _test_ZipWriteStr(self, zinfo_or_arcname, contents, extra_args=None): extra_args = dict(extra_args or {}) zip_file = tempfile.NamedTemporaryFile(delete=False) zip_file_name = zip_file.name zip_file.close() zip_file = zipfile.ZipFile(zip_file_name, "w") try: expected_compress_type = extra_args.get("compress_type", zipfile.ZIP_STORED) time.sleep(5) # Make sure the atime/mtime will change measurably. if not isinstance(zinfo_or_arcname, zipfile.ZipInfo): arcname = zinfo_or_arcname expected_mode = extra_args.get("perms", 0o644) else: arcname = zinfo_or_arcname.filename if zinfo_or_arcname.external_attr: zinfo_perms = zinfo_or_arcname.external_attr >> 16 else: zinfo_perms = 0o600 expected_mode = extra_args.get("perms", zinfo_perms) common.ZipWriteStr(zip_file, zinfo_or_arcname, contents, **extra_args) common.ZipClose(zip_file) self._verify(zip_file, zip_file_name, arcname, sha1(contents).hexdigest(), expected_mode=expected_mode, expected_compress_type=expected_compress_type) finally: os.remove(zip_file_name) def _test_ZipWriteStr_large_file(self, large, small, extra_args=None): extra_args = dict(extra_args or {}) zip_file = tempfile.NamedTemporaryFile(delete=False) zip_file_name = zip_file.name test_file = tempfile.NamedTemporaryFile(delete=False) test_file_name = test_file.name arcname_large = test_file_name arcname_small = "bar" # File names within an archive strip the leading slash. if arcname_large[0] == "/": arcname_large = arcname_large[1:] zip_file.close() zip_file = zipfile.ZipFile(zip_file_name, "w") try: sha1_hash = sha1() for data in large: sha1_hash.update(data) test_file.write(data) test_file.close() expected_stat = os.stat(test_file_name) expected_mode = 0o644 expected_compress_type = extra_args.get("compress_type", zipfile.ZIP_STORED) time.sleep(5) # Make sure the atime/mtime will change measurably. common.ZipWrite(zip_file, test_file_name, **extra_args) common.ZipWriteStr(zip_file, arcname_small, small, **extra_args) common.ZipClose(zip_file) # Verify the contents written by ZipWrite(). self._verify(zip_file, zip_file_name, arcname_large, sha1_hash.hexdigest(), test_file_name, expected_stat, expected_mode, expected_compress_type) # Verify the contents written by ZipWriteStr(). self._verify(zip_file, zip_file_name, arcname_small, sha1(small).hexdigest(), expected_compress_type=expected_compress_type) finally: os.remove(zip_file_name) os.remove(test_file_name) def _test_reset_ZIP64_LIMIT(self, func, *args): default_limit = (1 << 31) - 1 self.assertEqual(default_limit, zipfile.ZIP64_LIMIT) func(*args) self.assertEqual(default_limit, zipfile.ZIP64_LIMIT) def test_ZipWrite(self): file_contents = os.urandom(1024) self._test_ZipWrite(file_contents) def test_ZipWrite_with_opts(self): file_contents = os.urandom(1024) self._test_ZipWrite(file_contents, { "arcname": "foobar", "perms": 0o777, "compress_type": zipfile.ZIP_DEFLATED, }) self._test_ZipWrite(file_contents, { "arcname": "foobar", "perms": 0o700, "compress_type": zipfile.ZIP_STORED, }) def test_ZipWrite_large_file(self): file_contents = get_2gb_string() self._test_ZipWrite(file_contents, { "compress_type": zipfile.ZIP_DEFLATED, }) def test_ZipWrite_resets_ZIP64_LIMIT(self): self._test_reset_ZIP64_LIMIT(self._test_ZipWrite, "") def test_ZipWriteStr(self): random_string = os.urandom(1024) # Passing arcname self._test_ZipWriteStr("foo", random_string) # Passing zinfo zinfo = zipfile.ZipInfo(filename="foo") self._test_ZipWriteStr(zinfo, random_string) # Timestamp in the zinfo should be overwritten. zinfo.date_time = (2015, 3, 1, 15, 30, 0) self._test_ZipWriteStr(zinfo, random_string) def test_ZipWriteStr_with_opts(self): random_string = os.urandom(1024) # Passing arcname self._test_ZipWriteStr("foo", random_string, { "perms": 0o700, "compress_type": zipfile.ZIP_DEFLATED, }) self._test_ZipWriteStr("bar", random_string, { "compress_type": zipfile.ZIP_STORED, }) # Passing zinfo zinfo = zipfile.ZipInfo(filename="foo") self._test_ZipWriteStr(zinfo, random_string, { "compress_type": zipfile.ZIP_DEFLATED, }) self._test_ZipWriteStr(zinfo, random_string, { "perms": 0o600, "compress_type": zipfile.ZIP_STORED, }) self._test_ZipWriteStr(zinfo, random_string, { "perms": 0o000, "compress_type": zipfile.ZIP_STORED, }) def test_ZipWriteStr_large_file(self): # zipfile.writestr() doesn't work when the str size is over 2GiB even with # the workaround. We will only test the case of writing a string into a # large archive. long_string = get_2gb_string() short_string = os.urandom(1024) self._test_ZipWriteStr_large_file(long_string, short_string, { "compress_type": zipfile.ZIP_DEFLATED, }) def test_ZipWriteStr_resets_ZIP64_LIMIT(self): self._test_reset_ZIP64_LIMIT(self._test_ZipWriteStr, 'foo', b'') zinfo = zipfile.ZipInfo(filename="foo") self._test_reset_ZIP64_LIMIT(self._test_ZipWriteStr, zinfo, b'') def test_bug21309935(self): zip_file = tempfile.NamedTemporaryFile(delete=False) zip_file_name = zip_file.name zip_file.close() try: random_string = os.urandom(1024) zip_file = zipfile.ZipFile(zip_file_name, "w") # Default perms should be 0o644 when passing the filename. common.ZipWriteStr(zip_file, "foo", random_string) # Honor the specified perms. common.ZipWriteStr(zip_file, "bar", random_string, perms=0o755) # The perms in zinfo should be untouched. zinfo = zipfile.ZipInfo(filename="baz") zinfo.external_attr = 0o740 << 16 common.ZipWriteStr(zip_file, zinfo, random_string) # Explicitly specified perms has the priority. zinfo = zipfile.ZipInfo(filename="qux") zinfo.external_attr = 0o700 << 16 common.ZipWriteStr(zip_file, zinfo, random_string, perms=0o400) common.ZipClose(zip_file) self._verify(zip_file, zip_file_name, "foo", sha1(random_string).hexdigest(), expected_mode=0o644) self._verify(zip_file, zip_file_name, "bar", sha1(random_string).hexdigest(), expected_mode=0o755) self._verify(zip_file, zip_file_name, "baz", sha1(random_string).hexdigest(), expected_mode=0o740) self._verify(zip_file, zip_file_name, "qux", sha1(random_string).hexdigest(), expected_mode=0o400) finally: os.remove(zip_file_name) @test_utils.SkipIfExternalToolsUnavailable() def test_ZipDelete(self): zip_file = tempfile.NamedTemporaryFile(delete=False, suffix='.zip') output_zip = zipfile.ZipFile(zip_file.name, 'w', compression=zipfile.ZIP_DEFLATED) with tempfile.NamedTemporaryFile() as entry_file: entry_file.write(os.urandom(1024)) common.ZipWrite(output_zip, entry_file.name, arcname='Test1') common.ZipWrite(output_zip, entry_file.name, arcname='Test2') common.ZipWrite(output_zip, entry_file.name, arcname='Test3') common.ZipClose(output_zip) zip_file.close() try: common.ZipDelete(zip_file.name, 'Test2') with zipfile.ZipFile(zip_file.name, 'r') as check_zip: entries = check_zip.namelist() self.assertTrue('Test1' in entries) self.assertFalse('Test2' in entries) self.assertTrue('Test3' in entries) self.assertRaises( common.ExternalError, common.ZipDelete, zip_file.name, 'Test2') with zipfile.ZipFile(zip_file.name, 'r') as check_zip: entries = check_zip.namelist() self.assertTrue('Test1' in entries) self.assertFalse('Test2' in entries) self.assertTrue('Test3' in entries) common.ZipDelete(zip_file.name, ['Test3']) with zipfile.ZipFile(zip_file.name, 'r') as check_zip: entries = check_zip.namelist() self.assertTrue('Test1' in entries) self.assertFalse('Test2' in entries) self.assertFalse('Test3' in entries) common.ZipDelete(zip_file.name, ['Test1', 'Test2']) with zipfile.ZipFile(zip_file.name, 'r') as check_zip: entries = check_zip.namelist() self.assertFalse('Test1' in entries) self.assertFalse('Test2' in entries) self.assertFalse('Test3' in entries) finally: os.remove(zip_file.name) @staticmethod def _test_UnzipTemp_createZipFile(): zip_file = common.MakeTempFile(suffix='.zip') output_zip = zipfile.ZipFile( zip_file, 'w', compression=zipfile.ZIP_DEFLATED) contents = os.urandom(1024) with tempfile.NamedTemporaryFile() as entry_file: entry_file.write(contents) common.ZipWrite(output_zip, entry_file.name, arcname='Test1') common.ZipWrite(output_zip, entry_file.name, arcname='Test2') common.ZipWrite(output_zip, entry_file.name, arcname='Foo3') common.ZipWrite(output_zip, entry_file.name, arcname='Bar4') common.ZipWrite(output_zip, entry_file.name, arcname='Dir5/Baz5') common.ZipClose(output_zip) common.ZipClose(output_zip) return zip_file @test_utils.SkipIfExternalToolsUnavailable() def test_UnzipTemp(self): zip_file = self._test_UnzipTemp_createZipFile() unzipped_dir = common.UnzipTemp(zip_file) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test2'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Foo3'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Bar4'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5'))) @test_utils.SkipIfExternalToolsUnavailable() def test_UnzipTemp_withPatterns(self): zip_file = self._test_UnzipTemp_createZipFile() unzipped_dir = common.UnzipTemp(zip_file, ['Test1']) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5'))) unzipped_dir = common.UnzipTemp(zip_file, ['Test1', 'Foo3']) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Foo3'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5'))) unzipped_dir = common.UnzipTemp(zip_file, ['Test*', 'Foo3*']) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test2'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Foo3'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5'))) unzipped_dir = common.UnzipTemp(zip_file, ['*Test1', '*Baz*']) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5'))) def test_UnzipTemp_withEmptyPatterns(self): zip_file = self._test_UnzipTemp_createZipFile() unzipped_dir = common.UnzipTemp(zip_file, []) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test1'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5'))) @test_utils.SkipIfExternalToolsUnavailable() def test_UnzipTemp_withPartiallyMatchingPatterns(self): zip_file = self._test_UnzipTemp_createZipFile() unzipped_dir = common.UnzipTemp(zip_file, ['Test*', 'Nonexistent*']) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test1'))) self.assertTrue(os.path.exists(os.path.join(unzipped_dir, 'Test2'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5'))) def test_UnzipTemp_withNoMatchingPatterns(self): zip_file = self._test_UnzipTemp_createZipFile() unzipped_dir = common.UnzipTemp(zip_file, ['Foo4', 'Nonexistent*']) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test1'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Test2'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Foo3'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Bar4'))) self.assertFalse(os.path.exists(os.path.join(unzipped_dir, 'Dir5/Baz5'))) class CommonApkUtilsTest(test_utils.ReleaseToolsTestCase): """Tests the APK utils related functions.""" APKCERTS_TXT1 = ( 'name="RecoveryLocalizer.apk" certificate="certs/devkey.x509.pem"' ' private_key="certs/devkey.pk8"\n' 'name="Settings.apk"' ' certificate="build/make/target/product/security/platform.x509.pem"' ' private_key="build/make/target/product/security/platform.pk8"\n' 'name="TV.apk" certificate="PRESIGNED" private_key=""\n' ) APKCERTS_CERTMAP1 = { 'RecoveryLocalizer.apk' : 'certs/devkey', 'Settings.apk' : 'build/make/target/product/security/platform', 'TV.apk' : 'PRESIGNED', } APKCERTS_TXT2 = ( 'name="Compressed1.apk" certificate="certs/compressed1.x509.pem"' ' private_key="certs/compressed1.pk8" compressed="gz"\n' 'name="Compressed2a.apk" certificate="certs/compressed2.x509.pem"' ' private_key="certs/compressed2.pk8" compressed="gz"\n' 'name="Compressed2b.apk" certificate="certs/compressed2.x509.pem"' ' private_key="certs/compressed2.pk8" compressed="gz"\n' 'name="Compressed3.apk" certificate="certs/compressed3.x509.pem"' ' private_key="certs/compressed3.pk8" compressed="gz"\n' ) APKCERTS_CERTMAP2 = { 'Compressed1.apk' : 'certs/compressed1', 'Compressed2a.apk' : 'certs/compressed2', 'Compressed2b.apk' : 'certs/compressed2', 'Compressed3.apk' : 'certs/compressed3', } APKCERTS_TXT3 = ( 'name="Compressed4.apk" certificate="certs/compressed4.x509.pem"' ' private_key="certs/compressed4.pk8" compressed="xz"\n' ) APKCERTS_CERTMAP3 = { 'Compressed4.apk' : 'certs/compressed4', } def setUp(self): self.testdata_dir = test_utils.get_testdata_dir() @staticmethod def _write_apkcerts_txt(apkcerts_txt, additional=None): if additional is None: additional = [] target_files = common.MakeTempFile(suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: target_files_zip.writestr('META/apkcerts.txt', apkcerts_txt) for entry in additional: target_files_zip.writestr(entry, '') return target_files def test_ReadApkCerts_NoncompressedApks(self): target_files = self._write_apkcerts_txt(self.APKCERTS_TXT1) with zipfile.ZipFile(target_files, 'r') as input_zip: certmap, ext = common.ReadApkCerts(input_zip) self.assertDictEqual(self.APKCERTS_CERTMAP1, certmap) self.assertIsNone(ext) def test_ReadApkCerts_CompressedApks(self): # We have "installed" Compressed1.apk.gz only. Note that Compressed3.apk is # not stored in '.gz' format, so it shouldn't be considered as installed. target_files = self._write_apkcerts_txt( self.APKCERTS_TXT2, ['Compressed1.apk.gz', 'Compressed3.apk']) with zipfile.ZipFile(target_files, 'r') as input_zip: certmap, ext = common.ReadApkCerts(input_zip) self.assertDictEqual(self.APKCERTS_CERTMAP2, certmap) self.assertEqual('.gz', ext) # Alternative case with '.xz'. target_files = self._write_apkcerts_txt( self.APKCERTS_TXT3, ['Compressed4.apk.xz']) with zipfile.ZipFile(target_files, 'r') as input_zip: certmap, ext = common.ReadApkCerts(input_zip) self.assertDictEqual(self.APKCERTS_CERTMAP3, certmap) self.assertEqual('.xz', ext) def test_ReadApkCerts_CompressedAndNoncompressedApks(self): target_files = self._write_apkcerts_txt( self.APKCERTS_TXT1 + self.APKCERTS_TXT2, ['Compressed1.apk.gz', 'Compressed3.apk']) with zipfile.ZipFile(target_files, 'r') as input_zip: certmap, ext = common.ReadApkCerts(input_zip) certmap_merged = self.APKCERTS_CERTMAP1.copy() certmap_merged.update(self.APKCERTS_CERTMAP2) self.assertDictEqual(certmap_merged, certmap) self.assertEqual('.gz', ext) def test_ReadApkCerts_MultipleCompressionMethods(self): target_files = self._write_apkcerts_txt( self.APKCERTS_TXT2 + self.APKCERTS_TXT3, ['Compressed1.apk.gz', 'Compressed4.apk.xz']) with zipfile.ZipFile(target_files, 'r') as input_zip: self.assertRaises(ValueError, common.ReadApkCerts, input_zip) def test_ReadApkCerts_MismatchingKeys(self): malformed_apkcerts_txt = ( 'name="App1.apk" certificate="certs/cert1.x509.pem"' ' private_key="certs/cert2.pk8"\n' ) target_files = self._write_apkcerts_txt(malformed_apkcerts_txt) with zipfile.ZipFile(target_files, 'r') as input_zip: self.assertRaises(ValueError, common.ReadApkCerts, input_zip) def test_ExtractPublicKey(self): cert = os.path.join(self.testdata_dir, 'testkey.x509.pem') pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem') with open(pubkey) as pubkey_fp: self.assertEqual(pubkey_fp.read(), common.ExtractPublicKey(cert)) def test_ExtractPublicKey_invalidInput(self): wrong_input = os.path.join(self.testdata_dir, 'testkey.pk8') self.assertRaises(AssertionError, common.ExtractPublicKey, wrong_input) @test_utils.SkipIfExternalToolsUnavailable() def test_ExtractAvbPublicKey(self): privkey = os.path.join(self.testdata_dir, 'testkey.key') pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem') extracted_from_privkey = common.ExtractAvbPublicKey('avbtool', privkey) extracted_from_pubkey = common.ExtractAvbPublicKey('avbtool', pubkey) with open(extracted_from_privkey, 'rb') as privkey_fp, \ open(extracted_from_pubkey, 'rb') as pubkey_fp: self.assertEqual(privkey_fp.read(), pubkey_fp.read()) def test_ParseCertificate(self): cert = os.path.join(self.testdata_dir, 'testkey.x509.pem') cmd = ['openssl', 'x509', '-in', cert, '-outform', 'DER'] proc = common.Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=False) expected, _ = proc.communicate() self.assertEqual(0, proc.returncode) with open(cert) as cert_fp: actual = common.ParseCertificate(cert_fp.read()) self.assertEqual(expected, actual) @test_utils.SkipIfExternalToolsUnavailable() def test_GetMinSdkVersion(self): test_app = os.path.join(self.testdata_dir, 'TestApp.apk') self.assertEqual('24', common.GetMinSdkVersion(test_app)) @test_utils.SkipIfExternalToolsUnavailable() def test_GetMinSdkVersion_invalidInput(self): self.assertRaises( common.ExternalError, common.GetMinSdkVersion, 'does-not-exist.apk') @test_utils.SkipIfExternalToolsUnavailable() def test_GetMinSdkVersionInt(self): test_app = os.path.join(self.testdata_dir, 'TestApp.apk') self.assertEqual(24, common.GetMinSdkVersionInt(test_app, {})) @test_utils.SkipIfExternalToolsUnavailable() def test_GetMinSdkVersionInt_invalidInput(self): self.assertRaises( common.ExternalError, common.GetMinSdkVersionInt, 'does-not-exist.apk', {}) class CommonUtilsTest(test_utils.ReleaseToolsTestCase): def setUp(self): self.testdata_dir = test_utils.get_testdata_dir() @test_utils.SkipIfExternalToolsUnavailable() def test_GetSparseImage_emptyBlockMapFile(self): target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: target_files_zip.write( test_utils.construct_sparse_image([ (0xCAC1, 6), (0xCAC3, 3), (0xCAC1, 4)]), arcname='IMAGES/system.img') target_files_zip.writestr('IMAGES/system.map', '') target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 8)) target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3)) tempdir = common.UnzipTemp(target_files) with zipfile.ZipFile(target_files, 'r') as input_zip: sparse_image = common.GetSparseImage('system', tempdir, input_zip, False) self.assertDictEqual( { '__COPY': RangeSet("0"), '__NONZERO-0': RangeSet("1-5 9-12"), }, sparse_image.file_map) def test_GetSparseImage_missingImageFile(self): self.assertRaises( AssertionError, common.GetSparseImage, 'system2', self.testdata_dir, None, False) self.assertRaises( AssertionError, common.GetSparseImage, 'unknown', self.testdata_dir, None, False) @test_utils.SkipIfExternalToolsUnavailable() def test_GetSparseImage_missingBlockMapFile(self): target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: target_files_zip.write( test_utils.construct_sparse_image([ (0xCAC1, 6), (0xCAC3, 3), (0xCAC1, 4)]), arcname='IMAGES/system.img') target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 8)) target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3)) tempdir = common.UnzipTemp(target_files) with zipfile.ZipFile(target_files, 'r') as input_zip: self.assertRaises( AssertionError, common.GetSparseImage, 'system', tempdir, input_zip, False) @test_utils.SkipIfExternalToolsUnavailable() def test_GetSparseImage_sharedBlocks_notAllowed(self): """Tests the case of having overlapping blocks but disallowed.""" target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: target_files_zip.write( test_utils.construct_sparse_image([(0xCAC2, 16)]), arcname='IMAGES/system.img') # Block 10 is shared between two files. target_files_zip.writestr( 'IMAGES/system.map', '\n'.join([ '/system/file1 1-5 9-10', '/system/file2 10-12'])) target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7)) target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3)) tempdir = common.UnzipTemp(target_files) with zipfile.ZipFile(target_files, 'r') as input_zip: self.assertRaises( AssertionError, common.GetSparseImage, 'system', tempdir, input_zip, False) @test_utils.SkipIfExternalToolsUnavailable() def test_GetSparseImage_sharedBlocks_allowed(self): """Tests the case for target using BOARD_EXT4_SHARE_DUP_BLOCKS := true.""" target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: # Construct an image with a care_map of "0-5 9-12". target_files_zip.write( test_utils.construct_sparse_image([(0xCAC2, 16)]), arcname='IMAGES/system.img') # Block 10 is shared between two files. target_files_zip.writestr( 'IMAGES/system.map', '\n'.join([ '/system/file1 1-5 9-10', '/system/file2 10-12'])) target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7)) target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3)) tempdir = common.UnzipTemp(target_files) with zipfile.ZipFile(target_files, 'r') as input_zip: sparse_image = common.GetSparseImage('system', tempdir, input_zip, True) self.assertDictEqual( { '__COPY': RangeSet("0"), '__NONZERO-0': RangeSet("6-8 13-15"), '/system/file1': RangeSet("1-5 9-10"), '/system/file2': RangeSet("11-12"), }, sparse_image.file_map) # '/system/file2' should be marked with 'uses_shared_blocks', but not with # 'incomplete'. self.assertTrue( sparse_image.file_map['/system/file2'].extra['uses_shared_blocks']) self.assertNotIn( 'incomplete', sparse_image.file_map['/system/file2'].extra) # All other entries should look normal without any tags. self.assertFalse(sparse_image.file_map['__COPY'].extra) self.assertFalse(sparse_image.file_map['__NONZERO-0'].extra) self.assertFalse(sparse_image.file_map['/system/file1'].extra) @test_utils.SkipIfExternalToolsUnavailable() def test_GetSparseImage_incompleteRanges(self): """Tests the case of ext4 images with holes.""" target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: target_files_zip.write( test_utils.construct_sparse_image([(0xCAC2, 16)]), arcname='IMAGES/system.img') target_files_zip.writestr( 'IMAGES/system.map', '\n'.join([ '/system/file1 1-5 9-10', '/system/file2 11-12'])) target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7)) # '/system/file2' has less blocks listed (2) than actual (3). target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3)) tempdir = common.UnzipTemp(target_files) with zipfile.ZipFile(target_files, 'r') as input_zip: sparse_image = common.GetSparseImage('system', tempdir, input_zip, False) self.assertFalse(sparse_image.file_map['/system/file1'].extra) self.assertTrue(sparse_image.file_map['/system/file2'].extra['incomplete']) @test_utils.SkipIfExternalToolsUnavailable() def test_GetSparseImage_systemRootImage_filenameWithExtraLeadingSlash(self): target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: target_files_zip.write( test_utils.construct_sparse_image([(0xCAC2, 16)]), arcname='IMAGES/system.img') target_files_zip.writestr( 'IMAGES/system.map', '\n'.join([ '//system/file1 1-5 9-10', '//system/file2 11-12', '/system/app/file3 13-15'])) target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7)) # '/system/file2' has less blocks listed (2) than actual (3). target_files_zip.writestr('SYSTEM/file2', os.urandom(4096 * 3)) # '/system/app/file3' has less blocks listed (3) than actual (4). target_files_zip.writestr('SYSTEM/app/file3', os.urandom(4096 * 4)) tempdir = common.UnzipTemp(target_files) with zipfile.ZipFile(target_files, 'r') as input_zip: sparse_image = common.GetSparseImage('system', tempdir, input_zip, False) self.assertFalse(sparse_image.file_map['//system/file1'].extra) self.assertTrue(sparse_image.file_map['//system/file2'].extra['incomplete']) self.assertTrue( sparse_image.file_map['/system/app/file3'].extra['incomplete']) @test_utils.SkipIfExternalToolsUnavailable() def test_GetSparseImage_systemRootImage_nonSystemFiles(self): target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: target_files_zip.write( test_utils.construct_sparse_image([(0xCAC2, 16)]), arcname='IMAGES/system.img') target_files_zip.writestr( 'IMAGES/system.map', '\n'.join([ '//system/file1 1-5 9-10', '//init.rc 13-15'])) target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7)) # '/init.rc' has less blocks listed (3) than actual (4). target_files_zip.writestr('ROOT/init.rc', os.urandom(4096 * 4)) tempdir = common.UnzipTemp(target_files) with zipfile.ZipFile(target_files, 'r') as input_zip: sparse_image = common.GetSparseImage('system', tempdir, input_zip, False) self.assertFalse(sparse_image.file_map['//system/file1'].extra) self.assertTrue(sparse_image.file_map['//init.rc'].extra['incomplete']) @test_utils.SkipIfExternalToolsUnavailable() def test_GetSparseImage_fileNotFound(self): target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: target_files_zip.write( test_utils.construct_sparse_image([(0xCAC2, 16)]), arcname='IMAGES/system.img') target_files_zip.writestr( 'IMAGES/system.map', '\n'.join([ '//system/file1 1-5 9-10', '//system/file2 11-12'])) target_files_zip.writestr('SYSTEM/file1', os.urandom(4096 * 7)) tempdir = common.UnzipTemp(target_files) with zipfile.ZipFile(target_files, 'r') as input_zip: self.assertRaises( AssertionError, common.GetSparseImage, 'system', tempdir, input_zip, False) @test_utils.SkipIfExternalToolsUnavailable() def test_GetAvbChainedPartitionArg(self): pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem') info_dict = { 'avb_avbtool': 'avbtool', 'avb_system_key_path': pubkey, 'avb_system_rollback_index_location': 2, } args = common.GetAvbChainedPartitionArg('system', info_dict).split(':') self.assertEqual(3, len(args)) self.assertEqual('system', args[0]) self.assertEqual('2', args[1]) self.assertTrue(os.path.exists(args[2])) @test_utils.SkipIfExternalToolsUnavailable() def test_GetAvbChainedPartitionArg_withPrivateKey(self): key = os.path.join(self.testdata_dir, 'testkey.key') info_dict = { 'avb_avbtool': 'avbtool', 'avb_product_key_path': key, 'avb_product_rollback_index_location': 2, } args = common.GetAvbChainedPartitionArg('product', info_dict).split(':') self.assertEqual(3, len(args)) self.assertEqual('product', args[0]) self.assertEqual('2', args[1]) self.assertTrue(os.path.exists(args[2])) @test_utils.SkipIfExternalToolsUnavailable() def test_GetAvbChainedPartitionArg_withSpecifiedKey(self): info_dict = { 'avb_avbtool': 'avbtool', 'avb_system_key_path': 'does-not-exist', 'avb_system_rollback_index_location': 2, } pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem') args = common.GetAvbChainedPartitionArg( 'system', info_dict, pubkey).split(':') self.assertEqual(3, len(args)) self.assertEqual('system', args[0]) self.assertEqual('2', args[1]) self.assertTrue(os.path.exists(args[2])) @test_utils.SkipIfExternalToolsUnavailable() def test_GetAvbChainedPartitionArg_invalidKey(self): pubkey = os.path.join(self.testdata_dir, 'testkey_with_passwd.x509.pem') info_dict = { 'avb_avbtool': 'avbtool', 'avb_system_key_path': pubkey, 'avb_system_rollback_index_location': 2, } self.assertRaises( common.ExternalError, common.GetAvbChainedPartitionArg, 'system', info_dict) INFO_DICT_DEFAULT = { 'recovery_api_version': 3, 'fstab_version': 2, 'system_root_image': 'true', 'no_recovery' : 'true', 'recovery_as_boot': 'true', } def test_LoadListFromFile(self): file_path = os.path.join(self.testdata_dir, 'merge_config_framework_item_list') contents = common.LoadListFromFile(file_path) expected_contents = [ 'META/apkcerts.txt', 'META/filesystem_config.txt', 'META/root_filesystem_config.txt', 'META/system_manifest.xml', 'META/system_matrix.xml', 'META/update_engine_config.txt', 'PRODUCT/*', 'ROOT/*', 'SYSTEM/*', ] self.assertEqual(sorted(contents), sorted(expected_contents)) @staticmethod def _test_LoadInfoDict_createTargetFiles(info_dict, fstab_path): target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip') with zipfile.ZipFile(target_files, 'w') as target_files_zip: info_values = ''.join( ['{}={}\n'.format(k, v) for k, v in sorted(info_dict.items())]) common.ZipWriteStr(target_files_zip, 'META/misc_info.txt', info_values) FSTAB_TEMPLATE = "/dev/block/system {} ext4 ro,barrier=1 defaults" if info_dict.get('system_root_image') == 'true': fstab_values = FSTAB_TEMPLATE.format('/') else: fstab_values = FSTAB_TEMPLATE.format('/system') common.ZipWriteStr(target_files_zip, fstab_path, fstab_values) common.ZipWriteStr( target_files_zip, 'META/file_contexts', 'file-contexts') return target_files def test_LoadInfoDict(self): target_files = self._test_LoadInfoDict_createTargetFiles( self.INFO_DICT_DEFAULT, 'BOOT/RAMDISK/system/etc/recovery.fstab') with zipfile.ZipFile(target_files, 'r') as target_files_zip: loaded_dict = common.LoadInfoDict(target_files_zip) self.assertEqual(3, loaded_dict['recovery_api_version']) self.assertEqual(2, loaded_dict['fstab_version']) self.assertIn('/', loaded_dict['fstab']) self.assertIn('/system', loaded_dict['fstab']) def test_LoadInfoDict_legacyRecoveryFstabPath(self): target_files = self._test_LoadInfoDict_createTargetFiles( self.INFO_DICT_DEFAULT, 'BOOT/RAMDISK/etc/recovery.fstab') with zipfile.ZipFile(target_files, 'r') as target_files_zip: loaded_dict = common.LoadInfoDict(target_files_zip) self.assertEqual(3, loaded_dict['recovery_api_version']) self.assertEqual(2, loaded_dict['fstab_version']) self.assertIn('/', loaded_dict['fstab']) self.assertIn('/system', loaded_dict['fstab']) @test_utils.SkipIfExternalToolsUnavailable() def test_LoadInfoDict_dirInput(self): target_files = self._test_LoadInfoDict_createTargetFiles( self.INFO_DICT_DEFAULT, 'BOOT/RAMDISK/system/etc/recovery.fstab') unzipped = common.UnzipTemp(target_files) loaded_dict = common.LoadInfoDict(unzipped) self.assertEqual(3, loaded_dict['recovery_api_version']) self.assertEqual(2, loaded_dict['fstab_version']) self.assertIn('/', loaded_dict['fstab']) self.assertIn('/system', loaded_dict['fstab']) @test_utils.SkipIfExternalToolsUnavailable() def test_LoadInfoDict_dirInput_legacyRecoveryFstabPath(self): target_files = self._test_LoadInfoDict_createTargetFiles( self.INFO_DICT_DEFAULT, 'BOOT/RAMDISK/system/etc/recovery.fstab') unzipped = common.UnzipTemp(target_files) loaded_dict = common.LoadInfoDict(unzipped) self.assertEqual(3, loaded_dict['recovery_api_version']) self.assertEqual(2, loaded_dict['fstab_version']) self.assertIn('/', loaded_dict['fstab']) self.assertIn('/system', loaded_dict['fstab']) def test_LoadInfoDict_systemRootImageFalse(self): # Devices not using system-as-root nor recovery-as-boot. Non-A/B devices # launched prior to P will likely have this config. info_dict = copy.copy(self.INFO_DICT_DEFAULT) del info_dict['no_recovery'] del info_dict['system_root_image'] del info_dict['recovery_as_boot'] target_files = self._test_LoadInfoDict_createTargetFiles( info_dict, 'RECOVERY/RAMDISK/system/etc/recovery.fstab') with zipfile.ZipFile(target_files, 'r') as target_files_zip: loaded_dict = common.LoadInfoDict(target_files_zip) self.assertEqual(3, loaded_dict['recovery_api_version']) self.assertEqual(2, loaded_dict['fstab_version']) self.assertNotIn('/', loaded_dict['fstab']) self.assertIn('/system', loaded_dict['fstab']) def test_LoadInfoDict_recoveryAsBootFalse(self): # Devices using system-as-root, but with standalone recovery image. Non-A/B # devices launched since P will likely have this config. info_dict = copy.copy(self.INFO_DICT_DEFAULT) del info_dict['no_recovery'] del info_dict['recovery_as_boot'] target_files = self._test_LoadInfoDict_createTargetFiles( info_dict, 'RECOVERY/RAMDISK/system/etc/recovery.fstab') with zipfile.ZipFile(target_files, 'r') as target_files_zip: loaded_dict = common.LoadInfoDict(target_files_zip) self.assertEqual(3, loaded_dict['recovery_api_version']) self.assertEqual(2, loaded_dict['fstab_version']) self.assertIn('/', loaded_dict['fstab']) self.assertIn('/system', loaded_dict['fstab']) def test_LoadInfoDict_noRecoveryTrue(self): # Device doesn't have a recovery partition at all. info_dict = copy.copy(self.INFO_DICT_DEFAULT) del info_dict['recovery_as_boot'] target_files = self._test_LoadInfoDict_createTargetFiles( info_dict, 'RECOVERY/RAMDISK/system/etc/recovery.fstab') with zipfile.ZipFile(target_files, 'r') as target_files_zip: loaded_dict = common.LoadInfoDict(target_files_zip) self.assertEqual(3, loaded_dict['recovery_api_version']) self.assertEqual(2, loaded_dict['fstab_version']) self.assertIsNone(loaded_dict['fstab']) @test_utils.SkipIfExternalToolsUnavailable() def test_LoadInfoDict_missingMetaMiscInfoTxt(self): target_files = self._test_LoadInfoDict_createTargetFiles( self.INFO_DICT_DEFAULT, 'BOOT/RAMDISK/system/etc/recovery.fstab') common.ZipDelete(target_files, 'META/misc_info.txt') with zipfile.ZipFile(target_files, 'r') as target_files_zip: self.assertRaises(ValueError, common.LoadInfoDict, target_files_zip) @test_utils.SkipIfExternalToolsUnavailable() def test_LoadInfoDict_repacking(self): target_files = self._test_LoadInfoDict_createTargetFiles( self.INFO_DICT_DEFAULT, 'BOOT/RAMDISK/system/etc/recovery.fstab') unzipped = common.UnzipTemp(target_files) loaded_dict = common.LoadInfoDict(unzipped, True) self.assertEqual(3, loaded_dict['recovery_api_version']) self.assertEqual(2, loaded_dict['fstab_version']) self.assertIn('/', loaded_dict['fstab']) self.assertIn('/system', loaded_dict['fstab']) self.assertEqual( os.path.join(unzipped, 'ROOT'), loaded_dict['root_dir']) self.assertEqual( os.path.join(unzipped, 'META', 'root_filesystem_config.txt'), loaded_dict['root_fs_config']) def test_LoadInfoDict_repackingWithZipFileInput(self): target_files = self._test_LoadInfoDict_createTargetFiles( self.INFO_DICT_DEFAULT, 'BOOT/RAMDISK/system/etc/recovery.fstab') with zipfile.ZipFile(target_files, 'r') as target_files_zip: self.assertRaises( AssertionError, common.LoadInfoDict, target_files_zip, True) def test_MergeDynamicPartitionInfoDicts_ReturnsMergedDict(self): framework_dict = { 'super_partition_groups': 'group_a', 'dynamic_partition_list': 'system', 'super_group_a_list': 'system', } vendor_dict = { 'super_partition_groups': 'group_a group_b', 'dynamic_partition_list': 'vendor product', 'super_group_a_list': 'vendor', 'super_group_a_size': '1000', 'super_group_b_list': 'product', 'super_group_b_size': '2000', } merged_dict = common.MergeDynamicPartitionInfoDicts( framework_dict=framework_dict, vendor_dict=vendor_dict, size_prefix='super_', size_suffix='_size', list_prefix='super_', list_suffix='_list') expected_merged_dict = { 'super_partition_groups': 'group_a group_b', 'dynamic_partition_list': 'system vendor product', 'super_group_a_list': 'system vendor', 'super_group_a_size': '1000', 'super_group_b_list': 'product', 'super_group_b_size': '2000', } self.assertEqual(merged_dict, expected_merged_dict) def test_MergeDynamicPartitionInfoDicts_IgnoringFrameworkGroupSize(self): framework_dict = { 'super_partition_groups': 'group_a', 'dynamic_partition_list': 'system', 'super_group_a_list': 'system', 'super_group_a_size': '5000', } vendor_dict = { 'super_partition_groups': 'group_a group_b', 'dynamic_partition_list': 'vendor product', 'super_group_a_list': 'vendor', 'super_group_a_size': '1000', 'super_group_b_list': 'product', 'super_group_b_size': '2000', } merged_dict = common.MergeDynamicPartitionInfoDicts( framework_dict=framework_dict, vendor_dict=vendor_dict, size_prefix='super_', size_suffix='_size', list_prefix='super_', list_suffix='_list') expected_merged_dict = { 'super_partition_groups': 'group_a group_b', 'dynamic_partition_list': 'system vendor product', 'super_group_a_list': 'system vendor', 'super_group_a_size': '1000', 'super_group_b_list': 'product', 'super_group_b_size': '2000', } self.assertEqual(merged_dict, expected_merged_dict) def test_GetAvbPartitionArg(self): info_dict = {} cmd = common.GetAvbPartitionArg('system', '/path/to/system.img', info_dict) self.assertEqual( ['--include_descriptors_from_image', '/path/to/system.img'], cmd) @test_utils.SkipIfExternalToolsUnavailable() def test_AppendVBMetaArgsForPartition_vendorAsChainedPartition(self): testdata_dir = test_utils.get_testdata_dir() pubkey = os.path.join(testdata_dir, 'testkey.pubkey.pem') info_dict = { 'avb_avbtool': 'avbtool', 'avb_vendor_key_path': pubkey, 'avb_vendor_rollback_index_location': 5, } cmd = common.GetAvbPartitionArg('vendor', '/path/to/vendor.img', info_dict) self.assertEqual(2, len(cmd)) self.assertEqual('--chain_partition', cmd[0]) chained_partition_args = cmd[1].split(':') self.assertEqual(3, len(chained_partition_args)) self.assertEqual('vendor', chained_partition_args[0]) self.assertEqual('5', chained_partition_args[1]) self.assertTrue(os.path.exists(chained_partition_args[2])) class InstallRecoveryScriptFormatTest(test_utils.ReleaseToolsTestCase): """Checks the format of install-recovery.sh. Its format should match between common.py and validate_target_files.py. """ def setUp(self): self._tempdir = common.MakeTempDir() # Create a dummy dict that contains the fstab info for boot&recovery. self._info = {"fstab" : {}} dummy_fstab = [ "/dev/soc.0/by-name/boot /boot emmc defaults defaults", "/dev/soc.0/by-name/recovery /recovery emmc defaults defaults"] self._info["fstab"] = common.LoadRecoveryFSTab("\n".join, 2, dummy_fstab) # Construct the gzipped recovery.img and boot.img self.recovery_data = bytearray([ 0x1f, 0x8b, 0x08, 0x00, 0x81, 0x11, 0x02, 0x5a, 0x00, 0x03, 0x2b, 0x4a, 0x4d, 0xce, 0x2f, 0x4b, 0x2d, 0xaa, 0x04, 0x00, 0xc9, 0x93, 0x43, 0xf3, 0x08, 0x00, 0x00, 0x00 ]) # echo -n "boot" | gzip -f | hd self.boot_data = bytearray([ 0x1f, 0x8b, 0x08, 0x00, 0x8c, 0x12, 0x02, 0x5a, 0x00, 0x03, 0x4b, 0xca, 0xcf, 0x2f, 0x01, 0x00, 0xc4, 0xae, 0xed, 0x46, 0x04, 0x00, 0x00, 0x00 ]) def _out_tmp_sink(self, name, data, prefix="SYSTEM"): loc = os.path.join(self._tempdir, prefix, name) if not os.path.exists(os.path.dirname(loc)): os.makedirs(os.path.dirname(loc)) with open(loc, "wb") as f: f.write(data) def test_full_recovery(self): recovery_image = common.File("recovery.img", self.recovery_data) boot_image = common.File("boot.img", self.boot_data) self._info["full_recovery_image"] = "true" common.MakeRecoveryPatch(self._tempdir, self._out_tmp_sink, recovery_image, boot_image, self._info) validate_target_files.ValidateInstallRecoveryScript(self._tempdir, self._info) @test_utils.SkipIfExternalToolsUnavailable() def test_recovery_from_boot(self): recovery_image = common.File("recovery.img", self.recovery_data) self._out_tmp_sink("recovery.img", recovery_image.data, "IMAGES") boot_image = common.File("boot.img", self.boot_data) self._out_tmp_sink("boot.img", boot_image.data, "IMAGES") common.MakeRecoveryPatch(self._tempdir, self._out_tmp_sink, recovery_image, boot_image, self._info) validate_target_files.ValidateInstallRecoveryScript(self._tempdir, self._info) # Validate 'recovery-from-boot' with bonus argument. self._out_tmp_sink("etc/recovery-resource.dat", b"bonus", "SYSTEM") common.MakeRecoveryPatch(self._tempdir, self._out_tmp_sink, recovery_image, boot_image, self._info) validate_target_files.ValidateInstallRecoveryScript(self._tempdir, self._info) class MockScriptWriter(object): """A class that mocks edify_generator.EdifyGenerator.""" def __init__(self, enable_comments=False): self.lines = [] self.enable_comments = enable_comments def Comment(self, comment): if self.enable_comments: self.lines.append('# {}'.format(comment)) def AppendExtra(self, extra): self.lines.append(extra) def __str__(self): return '\n'.join(self.lines) class MockBlockDifference(object): def __init__(self, partition, tgt, src=None): self.partition = partition self.tgt = tgt self.src = src def WriteScript(self, script, _, progress=None, write_verify_script=False): if progress: script.AppendExtra("progress({})".format(progress)) script.AppendExtra("patch({});".format(self.partition)) if write_verify_script: self.WritePostInstallVerifyScript(script) def WritePostInstallVerifyScript(self, script): script.AppendExtra("verify({});".format(self.partition)) class FakeSparseImage(object): def __init__(self, size): self.blocksize = 4096 self.total_blocks = size // 4096 assert size % 4096 == 0, "{} is not a multiple of 4096".format(size) class DynamicPartitionsDifferenceTest(test_utils.ReleaseToolsTestCase): @staticmethod def get_op_list(output_path): with zipfile.ZipFile(output_path) as output_zip: with output_zip.open('dynamic_partitions_op_list') as op_list: return [line.decode().strip() for line in op_list.readlines() if not line.startswith(b'#')] def setUp(self): self.script = MockScriptWriter() self.output_path = common.MakeTempFile(suffix='.zip') def test_full(self): target_info = common.LoadDictionaryFromLines(""" dynamic_partition_list=system vendor super_partition_groups=group_foo super_group_foo_group_size={group_size} super_group_foo_partition_list=system vendor """.format(group_size=4 * GiB).split("\n")) block_diffs = [MockBlockDifference("system", FakeSparseImage(3 * GiB)), MockBlockDifference("vendor", FakeSparseImage(1 * GiB))] dp_diff = common.DynamicPartitionsDifference(target_info, block_diffs) with zipfile.ZipFile(self.output_path, 'w') as output_zip: dp_diff.WriteScript(self.script, output_zip, write_verify_script=True) self.assertEqual(str(self.script).strip(), """ assert(update_dynamic_partitions(package_extract_file("dynamic_partitions_op_list"))); patch(system); verify(system); unmap_partition("system"); patch(vendor); verify(vendor); unmap_partition("vendor"); """.strip()) lines = self.get_op_list(self.output_path) remove_all_groups = lines.index("remove_all_groups") add_group = lines.index("add_group group_foo 4294967296") add_vendor = lines.index("add vendor group_foo") add_system = lines.index("add system group_foo") resize_vendor = lines.index("resize vendor 1073741824") resize_system = lines.index("resize system 3221225472") self.assertLess(remove_all_groups, add_group, "Should add groups after removing all groups") self.assertLess(add_group, min(add_vendor, add_system), "Should add partitions after adding group") self.assertLess(add_system, resize_system, "Should resize system after adding it") self.assertLess(add_vendor, resize_vendor, "Should resize vendor after adding it") def test_inc_groups(self): source_info = common.LoadDictionaryFromLines(""" super_partition_groups=group_foo group_bar group_baz super_group_foo_group_size={group_foo_size} super_group_bar_group_size={group_bar_size} """.format(group_foo_size=4 * GiB, group_bar_size=3 * GiB).split("\n")) target_info = common.LoadDictionaryFromLines(""" super_partition_groups=group_foo group_baz group_qux super_group_foo_group_size={group_foo_size} super_group_baz_group_size={group_baz_size} super_group_qux_group_size={group_qux_size} """.format(group_foo_size=3 * GiB, group_baz_size=4 * GiB, group_qux_size=1 * GiB).split("\n")) dp_diff = common.DynamicPartitionsDifference(target_info, block_diffs=[], source_info_dict=source_info) with zipfile.ZipFile(self.output_path, 'w') as output_zip: dp_diff.WriteScript(self.script, output_zip, write_verify_script=True) lines = self.get_op_list(self.output_path) removed = lines.index("remove_group group_bar") shrunk = lines.index("resize_group group_foo 3221225472") grown = lines.index("resize_group group_baz 4294967296") added = lines.index("add_group group_qux 1073741824") self.assertLess(max(removed, shrunk), min(grown, added), "ops that remove / shrink partitions must precede ops that " "grow / add partitions") def test_incremental(self): source_info = common.LoadDictionaryFromLines(""" dynamic_partition_list=system vendor product system_ext super_partition_groups=group_foo super_group_foo_group_size={group_foo_size} super_group_foo_partition_list=system vendor product system_ext """.format(group_foo_size=4 * GiB).split("\n")) target_info = common.LoadDictionaryFromLines(""" dynamic_partition_list=system vendor product odm super_partition_groups=group_foo group_bar super_group_foo_group_size={group_foo_size} super_group_foo_partition_list=system vendor odm super_group_bar_group_size={group_bar_size} super_group_bar_partition_list=product """.format(group_foo_size=3 * GiB, group_bar_size=1 * GiB).split("\n")) block_diffs = [MockBlockDifference("system", FakeSparseImage(1536 * MiB), src=FakeSparseImage(1024 * MiB)), MockBlockDifference("vendor", FakeSparseImage(512 * MiB), src=FakeSparseImage(1024 * MiB)), MockBlockDifference("product", FakeSparseImage(1024 * MiB), src=FakeSparseImage(1024 * MiB)), MockBlockDifference("system_ext", None, src=FakeSparseImage(1024 * MiB)), MockBlockDifference("odm", FakeSparseImage(1024 * MiB), src=None)] dp_diff = common.DynamicPartitionsDifference(target_info, block_diffs, source_info_dict=source_info) with zipfile.ZipFile(self.output_path, 'w') as output_zip: dp_diff.WriteScript(self.script, output_zip, write_verify_script=True) metadata_idx = self.script.lines.index( 'assert(update_dynamic_partitions(package_extract_file(' '"dynamic_partitions_op_list")));') self.assertLess(self.script.lines.index('patch(vendor);'), metadata_idx) self.assertLess(metadata_idx, self.script.lines.index('verify(vendor);')) for p in ("product", "system", "odm"): patch_idx = self.script.lines.index("patch({});".format(p)) verify_idx = self.script.lines.index("verify({});".format(p)) self.assertLess(metadata_idx, patch_idx, "Should patch {} after updating metadata".format(p)) self.assertLess(patch_idx, verify_idx, "Should verify {} after patching".format(p)) self.assertNotIn("patch(system_ext);", self.script.lines) lines = self.get_op_list(self.output_path) remove = lines.index("remove system_ext") move_product_out = lines.index("move product default") shrink = lines.index("resize vendor 536870912") shrink_group = lines.index("resize_group group_foo 3221225472") add_group_bar = lines.index("add_group group_bar 1073741824") add_odm = lines.index("add odm group_foo") grow_existing = lines.index("resize system 1610612736") grow_added = lines.index("resize odm 1073741824") move_product_in = lines.index("move product group_bar") max_idx_move_partition_out_foo = max(remove, move_product_out, shrink) min_idx_move_partition_in_foo = min(add_odm, grow_existing, grow_added) self.assertLess(max_idx_move_partition_out_foo, shrink_group, "Must shrink group after partitions inside group are shrunk" " / removed") self.assertLess(add_group_bar, move_product_in, "Must add partitions to group after group is added") self.assertLess(max_idx_move_partition_out_foo, min_idx_move_partition_in_foo, "Must shrink partitions / remove partitions from group" "before adding / moving partitions into group") def test_remove_partition(self): source_info = common.LoadDictionaryFromLines(""" blockimgdiff_versions=3,4 use_dynamic_partitions=true dynamic_partition_list=foo super_partition_groups=group_foo super_group_foo_group_size={group_foo_size} super_group_foo_partition_list=foo """.format(group_foo_size=4 * GiB).split("\n")) target_info = common.LoadDictionaryFromLines(""" blockimgdiff_versions=3,4 use_dynamic_partitions=true super_partition_groups=group_foo super_group_foo_group_size={group_foo_size} """.format(group_foo_size=4 * GiB).split("\n")) common.OPTIONS.info_dict = target_info common.OPTIONS.target_info_dict = target_info common.OPTIONS.source_info_dict = source_info common.OPTIONS.cache_size = 4 * 4096 block_diffs = [common.BlockDifference("foo", EmptyImage(), src=DataImage("source", pad=True))] dp_diff = common.DynamicPartitionsDifference(target_info, block_diffs, source_info_dict=source_info) with zipfile.ZipFile(self.output_path, 'w') as output_zip: dp_diff.WriteScript(self.script, output_zip, write_verify_script=True) self.assertNotIn("block_image_update", str(self.script), "Removed partition should not be patched.") lines = self.get_op_list(self.output_path) self.assertEqual(lines, ["remove foo"])