834326ce7a
Both Extract...() functions don't need dynamic allocation for the writers, as those are strictly scoped. This CL changes heap allocation to stack allocation. Test: zip_archive_test Change-Id: Id727e4b9848235cd063cc67ecbe052d21ca21326
1233 lines
42 KiB
C++
1233 lines
42 KiB
C++
/*
|
|
* Copyright (C) 2008 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
/*
|
|
* Read-only access to Zip archives, with minimal heap allocation.
|
|
*/
|
|
|
|
#define LOG_TAG "ziparchive"
|
|
|
|
#include <assert.h>
|
|
#include <errno.h>
|
|
#include <fcntl.h>
|
|
#include <inttypes.h>
|
|
#include <limits.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include <time.h>
|
|
#include <unistd.h>
|
|
|
|
#include <memory>
|
|
#include <vector>
|
|
|
|
#include <android-base/file.h>
|
|
#include <android-base/logging.h>
|
|
#include <android-base/macros.h> // TEMP_FAILURE_RETRY may or may not be in unistd
|
|
#include <android-base/memory.h>
|
|
#include <log/log.h>
|
|
#include <utils/Compat.h>
|
|
#include <utils/FileMap.h>
|
|
#include "ziparchive/zip_archive.h"
|
|
#include "zlib.h"
|
|
|
|
#include "entry_name_utils-inl.h"
|
|
#include "zip_archive_common.h"
|
|
#include "zip_archive_private.h"
|
|
|
|
using android::base::get_unaligned;
|
|
|
|
// Used to turn on crc checks - verify that the content CRC matches the values
|
|
// specified in the local file header and the central directory.
|
|
static const bool kCrcChecksEnabled = false;
|
|
|
|
// This is for windows. If we don't open a file in binary mode, weird
|
|
// things will happen.
|
|
#ifndef O_BINARY
|
|
#define O_BINARY 0
|
|
#endif
|
|
|
|
// The maximum number of bytes to scan backwards for the EOCD start.
|
|
static const uint32_t kMaxEOCDSearch = kMaxCommentLen + sizeof(EocdRecord);
|
|
|
|
/*
|
|
* A Read-only Zip archive.
|
|
*
|
|
* We want "open" and "find entry by name" to be fast operations, and
|
|
* we want to use as little memory as possible. We memory-map the zip
|
|
* central directory, and load a hash table with pointers to the filenames
|
|
* (which aren't null-terminated). The other fields are at a fixed offset
|
|
* from the filename, so we don't need to extract those (but we do need
|
|
* to byte-read and endian-swap them every time we want them).
|
|
*
|
|
* It's possible that somebody has handed us a massive (~1GB) zip archive,
|
|
* so we can't expect to mmap the entire file.
|
|
*
|
|
* To speed comparisons when doing a lookup by name, we could make the mapping
|
|
* "private" (copy-on-write) and null-terminate the filenames after verifying
|
|
* the record structure. However, this requires a private mapping of
|
|
* every page that the Central Directory touches. Easier to tuck a copy
|
|
* of the string length into the hash table entry.
|
|
*/
|
|
|
|
/*
|
|
* Round up to the next highest power of 2.
|
|
*
|
|
* Found on http://graphics.stanford.edu/~seander/bithacks.html.
|
|
*/
|
|
static uint32_t RoundUpPower2(uint32_t val) {
|
|
val--;
|
|
val |= val >> 1;
|
|
val |= val >> 2;
|
|
val |= val >> 4;
|
|
val |= val >> 8;
|
|
val |= val >> 16;
|
|
val++;
|
|
|
|
return val;
|
|
}
|
|
|
|
static uint32_t ComputeHash(const ZipString& name) {
|
|
#if !defined(_WIN32)
|
|
return std::hash<std::string_view>{}(
|
|
std::string_view(reinterpret_cast<const char*>(name.name), name.name_length));
|
|
#else
|
|
// Remove this code path once the windows compiler knows how to compile the above statement.
|
|
uint32_t hash = 0;
|
|
uint16_t len = name.name_length;
|
|
const uint8_t* str = name.name;
|
|
|
|
while (len--) {
|
|
hash = hash * 31 + *str++;
|
|
}
|
|
|
|
return hash;
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* Convert a ZipEntry to a hash table index, verifying that it's in a
|
|
* valid range.
|
|
*/
|
|
static int64_t EntryToIndex(const ZipString* hash_table, const uint32_t hash_table_size,
|
|
const ZipString& name) {
|
|
const uint32_t hash = ComputeHash(name);
|
|
|
|
// NOTE: (hash_table_size - 1) is guaranteed to be non-negative.
|
|
uint32_t ent = hash & (hash_table_size - 1);
|
|
while (hash_table[ent].name != NULL) {
|
|
if (hash_table[ent] == name) {
|
|
return ent;
|
|
}
|
|
|
|
ent = (ent + 1) & (hash_table_size - 1);
|
|
}
|
|
|
|
ALOGV("Zip: Unable to find entry %.*s", name.name_length, name.name);
|
|
return kEntryNotFound;
|
|
}
|
|
|
|
/*
|
|
* Add a new entry to the hash table.
|
|
*/
|
|
static int32_t AddToHash(ZipString* hash_table, const uint64_t hash_table_size,
|
|
const ZipString& name) {
|
|
const uint64_t hash = ComputeHash(name);
|
|
uint32_t ent = hash & (hash_table_size - 1);
|
|
|
|
/*
|
|
* We over-allocated the table, so we're guaranteed to find an empty slot.
|
|
* Further, we guarantee that the hashtable size is not 0.
|
|
*/
|
|
while (hash_table[ent].name != NULL) {
|
|
if (hash_table[ent] == name) {
|
|
// We've found a duplicate entry. We don't accept it
|
|
ALOGW("Zip: Found duplicate entry %.*s", name.name_length, name.name);
|
|
return kDuplicateEntry;
|
|
}
|
|
ent = (ent + 1) & (hash_table_size - 1);
|
|
}
|
|
|
|
hash_table[ent].name = name.name;
|
|
hash_table[ent].name_length = name.name_length;
|
|
return 0;
|
|
}
|
|
|
|
static int32_t MapCentralDirectory0(const char* debug_file_name, ZipArchive* archive,
|
|
off64_t file_length, off64_t read_amount, uint8_t* scan_buffer) {
|
|
const off64_t search_start = file_length - read_amount;
|
|
|
|
if (!archive->mapped_zip.ReadAtOffset(scan_buffer, read_amount, search_start)) {
|
|
ALOGE("Zip: read %" PRId64 " from offset %" PRId64 " failed", static_cast<int64_t>(read_amount),
|
|
static_cast<int64_t>(search_start));
|
|
return kIoError;
|
|
}
|
|
|
|
/*
|
|
* Scan backward for the EOCD magic. In an archive without a trailing
|
|
* comment, we'll find it on the first try. (We may want to consider
|
|
* doing an initial minimal read; if we don't find it, retry with a
|
|
* second read as above.)
|
|
*/
|
|
int i = read_amount - sizeof(EocdRecord);
|
|
for (; i >= 0; i--) {
|
|
if (scan_buffer[i] == 0x50) {
|
|
uint32_t* sig_addr = reinterpret_cast<uint32_t*>(&scan_buffer[i]);
|
|
if (get_unaligned<uint32_t>(sig_addr) == EocdRecord::kSignature) {
|
|
ALOGV("+++ Found EOCD at buf+%d", i);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (i < 0) {
|
|
ALOGD("Zip: EOCD not found, %s is not zip", debug_file_name);
|
|
return kInvalidFile;
|
|
}
|
|
|
|
const off64_t eocd_offset = search_start + i;
|
|
const EocdRecord* eocd = reinterpret_cast<const EocdRecord*>(scan_buffer + i);
|
|
/*
|
|
* Verify that there's no trailing space at the end of the central directory
|
|
* and its comment.
|
|
*/
|
|
const off64_t calculated_length = eocd_offset + sizeof(EocdRecord) + eocd->comment_length;
|
|
if (calculated_length != file_length) {
|
|
ALOGW("Zip: %" PRId64 " extraneous bytes at the end of the central directory",
|
|
static_cast<int64_t>(file_length - calculated_length));
|
|
return kInvalidFile;
|
|
}
|
|
|
|
/*
|
|
* Grab the CD offset and size, and the number of entries in the
|
|
* archive and verify that they look reasonable.
|
|
*/
|
|
if (static_cast<off64_t>(eocd->cd_start_offset) + eocd->cd_size > eocd_offset) {
|
|
ALOGW("Zip: bad offsets (dir %" PRIu32 ", size %" PRIu32 ", eocd %" PRId64 ")",
|
|
eocd->cd_start_offset, eocd->cd_size, static_cast<int64_t>(eocd_offset));
|
|
#if defined(__ANDROID__)
|
|
if (eocd->cd_start_offset + eocd->cd_size <= eocd_offset) {
|
|
android_errorWriteLog(0x534e4554, "31251826");
|
|
}
|
|
#endif
|
|
return kInvalidOffset;
|
|
}
|
|
if (eocd->num_records == 0) {
|
|
ALOGW("Zip: empty archive?");
|
|
return kEmptyArchive;
|
|
}
|
|
|
|
ALOGV("+++ num_entries=%" PRIu32 " dir_size=%" PRIu32 " dir_offset=%" PRIu32, eocd->num_records,
|
|
eocd->cd_size, eocd->cd_start_offset);
|
|
|
|
/*
|
|
* It all looks good. Create a mapping for the CD, and set the fields
|
|
* in archive.
|
|
*/
|
|
|
|
if (!archive->InitializeCentralDirectory(debug_file_name,
|
|
static_cast<off64_t>(eocd->cd_start_offset),
|
|
static_cast<size_t>(eocd->cd_size))) {
|
|
ALOGE("Zip: failed to intialize central directory.\n");
|
|
return kMmapFailed;
|
|
}
|
|
|
|
archive->num_entries = eocd->num_records;
|
|
archive->directory_offset = eocd->cd_start_offset;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Find the zip Central Directory and memory-map it.
|
|
*
|
|
* On success, returns 0 after populating fields from the EOCD area:
|
|
* directory_offset
|
|
* directory_ptr
|
|
* num_entries
|
|
*/
|
|
static int32_t MapCentralDirectory(const char* debug_file_name, ZipArchive* archive) {
|
|
// Test file length. We use lseek64 to make sure the file
|
|
// is small enough to be a zip file (Its size must be less than
|
|
// 0xffffffff bytes).
|
|
off64_t file_length = archive->mapped_zip.GetFileLength();
|
|
if (file_length == -1) {
|
|
return kInvalidFile;
|
|
}
|
|
|
|
if (file_length > static_cast<off64_t>(0xffffffff)) {
|
|
ALOGV("Zip: zip file too long %" PRId64, static_cast<int64_t>(file_length));
|
|
return kInvalidFile;
|
|
}
|
|
|
|
if (file_length < static_cast<off64_t>(sizeof(EocdRecord))) {
|
|
ALOGV("Zip: length %" PRId64 " is too small to be zip", static_cast<int64_t>(file_length));
|
|
return kInvalidFile;
|
|
}
|
|
|
|
/*
|
|
* Perform the traditional EOCD snipe hunt.
|
|
*
|
|
* We're searching for the End of Central Directory magic number,
|
|
* which appears at the start of the EOCD block. It's followed by
|
|
* 18 bytes of EOCD stuff and up to 64KB of archive comment. We
|
|
* need to read the last part of the file into a buffer, dig through
|
|
* it to find the magic number, parse some values out, and use those
|
|
* to determine the extent of the CD.
|
|
*
|
|
* We start by pulling in the last part of the file.
|
|
*/
|
|
off64_t read_amount = kMaxEOCDSearch;
|
|
if (file_length < read_amount) {
|
|
read_amount = file_length;
|
|
}
|
|
|
|
std::vector<uint8_t> scan_buffer(read_amount);
|
|
int32_t result =
|
|
MapCentralDirectory0(debug_file_name, archive, file_length, read_amount, scan_buffer.data());
|
|
return result;
|
|
}
|
|
|
|
/*
|
|
* Parses the Zip archive's Central Directory. Allocates and populates the
|
|
* hash table.
|
|
*
|
|
* Returns 0 on success.
|
|
*/
|
|
static int32_t ParseZipArchive(ZipArchive* archive) {
|
|
const uint8_t* const cd_ptr = archive->central_directory.GetBasePtr();
|
|
const size_t cd_length = archive->central_directory.GetMapLength();
|
|
const uint16_t num_entries = archive->num_entries;
|
|
|
|
/*
|
|
* Create hash table. We have a minimum 75% load factor, possibly as
|
|
* low as 50% after we round off to a power of 2. There must be at
|
|
* least one unused entry to avoid an infinite loop during creation.
|
|
*/
|
|
archive->hash_table_size = RoundUpPower2(1 + (num_entries * 4) / 3);
|
|
archive->hash_table =
|
|
reinterpret_cast<ZipString*>(calloc(archive->hash_table_size, sizeof(ZipString)));
|
|
if (archive->hash_table == nullptr) {
|
|
ALOGW("Zip: unable to allocate the %u-entry hash_table, entry size: %zu",
|
|
archive->hash_table_size, sizeof(ZipString));
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* Walk through the central directory, adding entries to the hash
|
|
* table and verifying values.
|
|
*/
|
|
const uint8_t* const cd_end = cd_ptr + cd_length;
|
|
const uint8_t* ptr = cd_ptr;
|
|
for (uint16_t i = 0; i < num_entries; i++) {
|
|
if (ptr > cd_end - sizeof(CentralDirectoryRecord)) {
|
|
ALOGW("Zip: ran off the end (at %" PRIu16 ")", i);
|
|
#if defined(__ANDROID__)
|
|
android_errorWriteLog(0x534e4554, "36392138");
|
|
#endif
|
|
return -1;
|
|
}
|
|
|
|
const CentralDirectoryRecord* cdr = reinterpret_cast<const CentralDirectoryRecord*>(ptr);
|
|
if (cdr->record_signature != CentralDirectoryRecord::kSignature) {
|
|
ALOGW("Zip: missed a central dir sig (at %" PRIu16 ")", i);
|
|
return -1;
|
|
}
|
|
|
|
const off64_t local_header_offset = cdr->local_file_header_offset;
|
|
if (local_header_offset >= archive->directory_offset) {
|
|
ALOGW("Zip: bad LFH offset %" PRId64 " at entry %" PRIu16,
|
|
static_cast<int64_t>(local_header_offset), i);
|
|
return -1;
|
|
}
|
|
|
|
const uint16_t file_name_length = cdr->file_name_length;
|
|
const uint16_t extra_length = cdr->extra_field_length;
|
|
const uint16_t comment_length = cdr->comment_length;
|
|
const uint8_t* file_name = ptr + sizeof(CentralDirectoryRecord);
|
|
|
|
if (file_name + file_name_length > cd_end) {
|
|
ALOGW(
|
|
"Zip: file name boundary exceeds the central directory range, file_name_length: "
|
|
"%" PRIx16 ", cd_length: %zu",
|
|
file_name_length, cd_length);
|
|
return -1;
|
|
}
|
|
/* check that file name is valid UTF-8 and doesn't contain NUL (U+0000) characters */
|
|
if (!IsValidEntryName(file_name, file_name_length)) {
|
|
return -1;
|
|
}
|
|
|
|
/* add the CDE filename to the hash table */
|
|
ZipString entry_name;
|
|
entry_name.name = file_name;
|
|
entry_name.name_length = file_name_length;
|
|
const int add_result = AddToHash(archive->hash_table, archive->hash_table_size, entry_name);
|
|
if (add_result != 0) {
|
|
ALOGW("Zip: Error adding entry to hash table %d", add_result);
|
|
return add_result;
|
|
}
|
|
|
|
ptr += sizeof(CentralDirectoryRecord) + file_name_length + extra_length + comment_length;
|
|
if ((ptr - cd_ptr) > static_cast<int64_t>(cd_length)) {
|
|
ALOGW("Zip: bad CD advance (%tu vs %zu) at entry %" PRIu16, ptr - cd_ptr, cd_length, i);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
uint32_t lfh_start_bytes;
|
|
if (!archive->mapped_zip.ReadAtOffset(reinterpret_cast<uint8_t*>(&lfh_start_bytes),
|
|
sizeof(uint32_t), 0)) {
|
|
ALOGW("Zip: Unable to read header for entry at offset == 0.");
|
|
return -1;
|
|
}
|
|
|
|
if (lfh_start_bytes != LocalFileHeader::kSignature) {
|
|
ALOGW("Zip: Entry at offset zero has invalid LFH signature %" PRIx32, lfh_start_bytes);
|
|
#if defined(__ANDROID__)
|
|
android_errorWriteLog(0x534e4554, "64211847");
|
|
#endif
|
|
return -1;
|
|
}
|
|
|
|
ALOGV("+++ zip good scan %" PRIu16 " entries", num_entries);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int32_t OpenArchiveInternal(ZipArchive* archive, const char* debug_file_name) {
|
|
int32_t result = -1;
|
|
if ((result = MapCentralDirectory(debug_file_name, archive)) != 0) {
|
|
return result;
|
|
}
|
|
|
|
if ((result = ParseZipArchive(archive))) {
|
|
return result;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int32_t OpenArchiveFd(int fd, const char* debug_file_name, ZipArchiveHandle* handle,
|
|
bool assume_ownership) {
|
|
ZipArchive* archive = new ZipArchive(fd, assume_ownership);
|
|
*handle = archive;
|
|
return OpenArchiveInternal(archive, debug_file_name);
|
|
}
|
|
|
|
int32_t OpenArchive(const char* fileName, ZipArchiveHandle* handle) {
|
|
const int fd = open(fileName, O_RDONLY | O_BINARY, 0);
|
|
ZipArchive* archive = new ZipArchive(fd, true);
|
|
*handle = archive;
|
|
|
|
if (fd < 0) {
|
|
ALOGW("Unable to open '%s': %s", fileName, strerror(errno));
|
|
return kIoError;
|
|
}
|
|
|
|
return OpenArchiveInternal(archive, fileName);
|
|
}
|
|
|
|
int32_t OpenArchiveFromMemory(void* address, size_t length, const char* debug_file_name,
|
|
ZipArchiveHandle* handle) {
|
|
ZipArchive* archive = new ZipArchive(address, length);
|
|
*handle = archive;
|
|
return OpenArchiveInternal(archive, debug_file_name);
|
|
}
|
|
|
|
/*
|
|
* Close a ZipArchive, closing the file and freeing the contents.
|
|
*/
|
|
void CloseArchive(ZipArchiveHandle handle) {
|
|
ZipArchive* archive = reinterpret_cast<ZipArchive*>(handle);
|
|
ALOGV("Closing archive %p", archive);
|
|
delete archive;
|
|
}
|
|
|
|
static int32_t ValidateDataDescriptor(MappedZipFile& mapped_zip, ZipEntry* entry) {
|
|
uint8_t ddBuf[sizeof(DataDescriptor) + sizeof(DataDescriptor::kOptSignature)];
|
|
off64_t offset = entry->offset;
|
|
if (entry->method != kCompressStored) {
|
|
offset += entry->compressed_length;
|
|
} else {
|
|
offset += entry->uncompressed_length;
|
|
}
|
|
|
|
if (!mapped_zip.ReadAtOffset(ddBuf, sizeof(ddBuf), offset)) {
|
|
return kIoError;
|
|
}
|
|
|
|
const uint32_t ddSignature = *(reinterpret_cast<const uint32_t*>(ddBuf));
|
|
const uint16_t ddOffset = (ddSignature == DataDescriptor::kOptSignature) ? 4 : 0;
|
|
const DataDescriptor* descriptor = reinterpret_cast<const DataDescriptor*>(ddBuf + ddOffset);
|
|
|
|
// Validate that the values in the data descriptor match those in the central
|
|
// directory.
|
|
if (entry->compressed_length != descriptor->compressed_size ||
|
|
entry->uncompressed_length != descriptor->uncompressed_size ||
|
|
entry->crc32 != descriptor->crc32) {
|
|
ALOGW("Zip: size/crc32 mismatch. expected {%" PRIu32 ", %" PRIu32 ", %" PRIx32
|
|
"}, was {%" PRIu32 ", %" PRIu32 ", %" PRIx32 "}",
|
|
entry->compressed_length, entry->uncompressed_length, entry->crc32,
|
|
descriptor->compressed_size, descriptor->uncompressed_size, descriptor->crc32);
|
|
return kInconsistentInformation;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int32_t FindEntry(const ZipArchive* archive, const int ent, ZipEntry* data) {
|
|
const uint16_t nameLen = archive->hash_table[ent].name_length;
|
|
|
|
// Recover the start of the central directory entry from the filename
|
|
// pointer. The filename is the first entry past the fixed-size data,
|
|
// so we can just subtract back from that.
|
|
const uint8_t* ptr = archive->hash_table[ent].name;
|
|
ptr -= sizeof(CentralDirectoryRecord);
|
|
|
|
// This is the base of our mmapped region, we have to sanity check that
|
|
// the name that's in the hash table is a pointer to a location within
|
|
// this mapped region.
|
|
const uint8_t* base_ptr = archive->central_directory.GetBasePtr();
|
|
if (ptr < base_ptr || ptr > base_ptr + archive->central_directory.GetMapLength()) {
|
|
ALOGW("Zip: Invalid entry pointer");
|
|
return kInvalidOffset;
|
|
}
|
|
|
|
const CentralDirectoryRecord* cdr = reinterpret_cast<const CentralDirectoryRecord*>(ptr);
|
|
|
|
// The offset of the start of the central directory in the zipfile.
|
|
// We keep this lying around so that we can sanity check all our lengths
|
|
// and our per-file structures.
|
|
const off64_t cd_offset = archive->directory_offset;
|
|
|
|
// Fill out the compression method, modification time, crc32
|
|
// and other interesting attributes from the central directory. These
|
|
// will later be compared against values from the local file header.
|
|
data->method = cdr->compression_method;
|
|
data->mod_time = cdr->last_mod_date << 16 | cdr->last_mod_time;
|
|
data->crc32 = cdr->crc32;
|
|
data->compressed_length = cdr->compressed_size;
|
|
data->uncompressed_length = cdr->uncompressed_size;
|
|
|
|
// Figure out the local header offset from the central directory. The
|
|
// actual file data will begin after the local header and the name /
|
|
// extra comments.
|
|
const off64_t local_header_offset = cdr->local_file_header_offset;
|
|
if (local_header_offset + static_cast<off64_t>(sizeof(LocalFileHeader)) >= cd_offset) {
|
|
ALOGW("Zip: bad local hdr offset in zip");
|
|
return kInvalidOffset;
|
|
}
|
|
|
|
uint8_t lfh_buf[sizeof(LocalFileHeader)];
|
|
if (!archive->mapped_zip.ReadAtOffset(lfh_buf, sizeof(lfh_buf), local_header_offset)) {
|
|
ALOGW("Zip: failed reading lfh name from offset %" PRId64,
|
|
static_cast<int64_t>(local_header_offset));
|
|
return kIoError;
|
|
}
|
|
|
|
const LocalFileHeader* lfh = reinterpret_cast<const LocalFileHeader*>(lfh_buf);
|
|
|
|
if (lfh->lfh_signature != LocalFileHeader::kSignature) {
|
|
ALOGW("Zip: didn't find signature at start of lfh, offset=%" PRId64,
|
|
static_cast<int64_t>(local_header_offset));
|
|
return kInvalidOffset;
|
|
}
|
|
|
|
// Paranoia: Match the values specified in the local file header
|
|
// to those specified in the central directory.
|
|
|
|
// Warn if central directory and local file header don't agree on the use
|
|
// of a trailing Data Descriptor. The reference implementation is inconsistent
|
|
// and appears to use the LFH value during extraction (unzip) but the CD value
|
|
// while displayng information about archives (zipinfo). The spec remains
|
|
// silent on this inconsistency as well.
|
|
//
|
|
// For now, always use the version from the LFH but make sure that the values
|
|
// specified in the central directory match those in the data descriptor.
|
|
//
|
|
// NOTE: It's also worth noting that unzip *does* warn about inconsistencies in
|
|
// bit 11 (EFS: The language encoding flag, marking that filename and comment are
|
|
// encoded using UTF-8). This implementation does not check for the presence of
|
|
// that flag and always enforces that entry names are valid UTF-8.
|
|
if ((lfh->gpb_flags & kGPBDDFlagMask) != (cdr->gpb_flags & kGPBDDFlagMask)) {
|
|
ALOGW("Zip: gpb flag mismatch at bit 3. expected {%04" PRIx16 "}, was {%04" PRIx16 "}",
|
|
cdr->gpb_flags, lfh->gpb_flags);
|
|
}
|
|
|
|
// If there is no trailing data descriptor, verify that the central directory and local file
|
|
// header agree on the crc, compressed, and uncompressed sizes of the entry.
|
|
if ((lfh->gpb_flags & kGPBDDFlagMask) == 0) {
|
|
data->has_data_descriptor = 0;
|
|
if (data->compressed_length != lfh->compressed_size ||
|
|
data->uncompressed_length != lfh->uncompressed_size || data->crc32 != lfh->crc32) {
|
|
ALOGW("Zip: size/crc32 mismatch. expected {%" PRIu32 ", %" PRIu32 ", %" PRIx32
|
|
"}, was {%" PRIu32 ", %" PRIu32 ", %" PRIx32 "}",
|
|
data->compressed_length, data->uncompressed_length, data->crc32, lfh->compressed_size,
|
|
lfh->uncompressed_size, lfh->crc32);
|
|
return kInconsistentInformation;
|
|
}
|
|
} else {
|
|
data->has_data_descriptor = 1;
|
|
}
|
|
|
|
// 4.4.2.1: the upper byte of `version_made_by` gives the source OS. Unix is 3.
|
|
if ((cdr->version_made_by >> 8) == 3) {
|
|
data->unix_mode = (cdr->external_file_attributes >> 16) & 0xffff;
|
|
} else {
|
|
data->unix_mode = 0777;
|
|
}
|
|
|
|
// Check that the local file header name matches the declared
|
|
// name in the central directory.
|
|
if (lfh->file_name_length == nameLen) {
|
|
const off64_t name_offset = local_header_offset + sizeof(LocalFileHeader);
|
|
if (name_offset + lfh->file_name_length > cd_offset) {
|
|
ALOGW("Zip: Invalid declared length");
|
|
return kInvalidOffset;
|
|
}
|
|
|
|
std::vector<uint8_t> name_buf(nameLen);
|
|
if (!archive->mapped_zip.ReadAtOffset(name_buf.data(), nameLen, name_offset)) {
|
|
ALOGW("Zip: failed reading lfh name from offset %" PRId64, static_cast<int64_t>(name_offset));
|
|
return kIoError;
|
|
}
|
|
|
|
if (memcmp(archive->hash_table[ent].name, name_buf.data(), nameLen)) {
|
|
return kInconsistentInformation;
|
|
}
|
|
|
|
} else {
|
|
ALOGW("Zip: lfh name did not match central directory.");
|
|
return kInconsistentInformation;
|
|
}
|
|
|
|
const off64_t data_offset = local_header_offset + sizeof(LocalFileHeader) +
|
|
lfh->file_name_length + lfh->extra_field_length;
|
|
if (data_offset > cd_offset) {
|
|
ALOGW("Zip: bad data offset %" PRId64 " in zip", static_cast<int64_t>(data_offset));
|
|
return kInvalidOffset;
|
|
}
|
|
|
|
if (static_cast<off64_t>(data_offset + data->compressed_length) > cd_offset) {
|
|
ALOGW("Zip: bad compressed length in zip (%" PRId64 " + %" PRIu32 " > %" PRId64 ")",
|
|
static_cast<int64_t>(data_offset), data->compressed_length,
|
|
static_cast<int64_t>(cd_offset));
|
|
return kInvalidOffset;
|
|
}
|
|
|
|
if (data->method == kCompressStored &&
|
|
static_cast<off64_t>(data_offset + data->uncompressed_length) > cd_offset) {
|
|
ALOGW("Zip: bad uncompressed length in zip (%" PRId64 " + %" PRIu32 " > %" PRId64 ")",
|
|
static_cast<int64_t>(data_offset), data->uncompressed_length,
|
|
static_cast<int64_t>(cd_offset));
|
|
return kInvalidOffset;
|
|
}
|
|
|
|
data->offset = data_offset;
|
|
return 0;
|
|
}
|
|
|
|
struct IterationHandle {
|
|
uint32_t position;
|
|
// We're not using vector here because this code is used in the Windows SDK
|
|
// where the STL is not available.
|
|
ZipString prefix;
|
|
ZipString suffix;
|
|
ZipArchive* archive;
|
|
|
|
IterationHandle(const ZipString* in_prefix, const ZipString* in_suffix) {
|
|
if (in_prefix) {
|
|
uint8_t* name_copy = new uint8_t[in_prefix->name_length];
|
|
memcpy(name_copy, in_prefix->name, in_prefix->name_length);
|
|
prefix.name = name_copy;
|
|
prefix.name_length = in_prefix->name_length;
|
|
} else {
|
|
prefix.name = NULL;
|
|
prefix.name_length = 0;
|
|
}
|
|
if (in_suffix) {
|
|
uint8_t* name_copy = new uint8_t[in_suffix->name_length];
|
|
memcpy(name_copy, in_suffix->name, in_suffix->name_length);
|
|
suffix.name = name_copy;
|
|
suffix.name_length = in_suffix->name_length;
|
|
} else {
|
|
suffix.name = NULL;
|
|
suffix.name_length = 0;
|
|
}
|
|
}
|
|
|
|
~IterationHandle() {
|
|
delete[] prefix.name;
|
|
delete[] suffix.name;
|
|
}
|
|
};
|
|
|
|
int32_t StartIteration(ZipArchiveHandle handle, void** cookie_ptr, const ZipString* optional_prefix,
|
|
const ZipString* optional_suffix) {
|
|
ZipArchive* archive = reinterpret_cast<ZipArchive*>(handle);
|
|
|
|
if (archive == NULL || archive->hash_table == NULL) {
|
|
ALOGW("Zip: Invalid ZipArchiveHandle");
|
|
return kInvalidHandle;
|
|
}
|
|
|
|
IterationHandle* cookie = new IterationHandle(optional_prefix, optional_suffix);
|
|
cookie->position = 0;
|
|
cookie->archive = archive;
|
|
|
|
*cookie_ptr = cookie;
|
|
return 0;
|
|
}
|
|
|
|
void EndIteration(void* cookie) {
|
|
delete reinterpret_cast<IterationHandle*>(cookie);
|
|
}
|
|
|
|
int32_t FindEntry(const ZipArchiveHandle handle, const ZipString& entryName, ZipEntry* data) {
|
|
const ZipArchive* archive = reinterpret_cast<ZipArchive*>(handle);
|
|
if (entryName.name_length == 0) {
|
|
ALOGW("Zip: Invalid filename %.*s", entryName.name_length, entryName.name);
|
|
return kInvalidEntryName;
|
|
}
|
|
|
|
const int64_t ent = EntryToIndex(archive->hash_table, archive->hash_table_size, entryName);
|
|
|
|
if (ent < 0) {
|
|
ALOGV("Zip: Could not find entry %.*s", entryName.name_length, entryName.name);
|
|
return ent;
|
|
}
|
|
|
|
return FindEntry(archive, ent, data);
|
|
}
|
|
|
|
int32_t Next(void* cookie, ZipEntry* data, ZipString* name) {
|
|
IterationHandle* handle = reinterpret_cast<IterationHandle*>(cookie);
|
|
if (handle == NULL) {
|
|
return kInvalidHandle;
|
|
}
|
|
|
|
ZipArchive* archive = handle->archive;
|
|
if (archive == NULL || archive->hash_table == NULL) {
|
|
ALOGW("Zip: Invalid ZipArchiveHandle");
|
|
return kInvalidHandle;
|
|
}
|
|
|
|
const uint32_t currentOffset = handle->position;
|
|
const uint32_t hash_table_length = archive->hash_table_size;
|
|
const ZipString* hash_table = archive->hash_table;
|
|
|
|
for (uint32_t i = currentOffset; i < hash_table_length; ++i) {
|
|
if (hash_table[i].name != NULL &&
|
|
(handle->prefix.name_length == 0 || hash_table[i].StartsWith(handle->prefix)) &&
|
|
(handle->suffix.name_length == 0 || hash_table[i].EndsWith(handle->suffix))) {
|
|
handle->position = (i + 1);
|
|
const int error = FindEntry(archive, i, data);
|
|
if (!error) {
|
|
name->name = hash_table[i].name;
|
|
name->name_length = hash_table[i].name_length;
|
|
}
|
|
|
|
return error;
|
|
}
|
|
}
|
|
|
|
handle->position = 0;
|
|
return kIterationEnd;
|
|
}
|
|
|
|
// A Writer that writes data to a fixed size memory region.
|
|
// The size of the memory region must be equal to the total size of
|
|
// the data appended to it.
|
|
class MemoryWriter : public zip_archive::Writer {
|
|
public:
|
|
MemoryWriter(uint8_t* buf, size_t size) : Writer(), buf_(buf), size_(size), bytes_written_(0) {}
|
|
|
|
virtual bool Append(uint8_t* buf, size_t buf_size) override {
|
|
if (bytes_written_ + buf_size > size_) {
|
|
ALOGW("Zip: Unexpected size " ZD " (declared) vs " ZD " (actual)", size_,
|
|
bytes_written_ + buf_size);
|
|
return false;
|
|
}
|
|
|
|
memcpy(buf_ + bytes_written_, buf, buf_size);
|
|
bytes_written_ += buf_size;
|
|
return true;
|
|
}
|
|
|
|
private:
|
|
uint8_t* const buf_;
|
|
const size_t size_;
|
|
size_t bytes_written_;
|
|
};
|
|
|
|
// A Writer that appends data to a file |fd| at its current position.
|
|
// The file will be truncated to the end of the written data.
|
|
class FileWriter : public zip_archive::Writer {
|
|
public:
|
|
// Creates a FileWriter for |fd| and prepare to write |entry| to it,
|
|
// guaranteeing that the file descriptor is valid and that there's enough
|
|
// space on the volume to write out the entry completely and that the file
|
|
// is truncated to the correct length (no truncation if |fd| references a
|
|
// block device).
|
|
//
|
|
// Returns a valid FileWriter on success, |nullptr| if an error occurred.
|
|
static FileWriter Create(int fd, const ZipEntry* entry) {
|
|
const uint32_t declared_length = entry->uncompressed_length;
|
|
const off64_t current_offset = lseek64(fd, 0, SEEK_CUR);
|
|
if (current_offset == -1) {
|
|
ALOGW("Zip: unable to seek to current location on fd %d: %s", fd, strerror(errno));
|
|
return FileWriter{};
|
|
}
|
|
|
|
int result = 0;
|
|
#if defined(__linux__)
|
|
if (declared_length > 0) {
|
|
// Make sure we have enough space on the volume to extract the compressed
|
|
// entry. Note that the call to ftruncate below will change the file size but
|
|
// will not allocate space on disk and this call to fallocate will not
|
|
// change the file size.
|
|
// Note: fallocate is only supported by the following filesystems -
|
|
// btrfs, ext4, ocfs2, and xfs. Therefore fallocate might fail with
|
|
// EOPNOTSUPP error when issued in other filesystems.
|
|
// Hence, check for the return error code before concluding that the
|
|
// disk does not have enough space.
|
|
result = TEMP_FAILURE_RETRY(fallocate(fd, 0, current_offset, declared_length));
|
|
if (result == -1 && errno == ENOSPC) {
|
|
ALOGW("Zip: unable to allocate %" PRId64 " bytes at offset %" PRId64 ": %s",
|
|
static_cast<int64_t>(declared_length), static_cast<int64_t>(current_offset),
|
|
strerror(errno));
|
|
return FileWriter{};
|
|
}
|
|
}
|
|
#endif // __linux__
|
|
|
|
struct stat sb;
|
|
if (fstat(fd, &sb) == -1) {
|
|
ALOGW("Zip: unable to fstat file: %s", strerror(errno));
|
|
return FileWriter{};
|
|
}
|
|
|
|
// Block device doesn't support ftruncate(2).
|
|
if (!S_ISBLK(sb.st_mode)) {
|
|
result = TEMP_FAILURE_RETRY(ftruncate(fd, declared_length + current_offset));
|
|
if (result == -1) {
|
|
ALOGW("Zip: unable to truncate file to %" PRId64 ": %s",
|
|
static_cast<int64_t>(declared_length + current_offset), strerror(errno));
|
|
return FileWriter{};
|
|
}
|
|
}
|
|
|
|
return FileWriter(fd, declared_length);
|
|
}
|
|
|
|
FileWriter(FileWriter&& other)
|
|
: fd_(other.fd_),
|
|
declared_length_(other.declared_length_),
|
|
total_bytes_written_(other.total_bytes_written_) {
|
|
other.fd_ = -1;
|
|
}
|
|
|
|
bool IsValid() const { return fd_ != -1; }
|
|
|
|
virtual bool Append(uint8_t* buf, size_t buf_size) override {
|
|
if (total_bytes_written_ + buf_size > declared_length_) {
|
|
ALOGW("Zip: Unexpected size " ZD " (declared) vs " ZD " (actual)", declared_length_,
|
|
total_bytes_written_ + buf_size);
|
|
return false;
|
|
}
|
|
|
|
const bool result = android::base::WriteFully(fd_, buf, buf_size);
|
|
if (result) {
|
|
total_bytes_written_ += buf_size;
|
|
} else {
|
|
ALOGW("Zip: unable to write " ZD " bytes to file; %s", buf_size, strerror(errno));
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
private:
|
|
explicit FileWriter(const int fd = -1, const size_t declared_length = 0)
|
|
: Writer(), fd_(fd), declared_length_(declared_length), total_bytes_written_(0) {}
|
|
|
|
int fd_;
|
|
const size_t declared_length_;
|
|
size_t total_bytes_written_;
|
|
};
|
|
|
|
class EntryReader : public zip_archive::Reader {
|
|
public:
|
|
EntryReader(const MappedZipFile& zip_file, const ZipEntry* entry)
|
|
: Reader(), zip_file_(zip_file), entry_(entry) {}
|
|
|
|
virtual bool ReadAtOffset(uint8_t* buf, size_t len, uint32_t offset) const {
|
|
return zip_file_.ReadAtOffset(buf, len, entry_->offset + offset);
|
|
}
|
|
|
|
virtual ~EntryReader() {}
|
|
|
|
private:
|
|
const MappedZipFile& zip_file_;
|
|
const ZipEntry* entry_;
|
|
};
|
|
|
|
// This method is using libz macros with old-style-casts
|
|
#pragma GCC diagnostic push
|
|
#pragma GCC diagnostic ignored "-Wold-style-cast"
|
|
static inline int zlib_inflateInit2(z_stream* stream, int window_bits) {
|
|
return inflateInit2(stream, window_bits);
|
|
}
|
|
#pragma GCC diagnostic pop
|
|
|
|
namespace zip_archive {
|
|
|
|
// Moved out of line to avoid -Wweak-vtables.
|
|
Reader::~Reader() {}
|
|
Writer::~Writer() {}
|
|
|
|
int32_t Inflate(const Reader& reader, const uint32_t compressed_length,
|
|
const uint32_t uncompressed_length, Writer* writer, uint64_t* crc_out) {
|
|
const size_t kBufSize = 32768;
|
|
std::vector<uint8_t> read_buf(kBufSize);
|
|
std::vector<uint8_t> write_buf(kBufSize);
|
|
z_stream zstream;
|
|
int zerr;
|
|
|
|
/*
|
|
* Initialize the zlib stream struct.
|
|
*/
|
|
memset(&zstream, 0, sizeof(zstream));
|
|
zstream.zalloc = Z_NULL;
|
|
zstream.zfree = Z_NULL;
|
|
zstream.opaque = Z_NULL;
|
|
zstream.next_in = NULL;
|
|
zstream.avail_in = 0;
|
|
zstream.next_out = &write_buf[0];
|
|
zstream.avail_out = kBufSize;
|
|
zstream.data_type = Z_UNKNOWN;
|
|
|
|
/*
|
|
* Use the undocumented "negative window bits" feature to tell zlib
|
|
* that there's no zlib header waiting for it.
|
|
*/
|
|
zerr = zlib_inflateInit2(&zstream, -MAX_WBITS);
|
|
if (zerr != Z_OK) {
|
|
if (zerr == Z_VERSION_ERROR) {
|
|
ALOGE("Installed zlib is not compatible with linked version (%s)", ZLIB_VERSION);
|
|
} else {
|
|
ALOGW("Call to inflateInit2 failed (zerr=%d)", zerr);
|
|
}
|
|
|
|
return kZlibError;
|
|
}
|
|
|
|
auto zstream_deleter = [](z_stream* stream) {
|
|
inflateEnd(stream); /* free up any allocated structures */
|
|
};
|
|
|
|
std::unique_ptr<z_stream, decltype(zstream_deleter)> zstream_guard(&zstream, zstream_deleter);
|
|
|
|
const bool compute_crc = (crc_out != nullptr);
|
|
uint64_t crc = 0;
|
|
uint32_t remaining_bytes = compressed_length;
|
|
do {
|
|
/* read as much as we can */
|
|
if (zstream.avail_in == 0) {
|
|
const size_t read_size = (remaining_bytes > kBufSize) ? kBufSize : remaining_bytes;
|
|
const uint32_t offset = (compressed_length - remaining_bytes);
|
|
// Make sure to read at offset to ensure concurrent access to the fd.
|
|
if (!reader.ReadAtOffset(read_buf.data(), read_size, offset)) {
|
|
ALOGW("Zip: inflate read failed, getSize = %zu: %s", read_size, strerror(errno));
|
|
return kIoError;
|
|
}
|
|
|
|
remaining_bytes -= read_size;
|
|
|
|
zstream.next_in = &read_buf[0];
|
|
zstream.avail_in = read_size;
|
|
}
|
|
|
|
/* uncompress the data */
|
|
zerr = inflate(&zstream, Z_NO_FLUSH);
|
|
if (zerr != Z_OK && zerr != Z_STREAM_END) {
|
|
ALOGW("Zip: inflate zerr=%d (nIn=%p aIn=%u nOut=%p aOut=%u)", zerr, zstream.next_in,
|
|
zstream.avail_in, zstream.next_out, zstream.avail_out);
|
|
return kZlibError;
|
|
}
|
|
|
|
/* write when we're full or when we're done */
|
|
if (zstream.avail_out == 0 || (zerr == Z_STREAM_END && zstream.avail_out != kBufSize)) {
|
|
const size_t write_size = zstream.next_out - &write_buf[0];
|
|
if (!writer->Append(&write_buf[0], write_size)) {
|
|
return kIoError;
|
|
} else if (compute_crc) {
|
|
crc = crc32(crc, &write_buf[0], write_size);
|
|
}
|
|
|
|
zstream.next_out = &write_buf[0];
|
|
zstream.avail_out = kBufSize;
|
|
}
|
|
} while (zerr == Z_OK);
|
|
|
|
assert(zerr == Z_STREAM_END); /* other errors should've been caught */
|
|
|
|
// NOTE: zstream.adler is always set to 0, because we're using the -MAX_WBITS
|
|
// "feature" of zlib to tell it there won't be a zlib file header. zlib
|
|
// doesn't bother calculating the checksum in that scenario. We just do
|
|
// it ourselves above because there are no additional gains to be made by
|
|
// having zlib calculate it for us, since they do it by calling crc32 in
|
|
// the same manner that we have above.
|
|
if (compute_crc) {
|
|
*crc_out = crc;
|
|
}
|
|
|
|
if (zstream.total_out != uncompressed_length || remaining_bytes != 0) {
|
|
ALOGW("Zip: size mismatch on inflated file (%lu vs %" PRIu32 ")", zstream.total_out,
|
|
uncompressed_length);
|
|
return kInconsistentInformation;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
} // namespace zip_archive
|
|
|
|
static int32_t InflateEntryToWriter(MappedZipFile& mapped_zip, const ZipEntry* entry,
|
|
zip_archive::Writer* writer, uint64_t* crc_out) {
|
|
const EntryReader reader(mapped_zip, entry);
|
|
|
|
return zip_archive::Inflate(reader, entry->compressed_length, entry->uncompressed_length, writer,
|
|
crc_out);
|
|
}
|
|
|
|
static int32_t CopyEntryToWriter(MappedZipFile& mapped_zip, const ZipEntry* entry,
|
|
zip_archive::Writer* writer, uint64_t* crc_out) {
|
|
static const uint32_t kBufSize = 32768;
|
|
std::vector<uint8_t> buf(kBufSize);
|
|
|
|
const uint32_t length = entry->uncompressed_length;
|
|
uint32_t count = 0;
|
|
uint64_t crc = 0;
|
|
while (count < length) {
|
|
uint32_t remaining = length - count;
|
|
off64_t offset = entry->offset + count;
|
|
|
|
// Safe conversion because kBufSize is narrow enough for a 32 bit signed value.
|
|
const size_t block_size = (remaining > kBufSize) ? kBufSize : remaining;
|
|
|
|
// Make sure to read at offset to ensure concurrent access to the fd.
|
|
if (!mapped_zip.ReadAtOffset(buf.data(), block_size, offset)) {
|
|
ALOGW("CopyFileToFile: copy read failed, block_size = %zu, offset = %" PRId64 ": %s",
|
|
block_size, static_cast<int64_t>(offset), strerror(errno));
|
|
return kIoError;
|
|
}
|
|
|
|
if (!writer->Append(&buf[0], block_size)) {
|
|
return kIoError;
|
|
}
|
|
crc = crc32(crc, &buf[0], block_size);
|
|
count += block_size;
|
|
}
|
|
|
|
*crc_out = crc;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int32_t ExtractToWriter(ZipArchiveHandle handle, ZipEntry* entry, zip_archive::Writer* writer) {
|
|
ZipArchive* archive = reinterpret_cast<ZipArchive*>(handle);
|
|
const uint16_t method = entry->method;
|
|
|
|
// this should default to kUnknownCompressionMethod.
|
|
int32_t return_value = -1;
|
|
uint64_t crc = 0;
|
|
if (method == kCompressStored) {
|
|
return_value = CopyEntryToWriter(archive->mapped_zip, entry, writer, &crc);
|
|
} else if (method == kCompressDeflated) {
|
|
return_value = InflateEntryToWriter(archive->mapped_zip, entry, writer, &crc);
|
|
}
|
|
|
|
if (!return_value && entry->has_data_descriptor) {
|
|
return_value = ValidateDataDescriptor(archive->mapped_zip, entry);
|
|
if (return_value) {
|
|
return return_value;
|
|
}
|
|
}
|
|
|
|
// Validate that the CRC matches the calculated value.
|
|
if (kCrcChecksEnabled && (entry->crc32 != static_cast<uint32_t>(crc))) {
|
|
ALOGW("Zip: crc mismatch: expected %" PRIu32 ", was %" PRIu64, entry->crc32, crc);
|
|
return kInconsistentInformation;
|
|
}
|
|
|
|
return return_value;
|
|
}
|
|
|
|
int32_t ExtractToMemory(ZipArchiveHandle handle, ZipEntry* entry, uint8_t* begin, uint32_t size) {
|
|
MemoryWriter writer(begin, size);
|
|
return ExtractToWriter(handle, entry, &writer);
|
|
}
|
|
|
|
int32_t ExtractEntryToFile(ZipArchiveHandle handle, ZipEntry* entry, int fd) {
|
|
auto writer = FileWriter::Create(fd, entry);
|
|
if (!writer.IsValid()) {
|
|
return kIoError;
|
|
}
|
|
|
|
return ExtractToWriter(handle, entry, &writer);
|
|
}
|
|
|
|
const char* ErrorCodeString(int32_t error_code) {
|
|
// Make sure that the number of entries in kErrorMessages and ErrorCodes
|
|
// match.
|
|
static_assert((-kLastErrorCode + 1) == arraysize(kErrorMessages),
|
|
"(-kLastErrorCode + 1) != arraysize(kErrorMessages)");
|
|
|
|
const uint32_t idx = -error_code;
|
|
if (idx < arraysize(kErrorMessages)) {
|
|
return kErrorMessages[idx];
|
|
}
|
|
|
|
return "Unknown return code";
|
|
}
|
|
|
|
int GetFileDescriptor(const ZipArchiveHandle handle) {
|
|
return reinterpret_cast<ZipArchive*>(handle)->mapped_zip.GetFileDescriptor();
|
|
}
|
|
|
|
ZipString::ZipString(const char* entry_name) : name(reinterpret_cast<const uint8_t*>(entry_name)) {
|
|
size_t len = strlen(entry_name);
|
|
CHECK_LE(len, static_cast<size_t>(UINT16_MAX));
|
|
name_length = static_cast<uint16_t>(len);
|
|
}
|
|
|
|
#if !defined(_WIN32)
|
|
class ProcessWriter : public zip_archive::Writer {
|
|
public:
|
|
ProcessWriter(ProcessZipEntryFunction func, void* cookie)
|
|
: Writer(), proc_function_(func), cookie_(cookie) {}
|
|
|
|
virtual bool Append(uint8_t* buf, size_t buf_size) override {
|
|
return proc_function_(buf, buf_size, cookie_);
|
|
}
|
|
|
|
private:
|
|
ProcessZipEntryFunction proc_function_;
|
|
void* cookie_;
|
|
};
|
|
|
|
int32_t ProcessZipEntryContents(ZipArchiveHandle handle, ZipEntry* entry,
|
|
ProcessZipEntryFunction func, void* cookie) {
|
|
ProcessWriter writer(func, cookie);
|
|
return ExtractToWriter(handle, entry, &writer);
|
|
}
|
|
|
|
#endif //! defined(_WIN32)
|
|
|
|
int MappedZipFile::GetFileDescriptor() const {
|
|
if (!has_fd_) {
|
|
ALOGW("Zip: MappedZipFile doesn't have a file descriptor.");
|
|
return -1;
|
|
}
|
|
return fd_;
|
|
}
|
|
|
|
void* MappedZipFile::GetBasePtr() const {
|
|
if (has_fd_) {
|
|
ALOGW("Zip: MappedZipFile doesn't have a base pointer.");
|
|
return nullptr;
|
|
}
|
|
return base_ptr_;
|
|
}
|
|
|
|
off64_t MappedZipFile::GetFileLength() const {
|
|
if (has_fd_) {
|
|
off64_t result = lseek64(fd_, 0, SEEK_END);
|
|
if (result == -1) {
|
|
ALOGE("Zip: lseek on fd %d failed: %s", fd_, strerror(errno));
|
|
}
|
|
return result;
|
|
} else {
|
|
if (base_ptr_ == nullptr) {
|
|
ALOGE("Zip: invalid file map\n");
|
|
return -1;
|
|
}
|
|
return static_cast<off64_t>(data_length_);
|
|
}
|
|
}
|
|
|
|
// Attempts to read |len| bytes into |buf| at offset |off|.
|
|
bool MappedZipFile::ReadAtOffset(uint8_t* buf, size_t len, off64_t off) const {
|
|
if (has_fd_) {
|
|
if (!android::base::ReadFullyAtOffset(fd_, buf, len, off)) {
|
|
ALOGE("Zip: failed to read at offset %" PRId64 "\n", off);
|
|
return false;
|
|
}
|
|
} else {
|
|
if (off < 0 || off > static_cast<off64_t>(data_length_)) {
|
|
ALOGE("Zip: invalid offset: %" PRId64 ", data length: %" PRId64 "\n", off, data_length_);
|
|
return false;
|
|
}
|
|
memcpy(buf, static_cast<uint8_t*>(base_ptr_) + off, len);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
void CentralDirectory::Initialize(void* map_base_ptr, off64_t cd_start_offset, size_t cd_size) {
|
|
base_ptr_ = static_cast<uint8_t*>(map_base_ptr) + cd_start_offset;
|
|
length_ = cd_size;
|
|
}
|
|
|
|
bool ZipArchive::InitializeCentralDirectory(const char* debug_file_name, off64_t cd_start_offset,
|
|
size_t cd_size) {
|
|
if (mapped_zip.HasFd()) {
|
|
if (!directory_map->create(debug_file_name, mapped_zip.GetFileDescriptor(), cd_start_offset,
|
|
cd_size, true /* read only */)) {
|
|
return false;
|
|
}
|
|
|
|
CHECK_EQ(directory_map->getDataLength(), cd_size);
|
|
central_directory.Initialize(directory_map->getDataPtr(), 0 /*offset*/, cd_size);
|
|
} else {
|
|
if (mapped_zip.GetBasePtr() == nullptr) {
|
|
ALOGE("Zip: Failed to map central directory, bad mapped_zip base pointer\n");
|
|
return false;
|
|
}
|
|
if (static_cast<off64_t>(cd_start_offset) + static_cast<off64_t>(cd_size) >
|
|
mapped_zip.GetFileLength()) {
|
|
ALOGE(
|
|
"Zip: Failed to map central directory, offset exceeds mapped memory region ("
|
|
"start_offset %" PRId64 ", cd_size %zu, mapped_region_size %" PRId64 ")",
|
|
static_cast<int64_t>(cd_start_offset), cd_size, mapped_zip.GetFileLength());
|
|
return false;
|
|
}
|
|
|
|
central_directory.Initialize(mapped_zip.GetBasePtr(), cd_start_offset, cd_size);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
tm ZipEntry::GetModificationTime() const {
|
|
tm t = {};
|
|
|
|
t.tm_hour = (mod_time >> 11) & 0x1f;
|
|
t.tm_min = (mod_time >> 5) & 0x3f;
|
|
t.tm_sec = (mod_time & 0x1f) << 1;
|
|
|
|
t.tm_year = ((mod_time >> 25) & 0x7f) + 80;
|
|
t.tm_mon = ((mod_time >> 21) & 0xf) - 1;
|
|
t.tm_mday = (mod_time >> 16) & 0x1f;
|
|
|
|
return t;
|
|
}
|