/* * Copyright (C) 2017 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "utils.h" using namespace std::chrono_literals; static void WaitUntilAllThreadsExited(pid_t* tids, size_t tid_count) { // Wait until all children have exited. bool alive = true; while (alive) { alive = false; for (size_t i = 0; i < tid_count; ++i) { if (tids[i] != 0) { if (syscall(__NR_tgkill, getpid(), tids[i], 0) == 0) { alive = true; } else { EXPECT_ERRNO(ESRCH); tids[i] = 0; // Skip in next loop. } } } sched_yield(); } } class LeakChecker { public: LeakChecker() { // Avoid resizing and using memory later. // 64Ki is the default limit on VMAs per process. maps_.reserve(64*1024); Reset(); } ~LeakChecker() { Check(); } void Reset() { previous_size_ = GetMappingSize(); } void DumpTo(std::ostream& os) const { os << previous_size_; } private: size_t previous_size_; std::vector maps_; void Check() { auto current_size = GetMappingSize(); if (current_size > previous_size_) { FAIL() << "increase in process map size: " << previous_size_ << " -> " << current_size; } } size_t GetMappingSize() { if (!Maps::parse_maps(&maps_)) { err(1, "failed to parse maps"); } size_t result = 0; for (const map_record& map : maps_) { result += map.addr_end - map.addr_start; } return result; } }; std::ostream& operator<<(std::ostream& os, const LeakChecker& lc) { lc.DumpTo(os); return os; } // http://b/36045112 TEST(pthread_leak, join) { SKIP_WITH_NATIVE_BRIDGE; // http://b/37920774 // Warm up. HWASan allocates an extra page on the first iteration, but never after. pthread_t thread; ASSERT_EQ(0, pthread_create( &thread, nullptr, [](void*) -> void* { return nullptr; }, nullptr)); ASSERT_EQ(0, pthread_join(thread, nullptr)); LeakChecker lc; for (int i = 0; i < 100; ++i) { ASSERT_EQ(0, pthread_create( &thread, nullptr, [](void*) -> void* { return nullptr; }, nullptr)); ASSERT_EQ(0, pthread_join(thread, nullptr)); } } // http://b/36045112 TEST(pthread_leak, detach) { SKIP_WITH_NATIVE_BRIDGE; // http://b/37920774 LeakChecker lc; // Ancient devices with only 2 cores need a lower limit. // http://b/129924384 and https://issuetracker.google.com/142210680. const int thread_count = (sysconf(_SC_NPROCESSORS_CONF) > 2) ? 100 : 50; for (size_t pass = 0; pass < 1; ++pass) { struct thread_data { pthread_barrier_t* barrier; pid_t* tid; } threads[thread_count]; pthread_barrier_t barrier; ASSERT_EQ(pthread_barrier_init(&barrier, nullptr, thread_count + 1), 0); // Start child threads. pid_t tids[thread_count]; for (int i = 0; i < thread_count; ++i) { threads[i] = {&barrier, &tids[i]}; const auto thread_function = +[](void* ptr) -> void* { thread_data* data = static_cast(ptr); *data->tid = gettid(); pthread_barrier_wait(data->barrier); return nullptr; }; pthread_t thread; ASSERT_EQ(0, pthread_create(&thread, nullptr, thread_function, &threads[i])); ASSERT_EQ(0, pthread_detach(thread)); } pthread_barrier_wait(&barrier); ASSERT_EQ(pthread_barrier_destroy(&barrier), 0); WaitUntilAllThreadsExited(tids, thread_count); // TODO(b/158573595): the test is flaky without the warmup pass. if (pass == 0) lc.Reset(); } }