3e5754c6ad
Add benchmarks for mallinfo, and for retrieving RSS from /proc/self/statm, since we're considering using these for GC triggering. Add some static linkage specifiers, after running into a build problem due to a spurious conflict. Bug: 111447610 Test: Ran benchmarks Change-Id: Ie50d512294993882728c63ce51ec507590257d80
150 lines
4.4 KiB
C++
150 lines
4.4 KiB
C++
/*
|
|
* Copyright (C) 2017 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
// Our goal is to measure the cost of various C++ atomic operations.
|
|
// Android doesn't really control those. But since some of these operations can be quite
|
|
// expensive, this may be useful input for development of higher level code.
|
|
// Expected mappings from C++ atomics to hardware primitives can be found at
|
|
// http://www.cl.cam.ac.uk/~pes20/cpp/cpp0xmappings.html .
|
|
|
|
#include <atomic>
|
|
#include <mutex>
|
|
|
|
#include <benchmark/benchmark.h>
|
|
#include "util.h"
|
|
|
|
// We time atomic operations separated by a volatile (not atomic!) increment. This ensures
|
|
// that the compiler emits memory instructions (e.g. load or store) prior to any fence or the
|
|
// like. That in turn ensures that the CPU has outstanding memory operations when the fence
|
|
// is executed.
|
|
|
|
// In most respects, we compute best case values. Since there is only one thread, there are no
|
|
// coherence misses.
|
|
|
|
// We assume that the compiler is not smart enough to optimize away fences in a single-threaded
|
|
// program. If that changes, we'll need to add a second thread.
|
|
|
|
static volatile unsigned counter;
|
|
|
|
std::atomic<int> test_loc(0);
|
|
|
|
static volatile unsigned sink;
|
|
|
|
static std::mutex mtx;
|
|
|
|
void BM_atomic_empty(benchmark::State& state) {
|
|
while (state.KeepRunning()) {
|
|
++counter;
|
|
}
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_empty);
|
|
|
|
static void BM_atomic_load_relaxed(benchmark::State& state) {
|
|
unsigned result = 0;
|
|
while (state.KeepRunning()) {
|
|
result += test_loc.load(std::memory_order_relaxed);
|
|
++counter;
|
|
}
|
|
sink = result;
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_load_relaxed);
|
|
|
|
static void BM_atomic_load_acquire(benchmark::State& state) {
|
|
unsigned result = 0;
|
|
while (state.KeepRunning()) {
|
|
result += test_loc.load(std::memory_order_acquire);
|
|
++counter;
|
|
}
|
|
sink = result;
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_load_acquire);
|
|
|
|
static void BM_atomic_store_release(benchmark::State& state) {
|
|
int i = counter;
|
|
while (state.KeepRunning()) {
|
|
test_loc.store(++i, std::memory_order_release);
|
|
++counter;
|
|
}
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_store_release);
|
|
|
|
static void BM_atomic_store_seq_cst(benchmark::State& state) {
|
|
int i = counter;
|
|
while (state.KeepRunning()) {
|
|
test_loc.store(++i, std::memory_order_seq_cst);
|
|
++counter;
|
|
}
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_store_seq_cst);
|
|
|
|
static void BM_atomic_fetch_add_relaxed(benchmark::State& state) {
|
|
unsigned result = 0;
|
|
while (state.KeepRunning()) {
|
|
result += test_loc.fetch_add(1, std::memory_order_relaxed);
|
|
++counter;
|
|
}
|
|
sink = result;
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_fetch_add_relaxed);
|
|
|
|
static void BM_atomic_fetch_add_seq_cst(benchmark::State& state) {
|
|
unsigned result = 0;
|
|
while (state.KeepRunning()) {
|
|
result += test_loc.fetch_add(1, std::memory_order_seq_cst);
|
|
++counter;
|
|
}
|
|
sink = result;
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_fetch_add_seq_cst);
|
|
|
|
// The fence benchmarks include a relaxed load to make it much harder to optimize away
|
|
// the fence.
|
|
|
|
static void BM_atomic_acquire_fence(benchmark::State& state) {
|
|
unsigned result = 0;
|
|
while (state.KeepRunning()) {
|
|
result += test_loc.load(std::memory_order_relaxed);
|
|
std::atomic_thread_fence(std::memory_order_acquire);
|
|
++counter;
|
|
}
|
|
sink = result;
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_acquire_fence);
|
|
|
|
static void BM_atomic_seq_cst_fence(benchmark::State& state) {
|
|
unsigned result = 0;
|
|
while (state.KeepRunning()) {
|
|
result += test_loc.load(std::memory_order_relaxed);
|
|
std::atomic_thread_fence(std::memory_order_seq_cst);
|
|
++counter;
|
|
}
|
|
sink = result;
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_seq_cst_fence);
|
|
|
|
// For comparison, also throw in a critical section version:
|
|
|
|
static void BM_atomic_fetch_add_cs(benchmark::State& state) {
|
|
unsigned result = 0;
|
|
while (state.KeepRunning()) {
|
|
{
|
|
std::lock_guard<std::mutex> _(mtx);
|
|
result += ++counter;
|
|
}
|
|
}
|
|
sink = result;
|
|
}
|
|
BIONIC_BENCHMARK(BM_atomic_fetch_add_cs);
|