ec88556460
UserLifecycleTests test disables all Little cores in the course of the test, which causes attempts to add a process into /dev/cpuset/restricted cpuset cgroup to fail with ENOSPC error code, indicating that a process is joining a cpuset cgroup with no online cpus. Current libprocessgroup implementation will log an error on each such occurrence, which spams the logs and makes it hard to analyze test results. Because this situation does not happen in production environment (we do not offline cpus), we can prevent flooding the logs by identifying this case, logging an appropriate error one time and ignore all later similar errors. Bug: 158766131 Test: adb shell "echo 0 > /sys/devices/system/cpu/cpu[0-3]/online" Test: start some apps, observe libprocessgroup errors in the logcat Signed-off-by: Suren Baghdasaryan <surenb@google.com> Change-Id: Ia91d8839d86787569c255481bde077be51c43d93
645 lines
22 KiB
C++
645 lines
22 KiB
C++
/*
|
|
* Copyright (C) 2019 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
//#define LOG_NDEBUG 0
|
|
#define LOG_TAG "libprocessgroup"
|
|
|
|
#include <fcntl.h>
|
|
#include <task_profiles.h>
|
|
#include <string>
|
|
|
|
#include <android-base/file.h>
|
|
#include <android-base/logging.h>
|
|
#include <android-base/properties.h>
|
|
#include <android-base/stringprintf.h>
|
|
#include <android-base/strings.h>
|
|
#include <android-base/threads.h>
|
|
|
|
#include <cutils/android_filesystem_config.h>
|
|
|
|
#include <json/reader.h>
|
|
#include <json/value.h>
|
|
|
|
// To avoid issues in sdk_mac build
|
|
#if defined(__ANDROID__)
|
|
#include <sys/prctl.h>
|
|
#endif
|
|
|
|
using android::base::GetThreadId;
|
|
using android::base::GetUintProperty;
|
|
using android::base::StringPrintf;
|
|
using android::base::StringReplace;
|
|
using android::base::unique_fd;
|
|
using android::base::WriteStringToFile;
|
|
|
|
static constexpr const char* TASK_PROFILE_DB_FILE = "/etc/task_profiles.json";
|
|
static constexpr const char* TASK_PROFILE_DB_VENDOR_FILE = "/vendor/etc/task_profiles.json";
|
|
|
|
static constexpr const char* TEMPLATE_TASK_PROFILE_API_FILE =
|
|
"/etc/task_profiles/task_profiles_%u.json";
|
|
|
|
void ProfileAttribute::Reset(const CgroupController& controller, const std::string& file_name) {
|
|
controller_ = controller;
|
|
file_name_ = file_name;
|
|
}
|
|
|
|
bool ProfileAttribute::GetPathForTask(int tid, std::string* path) const {
|
|
std::string subgroup;
|
|
if (!controller()->GetTaskGroup(tid, &subgroup)) {
|
|
return false;
|
|
}
|
|
|
|
if (path == nullptr) {
|
|
return true;
|
|
}
|
|
|
|
if (subgroup.empty()) {
|
|
*path = StringPrintf("%s/%s", controller()->path(), file_name_.c_str());
|
|
} else {
|
|
*path = StringPrintf("%s/%s/%s", controller()->path(), subgroup.c_str(),
|
|
file_name_.c_str());
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool SetClampsAction::ExecuteForProcess(uid_t, pid_t) const {
|
|
// TODO: add support when kernel supports util_clamp
|
|
LOG(WARNING) << "SetClampsAction::ExecuteForProcess is not supported";
|
|
return false;
|
|
}
|
|
|
|
bool SetClampsAction::ExecuteForTask(int) const {
|
|
// TODO: add support when kernel supports util_clamp
|
|
LOG(WARNING) << "SetClampsAction::ExecuteForTask is not supported";
|
|
return false;
|
|
}
|
|
|
|
// To avoid issues in sdk_mac build
|
|
#if defined(__ANDROID__)
|
|
|
|
bool SetTimerSlackAction::IsTimerSlackSupported(int tid) {
|
|
auto file = StringPrintf("/proc/%d/timerslack_ns", tid);
|
|
|
|
return (access(file.c_str(), W_OK) == 0);
|
|
}
|
|
|
|
bool SetTimerSlackAction::ExecuteForTask(int tid) const {
|
|
static bool sys_supports_timerslack = IsTimerSlackSupported(tid);
|
|
|
|
// v4.6+ kernels support the /proc/<tid>/timerslack_ns interface.
|
|
// TODO: once we've backported this, log if the open(2) fails.
|
|
if (sys_supports_timerslack) {
|
|
auto file = StringPrintf("/proc/%d/timerslack_ns", tid);
|
|
if (!WriteStringToFile(std::to_string(slack_), file)) {
|
|
if (errno == ENOENT) {
|
|
// This happens when process is already dead
|
|
return true;
|
|
}
|
|
PLOG(ERROR) << "set_timerslack_ns write failed";
|
|
}
|
|
}
|
|
|
|
// TODO: Remove when /proc/<tid>/timerslack_ns interface is backported.
|
|
if (tid == 0 || tid == GetThreadId()) {
|
|
if (prctl(PR_SET_TIMERSLACK, slack_) == -1) {
|
|
PLOG(ERROR) << "set_timerslack_ns prctl failed";
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
#endif
|
|
|
|
bool SetAttributeAction::ExecuteForProcess(uid_t, pid_t pid) const {
|
|
return ExecuteForTask(pid);
|
|
}
|
|
|
|
bool SetAttributeAction::ExecuteForTask(int tid) const {
|
|
std::string path;
|
|
|
|
if (!attribute_->GetPathForTask(tid, &path)) {
|
|
LOG(ERROR) << "Failed to find cgroup for tid " << tid;
|
|
return false;
|
|
}
|
|
|
|
if (!WriteStringToFile(value_, path)) {
|
|
PLOG(ERROR) << "Failed to write '" << value_ << "' to " << path;
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool SetCgroupAction::IsAppDependentPath(const std::string& path) {
|
|
return path.find("<uid>", 0) != std::string::npos || path.find("<pid>", 0) != std::string::npos;
|
|
}
|
|
|
|
SetCgroupAction::SetCgroupAction(const CgroupController& c, const std::string& p)
|
|
: controller_(c), path_(p) {
|
|
// file descriptors for app-dependent paths can't be cached
|
|
if (IsAppDependentPath(path_)) {
|
|
// file descriptor is not cached
|
|
fd_.reset(FDS_APP_DEPENDENT);
|
|
return;
|
|
}
|
|
|
|
// file descriptor can be cached later on request
|
|
fd_.reset(FDS_NOT_CACHED);
|
|
}
|
|
|
|
void SetCgroupAction::EnableResourceCaching() {
|
|
std::lock_guard<std::mutex> lock(fd_mutex_);
|
|
if (fd_ != FDS_NOT_CACHED) {
|
|
return;
|
|
}
|
|
|
|
std::string tasks_path = controller_.GetTasksFilePath(path_);
|
|
|
|
if (access(tasks_path.c_str(), W_OK) != 0) {
|
|
// file is not accessible
|
|
fd_.reset(FDS_INACCESSIBLE);
|
|
return;
|
|
}
|
|
|
|
unique_fd fd(TEMP_FAILURE_RETRY(open(tasks_path.c_str(), O_WRONLY | O_CLOEXEC)));
|
|
if (fd < 0) {
|
|
PLOG(ERROR) << "Failed to cache fd '" << tasks_path << "'";
|
|
fd_.reset(FDS_INACCESSIBLE);
|
|
return;
|
|
}
|
|
|
|
fd_ = std::move(fd);
|
|
}
|
|
|
|
void SetCgroupAction::DropResourceCaching() {
|
|
std::lock_guard<std::mutex> lock(fd_mutex_);
|
|
if (fd_ == FDS_NOT_CACHED) {
|
|
return;
|
|
}
|
|
|
|
fd_.reset(FDS_NOT_CACHED);
|
|
}
|
|
|
|
bool SetCgroupAction::AddTidToCgroup(int tid, int fd, const char* controller_name) {
|
|
if (tid <= 0) {
|
|
return true;
|
|
}
|
|
|
|
std::string value = std::to_string(tid);
|
|
|
|
if (TEMP_FAILURE_RETRY(write(fd, value.c_str(), value.length())) == value.length()) {
|
|
return true;
|
|
}
|
|
|
|
// If the thread is in the process of exiting, don't flag an error
|
|
if (errno == ESRCH) {
|
|
return true;
|
|
}
|
|
|
|
// ENOSPC is returned when cpuset cgroup that we are joining has no online cpus
|
|
if (errno == ENOSPC && !strcmp(controller_name, "cpuset")) {
|
|
// This is an abnormal case happening only in testing, so report it only once
|
|
static bool empty_cpuset_reported = false;
|
|
|
|
if (empty_cpuset_reported) {
|
|
return true;
|
|
}
|
|
|
|
LOG(ERROR) << "Failed to add task '" << value
|
|
<< "' into cpuset because all cpus in that cpuset are offline";
|
|
empty_cpuset_reported = true;
|
|
} else {
|
|
PLOG(ERROR) << "AddTidToCgroup failed to write '" << value << "'; fd=" << fd;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool SetCgroupAction::ExecuteForProcess(uid_t uid, pid_t pid) const {
|
|
std::string procs_path = controller()->GetProcsFilePath(path_, uid, pid);
|
|
unique_fd tmp_fd(TEMP_FAILURE_RETRY(open(procs_path.c_str(), O_WRONLY | O_CLOEXEC)));
|
|
if (tmp_fd < 0) {
|
|
PLOG(WARNING) << "Failed to open " << procs_path;
|
|
return false;
|
|
}
|
|
if (!AddTidToCgroup(pid, tmp_fd, controller()->name())) {
|
|
LOG(ERROR) << "Failed to add task into cgroup";
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool SetCgroupAction::ExecuteForTask(int tid) const {
|
|
std::lock_guard<std::mutex> lock(fd_mutex_);
|
|
if (IsFdValid()) {
|
|
// fd is cached, reuse it
|
|
if (!AddTidToCgroup(tid, fd_, controller()->name())) {
|
|
LOG(ERROR) << "Failed to add task into cgroup";
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
if (fd_ == FDS_INACCESSIBLE) {
|
|
// no permissions to access the file, ignore
|
|
return true;
|
|
}
|
|
|
|
if (fd_ == FDS_APP_DEPENDENT) {
|
|
// application-dependent path can't be used with tid
|
|
PLOG(ERROR) << "Application profile can't be applied to a thread";
|
|
return false;
|
|
}
|
|
|
|
// fd was not cached because cached fd can't be used
|
|
std::string tasks_path = controller()->GetTasksFilePath(path_);
|
|
unique_fd tmp_fd(TEMP_FAILURE_RETRY(open(tasks_path.c_str(), O_WRONLY | O_CLOEXEC)));
|
|
if (tmp_fd < 0) {
|
|
PLOG(WARNING) << "Failed to open " << tasks_path << ": " << strerror(errno);
|
|
return false;
|
|
}
|
|
if (!AddTidToCgroup(tid, tmp_fd, controller()->name())) {
|
|
LOG(ERROR) << "Failed to add task into cgroup";
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool WriteFileAction::ExecuteForProcess(uid_t uid, pid_t pid) const {
|
|
std::string filepath(filepath_), value(value_);
|
|
|
|
filepath = StringReplace(filepath, "<uid>", std::to_string(uid), true);
|
|
filepath = StringReplace(filepath, "<pid>", std::to_string(pid), true);
|
|
value = StringReplace(value, "<uid>", std::to_string(uid), true);
|
|
value = StringReplace(value, "<pid>", std::to_string(pid), true);
|
|
|
|
if (!WriteStringToFile(value, filepath)) {
|
|
if (logfailures_) PLOG(ERROR) << "Failed to write '" << value << "' to " << filepath;
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool WriteFileAction::ExecuteForTask(int tid) const {
|
|
std::string filepath(filepath_), value(value_);
|
|
int uid = getuid();
|
|
|
|
filepath = StringReplace(filepath, "<uid>", std::to_string(uid), true);
|
|
filepath = StringReplace(filepath, "<pid>", std::to_string(tid), true);
|
|
value = StringReplace(value, "<uid>", std::to_string(uid), true);
|
|
value = StringReplace(value, "<pid>", std::to_string(tid), true);
|
|
|
|
if (!WriteStringToFile(value, filepath)) {
|
|
if (logfailures_) PLOG(ERROR) << "Failed to write '" << value << "' to " << filepath;
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool ApplyProfileAction::ExecuteForProcess(uid_t uid, pid_t pid) const {
|
|
for (const auto& profile : profiles_) {
|
|
if (!profile->ExecuteForProcess(uid, pid)) {
|
|
PLOG(WARNING) << "ExecuteForProcess failed for aggregate profile";
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool ApplyProfileAction::ExecuteForTask(int tid) const {
|
|
for (const auto& profile : profiles_) {
|
|
profile->ExecuteForTask(tid);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
void ApplyProfileAction::EnableResourceCaching() {
|
|
for (const auto& profile : profiles_) {
|
|
profile->EnableResourceCaching();
|
|
}
|
|
}
|
|
|
|
void ApplyProfileAction::DropResourceCaching() {
|
|
for (const auto& profile : profiles_) {
|
|
profile->DropResourceCaching();
|
|
}
|
|
}
|
|
|
|
void TaskProfile::MoveTo(TaskProfile* profile) {
|
|
profile->elements_ = std::move(elements_);
|
|
profile->res_cached_ = res_cached_;
|
|
}
|
|
|
|
bool TaskProfile::ExecuteForProcess(uid_t uid, pid_t pid) const {
|
|
for (const auto& element : elements_) {
|
|
if (!element->ExecuteForProcess(uid, pid)) {
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool TaskProfile::ExecuteForTask(int tid) const {
|
|
if (tid == 0) {
|
|
tid = GetThreadId();
|
|
}
|
|
for (const auto& element : elements_) {
|
|
if (!element->ExecuteForTask(tid)) {
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
void TaskProfile::EnableResourceCaching() {
|
|
if (res_cached_) {
|
|
return;
|
|
}
|
|
|
|
for (auto& element : elements_) {
|
|
element->EnableResourceCaching();
|
|
}
|
|
|
|
res_cached_ = true;
|
|
}
|
|
|
|
void TaskProfile::DropResourceCaching() {
|
|
if (!res_cached_) {
|
|
return;
|
|
}
|
|
|
|
for (auto& element : elements_) {
|
|
element->DropResourceCaching();
|
|
}
|
|
|
|
res_cached_ = false;
|
|
}
|
|
|
|
void TaskProfiles::DropResourceCaching() const {
|
|
for (auto& iter : profiles_) {
|
|
iter.second->DropResourceCaching();
|
|
}
|
|
}
|
|
|
|
TaskProfiles& TaskProfiles::GetInstance() {
|
|
// Deliberately leak this object to avoid a race between destruction on
|
|
// process exit and concurrent access from another thread.
|
|
static auto* instance = new TaskProfiles;
|
|
return *instance;
|
|
}
|
|
|
|
TaskProfiles::TaskProfiles() {
|
|
// load system task profiles
|
|
if (!Load(CgroupMap::GetInstance(), TASK_PROFILE_DB_FILE)) {
|
|
LOG(ERROR) << "Loading " << TASK_PROFILE_DB_FILE << " for [" << getpid() << "] failed";
|
|
}
|
|
|
|
// load API-level specific system task profiles if available
|
|
unsigned int api_level = GetUintProperty<unsigned int>("ro.product.first_api_level", 0);
|
|
if (api_level > 0) {
|
|
std::string api_profiles_path =
|
|
android::base::StringPrintf(TEMPLATE_TASK_PROFILE_API_FILE, api_level);
|
|
if (!access(api_profiles_path.c_str(), F_OK) || errno != ENOENT) {
|
|
if (!Load(CgroupMap::GetInstance(), api_profiles_path)) {
|
|
LOG(ERROR) << "Loading " << api_profiles_path << " for [" << getpid()
|
|
<< "] failed";
|
|
}
|
|
}
|
|
}
|
|
|
|
// load vendor task profiles if the file exists
|
|
if (!access(TASK_PROFILE_DB_VENDOR_FILE, F_OK) &&
|
|
!Load(CgroupMap::GetInstance(), TASK_PROFILE_DB_VENDOR_FILE)) {
|
|
LOG(ERROR) << "Loading " << TASK_PROFILE_DB_VENDOR_FILE << " for [" << getpid()
|
|
<< "] failed";
|
|
}
|
|
}
|
|
|
|
bool TaskProfiles::Load(const CgroupMap& cg_map, const std::string& file_name) {
|
|
std::string json_doc;
|
|
|
|
if (!android::base::ReadFileToString(file_name, &json_doc)) {
|
|
LOG(ERROR) << "Failed to read task profiles from " << file_name;
|
|
return false;
|
|
}
|
|
|
|
Json::CharReaderBuilder builder;
|
|
std::unique_ptr<Json::CharReader> reader(builder.newCharReader());
|
|
Json::Value root;
|
|
std::string errorMessage;
|
|
if (!reader->parse(&*json_doc.begin(), &*json_doc.end(), &root, &errorMessage)) {
|
|
LOG(ERROR) << "Failed to parse task profiles: " << errorMessage;
|
|
return false;
|
|
}
|
|
|
|
const Json::Value& attr = root["Attributes"];
|
|
for (Json::Value::ArrayIndex i = 0; i < attr.size(); ++i) {
|
|
std::string name = attr[i]["Name"].asString();
|
|
std::string controller_name = attr[i]["Controller"].asString();
|
|
std::string file_attr = attr[i]["File"].asString();
|
|
|
|
auto controller = cg_map.FindController(controller_name);
|
|
if (controller.HasValue()) {
|
|
auto iter = attributes_.find(name);
|
|
if (iter == attributes_.end()) {
|
|
attributes_[name] = std::make_unique<ProfileAttribute>(controller, file_attr);
|
|
} else {
|
|
iter->second->Reset(controller, file_attr);
|
|
}
|
|
} else {
|
|
LOG(WARNING) << "Controller " << controller_name << " is not found";
|
|
}
|
|
}
|
|
|
|
const Json::Value& profiles_val = root["Profiles"];
|
|
for (Json::Value::ArrayIndex i = 0; i < profiles_val.size(); ++i) {
|
|
const Json::Value& profile_val = profiles_val[i];
|
|
|
|
std::string profile_name = profile_val["Name"].asString();
|
|
const Json::Value& actions = profile_val["Actions"];
|
|
auto profile = std::make_shared<TaskProfile>();
|
|
|
|
for (Json::Value::ArrayIndex act_idx = 0; act_idx < actions.size(); ++act_idx) {
|
|
const Json::Value& action_val = actions[act_idx];
|
|
std::string action_name = action_val["Name"].asString();
|
|
const Json::Value& params_val = action_val["Params"];
|
|
if (action_name == "JoinCgroup") {
|
|
std::string controller_name = params_val["Controller"].asString();
|
|
std::string path = params_val["Path"].asString();
|
|
|
|
auto controller = cg_map.FindController(controller_name);
|
|
if (controller.HasValue()) {
|
|
profile->Add(std::make_unique<SetCgroupAction>(controller, path));
|
|
} else {
|
|
LOG(WARNING) << "JoinCgroup: controller " << controller_name << " is not found";
|
|
}
|
|
} else if (action_name == "SetTimerSlack") {
|
|
std::string slack_value = params_val["Slack"].asString();
|
|
char* end;
|
|
unsigned long slack;
|
|
|
|
slack = strtoul(slack_value.c_str(), &end, 10);
|
|
if (end > slack_value.c_str()) {
|
|
profile->Add(std::make_unique<SetTimerSlackAction>(slack));
|
|
} else {
|
|
LOG(WARNING) << "SetTimerSlack: invalid parameter: " << slack_value;
|
|
}
|
|
} else if (action_name == "SetAttribute") {
|
|
std::string attr_name = params_val["Name"].asString();
|
|
std::string attr_value = params_val["Value"].asString();
|
|
|
|
auto iter = attributes_.find(attr_name);
|
|
if (iter != attributes_.end()) {
|
|
profile->Add(
|
|
std::make_unique<SetAttributeAction>(iter->second.get(), attr_value));
|
|
} else {
|
|
LOG(WARNING) << "SetAttribute: unknown attribute: " << attr_name;
|
|
}
|
|
} else if (action_name == "SetClamps") {
|
|
std::string boost_value = params_val["Boost"].asString();
|
|
std::string clamp_value = params_val["Clamp"].asString();
|
|
char* end;
|
|
unsigned long boost;
|
|
|
|
boost = strtoul(boost_value.c_str(), &end, 10);
|
|
if (end > boost_value.c_str()) {
|
|
unsigned long clamp = strtoul(clamp_value.c_str(), &end, 10);
|
|
if (end > clamp_value.c_str()) {
|
|
profile->Add(std::make_unique<SetClampsAction>(boost, clamp));
|
|
} else {
|
|
LOG(WARNING) << "SetClamps: invalid parameter " << clamp_value;
|
|
}
|
|
} else {
|
|
LOG(WARNING) << "SetClamps: invalid parameter: " << boost_value;
|
|
}
|
|
} else if (action_name == "WriteFile") {
|
|
std::string attr_filepath = params_val["FilePath"].asString();
|
|
std::string attr_value = params_val["Value"].asString();
|
|
if (!attr_filepath.empty() && !attr_value.empty()) {
|
|
std::string attr_logfailures = params_val["LogFailures"].asString();
|
|
bool logfailures = attr_logfailures.empty() || attr_logfailures == "true";
|
|
profile->Add(std::make_unique<WriteFileAction>(attr_filepath, attr_value,
|
|
logfailures));
|
|
} else if (attr_filepath.empty()) {
|
|
LOG(WARNING) << "WriteFile: invalid parameter: "
|
|
<< "empty filepath";
|
|
} else if (attr_value.empty()) {
|
|
LOG(WARNING) << "WriteFile: invalid parameter: "
|
|
<< "empty value";
|
|
}
|
|
} else {
|
|
LOG(WARNING) << "Unknown profile action: " << action_name;
|
|
}
|
|
}
|
|
auto iter = profiles_.find(profile_name);
|
|
if (iter == profiles_.end()) {
|
|
profiles_[profile_name] = profile;
|
|
} else {
|
|
// Move the content rather that replace the profile because old profile might be
|
|
// referenced from an aggregate profile if vendor overrides task profiles
|
|
profile->MoveTo(iter->second.get());
|
|
profile.reset();
|
|
}
|
|
}
|
|
|
|
const Json::Value& aggregateprofiles_val = root["AggregateProfiles"];
|
|
for (Json::Value::ArrayIndex i = 0; i < aggregateprofiles_val.size(); ++i) {
|
|
const Json::Value& aggregateprofile_val = aggregateprofiles_val[i];
|
|
|
|
std::string aggregateprofile_name = aggregateprofile_val["Name"].asString();
|
|
const Json::Value& aggregateprofiles = aggregateprofile_val["Profiles"];
|
|
std::vector<std::shared_ptr<TaskProfile>> profiles;
|
|
bool ret = true;
|
|
|
|
for (Json::Value::ArrayIndex pf_idx = 0; pf_idx < aggregateprofiles.size(); ++pf_idx) {
|
|
std::string profile_name = aggregateprofiles[pf_idx].asString();
|
|
|
|
if (profile_name == aggregateprofile_name) {
|
|
LOG(WARNING) << "AggregateProfiles: recursive profile name: " << profile_name;
|
|
ret = false;
|
|
break;
|
|
} else if (profiles_.find(profile_name) == profiles_.end()) {
|
|
LOG(WARNING) << "AggregateProfiles: undefined profile name: " << profile_name;
|
|
ret = false;
|
|
break;
|
|
} else {
|
|
profiles.push_back(profiles_[profile_name]);
|
|
}
|
|
}
|
|
if (ret) {
|
|
auto profile = std::make_shared<TaskProfile>();
|
|
profile->Add(std::make_unique<ApplyProfileAction>(profiles));
|
|
profiles_[aggregateprofile_name] = profile;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
TaskProfile* TaskProfiles::GetProfile(const std::string& name) const {
|
|
auto iter = profiles_.find(name);
|
|
|
|
if (iter != profiles_.end()) {
|
|
return iter->second.get();
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
const ProfileAttribute* TaskProfiles::GetAttribute(const std::string& name) const {
|
|
auto iter = attributes_.find(name);
|
|
|
|
if (iter != attributes_.end()) {
|
|
return iter->second.get();
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
bool TaskProfiles::SetProcessProfiles(uid_t uid, pid_t pid,
|
|
const std::vector<std::string>& profiles) {
|
|
for (const auto& name : profiles) {
|
|
TaskProfile* profile = GetProfile(name);
|
|
if (profile != nullptr) {
|
|
if (!profile->ExecuteForProcess(uid, pid)) {
|
|
PLOG(WARNING) << "Failed to apply " << name << " process profile";
|
|
}
|
|
} else {
|
|
PLOG(WARNING) << "Failed to find " << name << "process profile";
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool TaskProfiles::SetTaskProfiles(int tid, const std::vector<std::string>& profiles,
|
|
bool use_fd_cache) {
|
|
for (const auto& name : profiles) {
|
|
TaskProfile* profile = GetProfile(name);
|
|
if (profile != nullptr) {
|
|
if (use_fd_cache) {
|
|
profile->EnableResourceCaching();
|
|
}
|
|
if (!profile->ExecuteForTask(tid)) {
|
|
PLOG(WARNING) << "Failed to apply " << name << " task profile";
|
|
}
|
|
} else {
|
|
PLOG(WARNING) << "Failed to find " << name << "task profile";
|
|
}
|
|
}
|
|
return true;
|
|
}
|