blob: ae7db45024b2b0cee513f4804cb5da25793409f3 [file] [log] [blame]
/*
* Copyright (C) 2018 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "mem_map_arena_pool.h"
#include <sys/mman.h>
#include <algorithm>
#include <cstddef>
#include <iomanip>
#include <numeric>
#include <android-base/logging.h>
#include "base/arena_allocator-inl.h"
#include "base/mem_map.h"
#include "base/systrace.h"
namespace art {
class MemMapArena final : public Arena {
public:
MemMapArena(size_t size, bool low_4gb, const char* name);
virtual ~MemMapArena();
void Release() override;
private:
static MemMap Allocate(size_t size, bool low_4gb, const char* name);
MemMap map_;
};
MemMapArena::MemMapArena(size_t size, bool low_4gb, const char* name)
: map_(Allocate(size, low_4gb, name)) {
memory_ = map_.Begin();
static_assert(ArenaAllocator::kArenaAlignment <= kPageSize,
"Arena should not need stronger alignment than kPageSize.");
DCHECK_ALIGNED(memory_, ArenaAllocator::kArenaAlignment);
size_ = map_.Size();
}
MemMap MemMapArena::Allocate(size_t size, bool low_4gb, const char* name) {
// Round up to a full page as that's the smallest unit of allocation for mmap()
// and we want to be able to use all memory that we actually allocate.
size = RoundUp(size, kPageSize);
std::string error_msg;
MemMap map = MemMap::MapAnonymous(name,
size,
PROT_READ | PROT_WRITE,
low_4gb,
&error_msg);
CHECK(map.IsValid()) << error_msg;
return map;
}
MemMapArena::~MemMapArena() {
// Destroys MemMap via std::unique_ptr<>.
}
void MemMapArena::Release() {
if (bytes_allocated_ > 0) {
map_.MadviseDontNeedAndZero();
bytes_allocated_ = 0;
}
}
MemMapArenaPool::MemMapArenaPool(bool low_4gb, const char* name)
: low_4gb_(low_4gb),
name_(name),
free_arenas_(nullptr) {
MemMap::Init();
}
MemMapArenaPool::~MemMapArenaPool() {
ReclaimMemory();
}
void MemMapArenaPool::ReclaimMemory() {
while (free_arenas_ != nullptr) {
Arena* arena = free_arenas_;
free_arenas_ = free_arenas_->next_;
delete arena;
}
}
void MemMapArenaPool::LockReclaimMemory() {
std::lock_guard<std::mutex> lock(lock_);
ReclaimMemory();
}
Arena* MemMapArenaPool::AllocArena(size_t size) {
Arena* ret = nullptr;
{
std::lock_guard<std::mutex> lock(lock_);
if (free_arenas_ != nullptr && LIKELY(free_arenas_->Size() >= size)) {
ret = free_arenas_;
free_arenas_ = free_arenas_->next_;
}
}
if (ret == nullptr) {
ret = new MemMapArena(size, low_4gb_, name_);
}
ret->Reset();
return ret;
}
void MemMapArenaPool::TrimMaps() {
ScopedTrace trace(__PRETTY_FUNCTION__);
std::lock_guard<std::mutex> lock(lock_);
for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) {
arena->Release();
}
}
size_t MemMapArenaPool::GetBytesAllocated() const {
size_t total = 0;
std::lock_guard<std::mutex> lock(lock_);
for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) {
total += arena->GetBytesAllocated();
}
return total;
}
void MemMapArenaPool::FreeArenaChain(Arena* first) {
if (kRunningOnMemoryTool) {
for (Arena* arena = first; arena != nullptr; arena = arena->next_) {
MEMORY_TOOL_MAKE_UNDEFINED(arena->memory_, arena->bytes_allocated_);
}
}
if (arena_allocator::kArenaAllocatorPreciseTracking) {
// Do not reuse arenas when tracking.
while (first != nullptr) {
Arena* next = first->next_;
delete first;
first = next;
}
return;
}
if (first != nullptr) {
Arena* last = first;
while (last->next_ != nullptr) {
last = last->next_;
}
std::lock_guard<std::mutex> lock(lock_);
last->next_ = free_arenas_;
free_arenas_ = first;
}
}
} // namespace art