blob: 10f7f3183d6a4a6029e0aafc0ab1dd46e7274ffa [file] [log] [blame]
/*
* Copyright (C) 2013 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_LIBARTBASE_BASE_ARENA_ALLOCATOR_H_
#define ART_LIBARTBASE_BASE_ARENA_ALLOCATOR_H_
#include <stddef.h>
#include <stdint.h>
#include "bit_utils.h"
#include "debug_stack.h"
#include "dchecked_vector.h"
#include "macros.h"
#include "memory_tool.h"
namespace art {
class Arena;
class ArenaPool;
class ArenaAllocator;
class ArenaStack;
class ScopedArenaAllocator;
class MemStats;
template <typename T>
class ArenaAllocatorAdapter;
static constexpr bool kArenaAllocatorCountAllocations = false;
// Type of allocation for memory tuning.
enum ArenaAllocKind {
kArenaAllocMisc,
kArenaAllocSwitchTable,
kArenaAllocSlowPaths,
kArenaAllocGrowableBitMap,
kArenaAllocSTL,
kArenaAllocGraphBuilder,
kArenaAllocGraph,
kArenaAllocBasicBlock,
kArenaAllocBlockList,
kArenaAllocReversePostOrder,
kArenaAllocLinearOrder,
kArenaAllocReachabilityGraph,
kArenaAllocConstantsMap,
kArenaAllocPredecessors,
kArenaAllocSuccessors,
kArenaAllocDominated,
kArenaAllocInstruction,
kArenaAllocConstructorFenceInputs,
kArenaAllocInvokeInputs,
kArenaAllocPhiInputs,
kArenaAllocTypeCheckInputs,
kArenaAllocLoopInfo,
kArenaAllocLoopInfoBackEdges,
kArenaAllocTryCatchInfo,
kArenaAllocUseListNode,
kArenaAllocEnvironment,
kArenaAllocEnvironmentVRegs,
kArenaAllocEnvironmentLocations,
kArenaAllocLocationSummary,
kArenaAllocSsaBuilder,
kArenaAllocMoveOperands,
kArenaAllocCodeBuffer,
kArenaAllocStackMaps,
kArenaAllocOptimization,
kArenaAllocGvn,
kArenaAllocInductionVarAnalysis,
kArenaAllocBoundsCheckElimination,
kArenaAllocDCE,
kArenaAllocLSA,
kArenaAllocLSE,
kArenaAllocCFRE,
kArenaAllocLICM,
kArenaAllocWBE,
kArenaAllocLoopOptimization,
kArenaAllocSsaLiveness,
kArenaAllocSsaPhiElimination,
kArenaAllocReferenceTypePropagation,
kArenaAllocSelectGenerator,
kArenaAllocSideEffectsAnalysis,
kArenaAllocRegisterAllocator,
kArenaAllocRegisterAllocatorValidate,
kArenaAllocStackMapStream,
kArenaAllocBitTableBuilder,
kArenaAllocVectorNode,
kArenaAllocCodeGenerator,
kArenaAllocAssembler,
kArenaAllocParallelMoveResolver,
kArenaAllocGraphChecker,
kArenaAllocVerifier,
kArenaAllocCallingConvention,
kArenaAllocCHA,
kArenaAllocScheduler,
kArenaAllocProfile,
kArenaAllocSuperblockCloner,
kArenaAllocTransaction,
kNumArenaAllocKinds
};
template <bool kCount>
class ArenaAllocatorStatsImpl;
template <>
class ArenaAllocatorStatsImpl<false> {
public:
ArenaAllocatorStatsImpl() = default;
ArenaAllocatorStatsImpl(const ArenaAllocatorStatsImpl& other) = default;
ArenaAllocatorStatsImpl& operator = (const ArenaAllocatorStatsImpl& other) = delete;
void Copy([[maybe_unused]] const ArenaAllocatorStatsImpl& other) {}
void RecordAlloc([[maybe_unused]] size_t bytes, [[maybe_unused]] ArenaAllocKind kind) {}
size_t NumAllocations() const { return 0u; }
size_t BytesAllocated() const { return 0u; }
void Dump([[maybe_unused]] std::ostream& os,
[[maybe_unused]] const Arena* first,
[[maybe_unused]] ssize_t lost_bytes_adjustment) const {}
};
template <bool kCount>
class ArenaAllocatorStatsImpl {
public:
ArenaAllocatorStatsImpl();
ArenaAllocatorStatsImpl(const ArenaAllocatorStatsImpl& other) = default;
ArenaAllocatorStatsImpl& operator = (const ArenaAllocatorStatsImpl& other) = delete;
void Copy(const ArenaAllocatorStatsImpl& other);
void RecordAlloc(size_t bytes, ArenaAllocKind kind);
size_t NumAllocations() const;
size_t BytesAllocated() const;
void Dump(std::ostream& os, const Arena* first, ssize_t lost_bytes_adjustment) const;
private:
size_t num_allocations_;
dchecked_vector<size_t> alloc_stats_; // Bytes used by various allocation kinds.
static const char* const kAllocNames[];
};
using ArenaAllocatorStats = ArenaAllocatorStatsImpl<kArenaAllocatorCountAllocations>;
class ArenaAllocatorMemoryTool {
public:
static constexpr bool IsRunningOnMemoryTool() { return kMemoryToolIsAvailable; }
void MakeDefined(void* ptr, size_t size) {
if (UNLIKELY(IsRunningOnMemoryTool())) {
DoMakeDefined(ptr, size);
}
}
void MakeUndefined(void* ptr, size_t size) {
if (UNLIKELY(IsRunningOnMemoryTool())) {
DoMakeUndefined(ptr, size);
}
}
void MakeInaccessible(void* ptr, size_t size) {
if (UNLIKELY(IsRunningOnMemoryTool())) {
DoMakeInaccessible(ptr, size);
}
}
private:
void DoMakeDefined(void* ptr, size_t size);
void DoMakeUndefined(void* ptr, size_t size);
void DoMakeInaccessible(void* ptr, size_t size);
};
class Arena {
public:
Arena() : bytes_allocated_(0), memory_(nullptr), size_(0), next_(nullptr) {}
virtual ~Arena() { }
// Reset is for pre-use and uses memset for performance.
void Reset();
// Release is used inbetween uses and uses madvise for memory usage.
virtual void Release() { }
uint8_t* Begin() const {
return memory_;
}
uint8_t* End() const { return memory_ + size_; }
size_t Size() const {
return size_;
}
size_t RemainingSpace() const {
return Size() - bytes_allocated_;
}
size_t GetBytesAllocated() const {
return bytes_allocated_;
}
// Return true if ptr is contained in the arena.
bool Contains(const void* ptr) const { return memory_ <= ptr && ptr < memory_ + size_; }
Arena* Next() const { return next_; }
protected:
size_t bytes_allocated_;
uint8_t* memory_;
size_t size_;
Arena* next_;
friend class MallocArenaPool;
friend class MemMapArenaPool;
friend class ArenaAllocator;
friend class ArenaStack;
friend class ScopedArenaAllocator;
template <bool kCount> friend class ArenaAllocatorStatsImpl;
friend class ArenaAllocatorTest;
private:
DISALLOW_COPY_AND_ASSIGN(Arena);
};
class ArenaPool {
public:
virtual ~ArenaPool() = default;
virtual Arena* AllocArena(size_t size) = 0;
virtual void FreeArenaChain(Arena* first) = 0;
virtual size_t GetBytesAllocated() const = 0;
virtual void ReclaimMemory() = 0;
virtual void LockReclaimMemory() = 0;
// Trim the maps in arenas by madvising, used by JIT to reduce memory usage.
virtual void TrimMaps() = 0;
protected:
ArenaPool() = default;
private:
DISALLOW_COPY_AND_ASSIGN(ArenaPool);
};
// Fast single-threaded allocator for zero-initialized memory chunks.
//
// Memory is allocated from ArenaPool in large chunks and then rationed through
// the ArenaAllocator. It's returned to the ArenaPool only when the ArenaAllocator
// is destroyed.
class ArenaAllocator
: private DebugStackRefCounter, private ArenaAllocatorStats, private ArenaAllocatorMemoryTool {
public:
explicit ArenaAllocator(ArenaPool* pool);
~ArenaAllocator();
using ArenaAllocatorMemoryTool::IsRunningOnMemoryTool;
using ArenaAllocatorMemoryTool::MakeDefined;
using ArenaAllocatorMemoryTool::MakeUndefined;
using ArenaAllocatorMemoryTool::MakeInaccessible;
// Get adapter for use in STL containers. See arena_containers.h .
ArenaAllocatorAdapter<void> Adapter(ArenaAllocKind kind = kArenaAllocSTL);
// Returns zeroed memory.
void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE {
if (UNLIKELY(IsRunningOnMemoryTool())) {
return AllocWithMemoryTool(bytes, kind);
}
bytes = RoundUp(bytes, kAlignment);
ArenaAllocatorStats::RecordAlloc(bytes, kind);
if (UNLIKELY(bytes > static_cast<size_t>(end_ - ptr_))) {
return AllocFromNewArena(bytes);
}
uint8_t* ret = ptr_;
DCHECK_ALIGNED(ret, kAlignment);
ptr_ += bytes;
return ret;
}
// Returns zeroed memory.
void* AllocAlign16(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE {
// It is an error to request 16-byte aligned allocation of unaligned size.
DCHECK_ALIGNED(bytes, 16);
if (UNLIKELY(IsRunningOnMemoryTool())) {
return AllocWithMemoryToolAlign16(bytes, kind);
}
uintptr_t padding =
RoundUp(reinterpret_cast<uintptr_t>(ptr_), 16) - reinterpret_cast<uintptr_t>(ptr_);
ArenaAllocatorStats::RecordAlloc(bytes, kind);
if (UNLIKELY(padding + bytes > static_cast<size_t>(end_ - ptr_))) {
static_assert(kArenaAlignment >= 16, "Expecting sufficient alignment for new Arena.");
return AllocFromNewArena(bytes);
}
ptr_ += padding;
uint8_t* ret = ptr_;
DCHECK_ALIGNED(ret, 16);
ptr_ += bytes;
return ret;
}
// Realloc never frees the input pointer, it is the caller's job to do this if necessary.
void* Realloc(void* ptr,
size_t ptr_size,
size_t new_size,
ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE {
DCHECK_GE(new_size, ptr_size);
DCHECK_EQ(ptr == nullptr, ptr_size == 0u);
// We always allocate aligned.
const size_t aligned_ptr_size = RoundUp(ptr_size, kAlignment);
auto* end = reinterpret_cast<uint8_t*>(ptr) + aligned_ptr_size;
// If we haven't allocated anything else, we can safely extend.
if (end == ptr_) {
// Red zone prevents end == ptr_ (unless input = allocator state = null).
DCHECK(!IsRunningOnMemoryTool() || ptr_ == nullptr);
const size_t aligned_new_size = RoundUp(new_size, kAlignment);
const size_t size_delta = aligned_new_size - aligned_ptr_size;
// Check remain space.
const size_t remain = end_ - ptr_;
if (remain >= size_delta) {
ptr_ += size_delta;
ArenaAllocatorStats::RecordAlloc(size_delta, kind);
DCHECK_ALIGNED(ptr_, kAlignment);
return ptr;
}
}
auto* new_ptr = Alloc(new_size, kind); // Note: Alloc will take care of aligning new_size.
memcpy(new_ptr, ptr, ptr_size);
// TODO: Call free on ptr if linear alloc supports free.
return new_ptr;
}
template <typename T>
T* Alloc(ArenaAllocKind kind = kArenaAllocMisc) {
return AllocArray<T>(1, kind);
}
template <typename T>
T* AllocArray(size_t length, ArenaAllocKind kind = kArenaAllocMisc) {
return static_cast<T*>(Alloc(length * sizeof(T), kind));
}
size_t BytesAllocated() const;
MemStats GetMemStats() const;
// The BytesUsed method sums up bytes allocated from arenas in arena_head_ and nodes.
// TODO: Change BytesAllocated to this behavior?
size_t BytesUsed() const;
ArenaPool* GetArenaPool() const {
return pool_;
}
Arena* GetHeadArena() const {
return arena_head_;
}
uint8_t* CurrentPtr() const {
return ptr_;
}
size_t CurrentArenaUnusedBytes() const {
DCHECK_LE(ptr_, end_);
return end_ - ptr_;
}
// Resets the current arena in use, which will force us to get a new arena
// on next allocation.
void ResetCurrentArena();
bool Contains(const void* ptr) const;
// The alignment guaranteed for individual allocations.
static constexpr size_t kAlignment = 8u;
// The alignment required for the whole Arena rather than individual allocations.
static constexpr size_t kArenaAlignment = 16u;
// Extra bytes required by the memory tool.
static constexpr size_t kMemoryToolRedZoneBytes = 8u;
private:
void* AllocWithMemoryTool(size_t bytes, ArenaAllocKind kind);
void* AllocWithMemoryToolAlign16(size_t bytes, ArenaAllocKind kind);
uint8_t* AllocFromNewArena(size_t bytes);
uint8_t* AllocFromNewArenaWithMemoryTool(size_t bytes);
void UpdateBytesAllocated();
ArenaPool* pool_;
uint8_t* begin_;
uint8_t* end_;
uint8_t* ptr_;
Arena* arena_head_;
template <typename U>
friend class ArenaAllocatorAdapter;
friend class ArenaAllocatorTest;
DISALLOW_COPY_AND_ASSIGN(ArenaAllocator);
}; // ArenaAllocator
class MemStats {
public:
MemStats(const char* name,
const ArenaAllocatorStats* stats,
const Arena* first_arena,
ssize_t lost_bytes_adjustment = 0);
void Dump(std::ostream& os) const;
private:
const char* const name_;
const ArenaAllocatorStats* const stats_;
const Arena* const first_arena_;
const ssize_t lost_bytes_adjustment_;
}; // MemStats
} // namespace art
#endif // ART_LIBARTBASE_BASE_ARENA_ALLOCATOR_H_