blob: ae59d0ade1285022c45e112ef70a2431e7a6b6eb [file] [log] [blame]
/*
* Copyright (C) 2015 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_RUNTIME_JIT_PROFILING_INFO_H_
#define ART_RUNTIME_JIT_PROFILING_INFO_H_
#include <vector>
#include "base/macros.h"
#include "base/value_object.h"
#include "gc_root.h"
#include "interpreter/mterp/nterp.h"
#include "offsets.h"
namespace art HIDDEN {
class ArtMethod;
class CompilerOptions;
class ProfilingInfo;
namespace jit {
class Jit;
class JitCodeCache;
} // namespace jit
namespace mirror {
class Class;
} // namespace mirror
// Structure to store the classes seen at runtime for a specific instruction.
// Once the classes_ array is full, we consider the INVOKE to be megamorphic.
class InlineCache {
public:
// This is hard coded in the assembly stub art_quick_update_inline_cache.
static constexpr uint8_t kIndividualCacheSize = 5;
static constexpr MemberOffset ClassesOffset() {
return MemberOffset(OFFSETOF_MEMBER(InlineCache, classes_));
}
// Encode the list of `dex_pcs` to fit into an uint32_t.
static uint32_t EncodeDexPc(ArtMethod* method,
const std::vector<uint32_t>& dex_pcs,
uint32_t inline_max_code_units)
REQUIRES_SHARED(Locks::mutator_lock_);
// Return the maximum inlining depth that we support to encode a list of dex
// pcs.
static uint32_t MaxDexPcEncodingDepth(ArtMethod* method,
uint32_t inline_max_code_units)
REQUIRES_SHARED(Locks::mutator_lock_);
private:
uint32_t dex_pc_;
GcRoot<mirror::Class> classes_[kIndividualCacheSize];
friend class jit::JitCodeCache;
friend class ProfilingInfo;
DISALLOW_COPY_AND_ASSIGN(InlineCache);
};
class BranchCache {
public:
static constexpr MemberOffset FalseOffset() {
return MemberOffset(OFFSETOF_MEMBER(BranchCache, false_));
}
static constexpr MemberOffset TrueOffset() {
return MemberOffset(OFFSETOF_MEMBER(BranchCache, true_));
}
uint32_t GetExecutionCount() const {
return true_ + false_;
}
uint16_t GetTrue() const {
return true_;
}
uint16_t GetFalse() const {
return false_;
}
private:
uint32_t dex_pc_;
uint16_t false_;
uint16_t true_;
friend class ProfilingInfo;
DISALLOW_COPY_AND_ASSIGN(BranchCache);
};
/**
* Profiling info for a method, created and filled by the interpreter once the
* method is warm, and used by the compiler to drive optimizations.
*/
class ProfilingInfo {
public:
// Create a ProfilingInfo for 'method'.
EXPORT static ProfilingInfo* Create(Thread* self,
ArtMethod* method,
const std::vector<uint32_t>& inline_cache_entries)
REQUIRES_SHARED(Locks::mutator_lock_);
// Add information from an executed INVOKE instruction to the profile.
void AddInvokeInfo(uint32_t dex_pc, mirror::Class* cls)
// Method should not be interruptible, as it manipulates the ProfilingInfo
// which can be concurrently collected.
REQUIRES(Roles::uninterruptible_)
REQUIRES_SHARED(Locks::mutator_lock_);
ArtMethod* GetMethod() const {
return method_;
}
InlineCache* GetInlineCache(uint32_t dex_pc);
BranchCache* GetBranchCache(uint32_t dex_pc);
InlineCache* GetInlineCaches() {
return reinterpret_cast<InlineCache*>(
reinterpret_cast<uintptr_t>(this) + sizeof(ProfilingInfo));
}
BranchCache* GetBranchCaches() {
return reinterpret_cast<BranchCache*>(
reinterpret_cast<uintptr_t>(this) + sizeof(ProfilingInfo) +
number_of_inline_caches_ * sizeof(InlineCache));
}
static size_t ComputeSize(uint32_t number_of_inline_caches, uint32_t number_of_branch_caches) {
return sizeof(ProfilingInfo) +
number_of_inline_caches * sizeof(InlineCache) +
number_of_branch_caches * sizeof(BranchCache);
}
// Increments the number of times this method is currently being inlined.
// Returns whether it was successful, that is it could increment without
// overflowing.
bool IncrementInlineUse() {
if (current_inline_uses_ == std::numeric_limits<uint16_t>::max()) {
return false;
}
current_inline_uses_++;
return true;
}
void DecrementInlineUse() {
DCHECK_GT(current_inline_uses_, 0);
current_inline_uses_--;
}
bool IsInUseByCompiler() const {
return current_inline_uses_ > 0;
}
static constexpr MemberOffset BaselineHotnessCountOffset() {
return MemberOffset(OFFSETOF_MEMBER(ProfilingInfo, baseline_hotness_count_));
}
void ResetCounter() {
baseline_hotness_count_ = GetOptimizeThreshold();
}
bool CounterHasChanged() const {
return baseline_hotness_count_ != GetOptimizeThreshold();
}
uint16_t GetBaselineHotnessCount() const {
return baseline_hotness_count_;
}
static uint16_t GetOptimizeThreshold();
private:
ProfilingInfo(ArtMethod* method,
const std::vector<uint32_t>& inline_cache_entries,
const std::vector<uint32_t>& branch_cache_entries);
// Hotness count for methods compiled with the JIT baseline compiler. Once
// a threshold is hit (currentily the maximum value of uint16_t), we will
// JIT compile optimized the method.
uint16_t baseline_hotness_count_;
// Method this profiling info is for.
// Not 'const' as JVMTI introduces obsolete methods that we implement by creating new ArtMethods.
// See JitCodeCache::MoveObsoleteMethod.
ArtMethod* method_;
// Number of invokes we are profiling in the ArtMethod.
const uint32_t number_of_inline_caches_;
// Number of branches we are profiling in the ArtMethod.
const uint32_t number_of_branch_caches_;
// When the compiler inlines the method associated to this ProfilingInfo,
// it updates this counter so that the GC does not try to clear the inline caches.
uint16_t current_inline_uses_;
// Memory following the object:
// - Dynamically allocated array of `InlineCache` of size `number_of_inline_caches_`.
// - Dynamically allocated array of `BranchCache of size `number_of_branch_caches_`.
friend class jit::JitCodeCache;
DISALLOW_COPY_AND_ASSIGN(ProfilingInfo);
};
class ScopedProfilingInfoUse : public ValueObject {
public:
ScopedProfilingInfoUse(jit::Jit* jit, ArtMethod* method, Thread* self);
~ScopedProfilingInfoUse();
ProfilingInfo* GetProfilingInfo() const { return profiling_info_; }
private:
jit::Jit* const jit_;
ArtMethod* const method_;
Thread* const self_;
ProfilingInfo* const profiling_info_;
DISALLOW_COPY_AND_ASSIGN(ScopedProfilingInfoUse);
};
} // namespace art
#endif // ART_RUNTIME_JIT_PROFILING_INFO_H_