blob: 0a95d499738323bed7437b7a911cceb2352f4ed7 [file] [log] [blame]
/*
* Copyright (C) 2011 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <stdio.h>
#include <stdlib.h>
#include <fstream>
#include <iostream>
#include <map>
#include <set>
#include <string>
#include <unordered_map>
#include <unordered_set>
#include <vector>
#include "android-base/stringprintf.h"
#include "android-base/strings.h"
#include "arch/instruction_set_features.h"
#include "art_field-inl.h"
#include "art_method-inl.h"
#include "base/stl_util.h"
#include "base/unix_file/fd_file.h"
#include "class_linker-inl.h"
#include "class_linker.h"
#include "debug/elf_debug_writer.h"
#include "debug/method_debug_info.h"
#include "dex_file-inl.h"
#include "dex_instruction-inl.h"
#include "disassembler.h"
#include "elf_builder.h"
#include "gc/accounting/space_bitmap-inl.h"
#include "gc/space/image_space.h"
#include "gc/space/large_object_space.h"
#include "gc/space/space-inl.h"
#include "image-inl.h"
#include "imtable-inl.h"
#include "indenter.h"
#include "interpreter/unstarted_runtime.h"
#include "linker/buffered_output_stream.h"
#include "linker/file_output_stream.h"
#include "mirror/array-inl.h"
#include "mirror/class-inl.h"
#include "mirror/dex_cache-inl.h"
#include "mirror/object-inl.h"
#include "mirror/object_array-inl.h"
#include "nativehelper/ScopedLocalRef.h"
#include "oat.h"
#include "oat_file-inl.h"
#include "oat_file_manager.h"
#include "os.h"
#include "safe_map.h"
#include "scoped_thread_state_change-inl.h"
#include "stack.h"
#include "stack_map.h"
#include "string_reference.h"
#include "thread_list.h"
#include "type_lookup_table.h"
#include "vdex_file.h"
#include "verifier/method_verifier.h"
#include "verifier/verifier_deps.h"
#include "well_known_classes.h"
#include <sys/stat.h>
#include "cmdline.h"
namespace art {
using android::base::StringPrintf;
const char* image_methods_descriptions_[] = {
"kResolutionMethod",
"kImtConflictMethod",
"kImtUnimplementedMethod",
"kSaveAllCalleeSavesMethod",
"kSaveRefsOnlyMethod",
"kSaveRefsAndArgsMethod",
"kSaveEverythingMethod",
};
const char* image_roots_descriptions_[] = {
"kDexCaches",
"kClassRoots",
"kClassLoader",
};
// Map is so that we don't allocate multiple dex files for the same OatDexFile.
static std::map<const OatFile::OatDexFile*,
std::unique_ptr<const DexFile>> opened_dex_files;
const DexFile* OpenDexFile(const OatFile::OatDexFile* oat_dex_file, std::string* error_msg) {
DCHECK(oat_dex_file != nullptr);
auto it = opened_dex_files.find(oat_dex_file);
if (it != opened_dex_files.end()) {
return it->second.get();
}
const DexFile* ret = oat_dex_file->OpenDexFile(error_msg).release();
opened_dex_files.emplace(oat_dex_file, std::unique_ptr<const DexFile>(ret));
return ret;
}
template <typename ElfTypes>
class OatSymbolizer FINAL {
public:
OatSymbolizer(const OatFile* oat_file, const std::string& output_name, bool no_bits) :
oat_file_(oat_file),
builder_(nullptr),
output_name_(output_name.empty() ? "symbolized.oat" : output_name),
no_bits_(no_bits) {
}
bool Symbolize() {
const InstructionSet isa = oat_file_->GetOatHeader().GetInstructionSet();
std::unique_ptr<const InstructionSetFeatures> features = InstructionSetFeatures::FromBitmap(
isa, oat_file_->GetOatHeader().GetInstructionSetFeaturesBitmap());
std::unique_ptr<File> elf_file(OS::CreateEmptyFile(output_name_.c_str()));
if (elf_file == nullptr) {
return false;
}
std::unique_ptr<BufferedOutputStream> output_stream =
std::make_unique<BufferedOutputStream>(std::make_unique<FileOutputStream>(elf_file.get()));
builder_.reset(new ElfBuilder<ElfTypes>(isa, features.get(), output_stream.get()));
builder_->Start();
auto* rodata = builder_->GetRoData();
auto* text = builder_->GetText();
auto* bss = builder_->GetBss();
const uint8_t* rodata_begin = oat_file_->Begin();
const size_t rodata_size = oat_file_->GetOatHeader().GetExecutableOffset();
if (no_bits_) {
rodata->WriteNoBitsSection(rodata_size);
} else {
rodata->Start();
rodata->WriteFully(rodata_begin, rodata_size);
rodata->End();
}
const uint8_t* text_begin = oat_file_->Begin() + rodata_size;
const size_t text_size = oat_file_->End() - text_begin;
if (no_bits_) {
text->WriteNoBitsSection(text_size);
} else {
text->Start();
text->WriteFully(text_begin, text_size);
text->End();
}
if (oat_file_->BssSize() != 0) {
bss->WriteNoBitsSection(oat_file_->BssSize());
}
if (isa == kMips || isa == kMips64) {
builder_->WriteMIPSabiflagsSection();
}
builder_->PrepareDynamicSection(elf_file->GetPath(),
rodata_size,
text_size,
oat_file_->BssSize(),
oat_file_->BssMethodsOffset(),
oat_file_->BssRootsOffset());
builder_->WriteDynamicSection();
Walk();
for (const auto& trampoline : debug::MakeTrampolineInfos(oat_file_->GetOatHeader())) {
method_debug_infos_.push_back(trampoline);
}
debug::WriteDebugInfo(builder_.get(),
ArrayRef<const debug::MethodDebugInfo>(method_debug_infos_),
dwarf::DW_DEBUG_FRAME_FORMAT,
true /* write_oat_patches */);
builder_->End();
bool ret_value = builder_->Good();
builder_.reset();
output_stream.reset();
if (elf_file->FlushCloseOrErase() != 0) {
return false;
}
elf_file.reset();
return ret_value;
}
void Walk() {
std::vector<const OatFile::OatDexFile*> oat_dex_files = oat_file_->GetOatDexFiles();
for (size_t i = 0; i < oat_dex_files.size(); i++) {
const OatFile::OatDexFile* oat_dex_file = oat_dex_files[i];
CHECK(oat_dex_file != nullptr);
WalkOatDexFile(oat_dex_file);
}
}
void WalkOatDexFile(const OatFile::OatDexFile* oat_dex_file) {
std::string error_msg;
const DexFile* const dex_file = OpenDexFile(oat_dex_file, &error_msg);
if (dex_file == nullptr) {
return;
}
for (size_t class_def_index = 0;
class_def_index < dex_file->NumClassDefs();
class_def_index++) {
const OatFile::OatClass oat_class = oat_dex_file->GetOatClass(class_def_index);
OatClassType type = oat_class.GetType();
switch (type) {
case kOatClassAllCompiled:
case kOatClassSomeCompiled:
WalkOatClass(oat_class, *dex_file, class_def_index);
break;
case kOatClassNoneCompiled:
case kOatClassMax:
// Ignore.
break;
}
}
}
void WalkOatClass(const OatFile::OatClass& oat_class,
const DexFile& dex_file,
uint32_t class_def_index) {
const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_def_index);
const uint8_t* class_data = dex_file.GetClassData(class_def);
if (class_data == nullptr) { // empty class such as a marker interface?
return;
}
// Note: even if this is an interface or a native class, we still have to walk it, as there
// might be a static initializer.
ClassDataItemIterator it(dex_file, class_data);
uint32_t class_method_idx = 0;
it.SkipAllFields();
for (; it.HasNextDirectMethod() || it.HasNextVirtualMethod(); it.Next()) {
WalkOatMethod(oat_class.GetOatMethod(class_method_idx++),
dex_file,
class_def_index,
it.GetMemberIndex(),
it.GetMethodCodeItem(),
it.GetMethodAccessFlags());
}
DCHECK(!it.HasNext());
}
void WalkOatMethod(const OatFile::OatMethod& oat_method,
const DexFile& dex_file,
uint32_t class_def_index,
uint32_t dex_method_index,
const DexFile::CodeItem* code_item,
uint32_t method_access_flags) {
if ((method_access_flags & kAccAbstract) != 0) {
// Abstract method, no code.
return;
}
const OatHeader& oat_header = oat_file_->GetOatHeader();
const OatQuickMethodHeader* method_header = oat_method.GetOatQuickMethodHeader();
if (method_header == nullptr || method_header->GetCodeSize() == 0) {
// No code.
return;
}
uint32_t entry_point = oat_method.GetCodeOffset() - oat_header.GetExecutableOffset();
// Clear Thumb2 bit.
const void* code_address = EntryPointToCodePointer(reinterpret_cast<void*>(entry_point));
debug::MethodDebugInfo info = debug::MethodDebugInfo();
info.trampoline_name = nullptr;
info.dex_file = &dex_file;
info.class_def_index = class_def_index;
info.dex_method_index = dex_method_index;
info.access_flags = method_access_flags;
info.code_item = code_item;
info.isa = oat_header.GetInstructionSet();
info.deduped = !seen_offsets_.insert(oat_method.GetCodeOffset()).second;
info.is_native_debuggable = oat_header.IsNativeDebuggable();
info.is_optimized = method_header->IsOptimized();
info.is_code_address_text_relative = true;
info.code_address = reinterpret_cast<uintptr_t>(code_address);
info.code_size = method_header->GetCodeSize();
info.frame_size_in_bytes = method_header->GetFrameSizeInBytes();
info.code_info = info.is_optimized ? method_header->GetOptimizedCodeInfoPtr() : nullptr;
info.cfi = ArrayRef<uint8_t>();
method_debug_infos_.push_back(info);
}
private:
const OatFile* oat_file_;
std::unique_ptr<ElfBuilder<ElfTypes> > builder_;
std::vector<debug::MethodDebugInfo> method_debug_infos_;
std::unordered_set<uint32_t> seen_offsets_;
const std::string output_name_;
bool no_bits_;
};
class OatDumperOptions {
public:
OatDumperOptions(bool dump_vmap,
bool dump_code_info_stack_maps,
bool disassemble_code,
bool absolute_addresses,
const char* class_filter,
const char* method_filter,
bool list_classes,
bool list_methods,
bool dump_header_only,
const char* export_dex_location,
const char* app_image,
const char* app_oat,
uint32_t addr2instr)
: dump_vmap_(dump_vmap),
dump_code_info_stack_maps_(dump_code_info_stack_maps),
disassemble_code_(disassemble_code),
absolute_addresses_(absolute_addresses),
class_filter_(class_filter),
method_filter_(method_filter),
list_classes_(list_classes),
list_methods_(list_methods),
dump_header_only_(dump_header_only),
export_dex_location_(export_dex_location),
app_image_(app_image),
app_oat_(app_oat),
addr2instr_(addr2instr),
class_loader_(nullptr) {}
const bool dump_vmap_;
const bool dump_code_info_stack_maps_;
const bool disassemble_code_;
const bool absolute_addresses_;
const char* const class_filter_;
const char* const method_filter_;
const bool list_classes_;
const bool list_methods_;
const bool dump_header_only_;
const char* const export_dex_location_;
const char* const app_image_;
const char* const app_oat_;
uint32_t addr2instr_;
Handle<mirror::ClassLoader>* class_loader_;
};
class OatDumper {
public:
OatDumper(const OatFile& oat_file, const OatDumperOptions& options)
: oat_file_(oat_file),
oat_dex_files_(oat_file.GetOatDexFiles()),
options_(options),
resolved_addr2instr_(0),
instruction_set_(oat_file_.GetOatHeader().GetInstructionSet()),
disassembler_(Disassembler::Create(instruction_set_,
new DisassemblerOptions(
options_.absolute_addresses_,
oat_file.Begin(),
oat_file.End(),
true /* can_read_literals_ */,
Is64BitInstructionSet(instruction_set_)
? &Thread::DumpThreadOffset<PointerSize::k64>
: &Thread::DumpThreadOffset<PointerSize::k32>))) {
CHECK(options_.class_loader_ != nullptr);
CHECK(options_.class_filter_ != nullptr);
CHECK(options_.method_filter_ != nullptr);
AddAllOffsets();
}
~OatDumper() {
delete disassembler_;
}
InstructionSet GetInstructionSet() {
return instruction_set_;
}
bool Dump(std::ostream& os) {
bool success = true;
const OatHeader& oat_header = oat_file_.GetOatHeader();
os << "MAGIC:\n";
os << oat_header.GetMagic() << "\n\n";
os << "LOCATION:\n";
os << oat_file_.GetLocation() << "\n\n";
os << "CHECKSUM:\n";
os << StringPrintf("0x%08x\n\n", oat_header.GetChecksum());
os << "INSTRUCTION SET:\n";
os << oat_header.GetInstructionSet() << "\n\n";
{
std::unique_ptr<const InstructionSetFeatures> features(
InstructionSetFeatures::FromBitmap(oat_header.GetInstructionSet(),
oat_header.GetInstructionSetFeaturesBitmap()));
os << "INSTRUCTION SET FEATURES:\n";
os << features->GetFeatureString() << "\n\n";
}
os << "DEX FILE COUNT:\n";
os << oat_header.GetDexFileCount() << "\n\n";
#define DUMP_OAT_HEADER_OFFSET(label, offset) \
os << label " OFFSET:\n"; \
os << StringPrintf("0x%08x", oat_header.offset()); \
if (oat_header.offset() != 0 && options_.absolute_addresses_) { \
os << StringPrintf(" (%p)", oat_file_.Begin() + oat_header.offset()); \
} \
os << StringPrintf("\n\n");
DUMP_OAT_HEADER_OFFSET("EXECUTABLE", GetExecutableOffset);
DUMP_OAT_HEADER_OFFSET("INTERPRETER TO INTERPRETER BRIDGE",
GetInterpreterToInterpreterBridgeOffset);
DUMP_OAT_HEADER_OFFSET("INTERPRETER TO COMPILED CODE BRIDGE",
GetInterpreterToCompiledCodeBridgeOffset);
DUMP_OAT_HEADER_OFFSET("JNI DLSYM LOOKUP",
GetJniDlsymLookupOffset);
DUMP_OAT_HEADER_OFFSET("QUICK GENERIC JNI TRAMPOLINE",
GetQuickGenericJniTrampolineOffset);
DUMP_OAT_HEADER_OFFSET("QUICK IMT CONFLICT TRAMPOLINE",
GetQuickImtConflictTrampolineOffset);
DUMP_OAT_HEADER_OFFSET("QUICK RESOLUTION TRAMPOLINE",
GetQuickResolutionTrampolineOffset);
DUMP_OAT_HEADER_OFFSET("QUICK TO INTERPRETER BRIDGE",
GetQuickToInterpreterBridgeOffset);
#undef DUMP_OAT_HEADER_OFFSET
os << "IMAGE PATCH DELTA:\n";
os << StringPrintf("%d (0x%08x)\n\n",
oat_header.GetImagePatchDelta(),
oat_header.GetImagePatchDelta());
os << "IMAGE FILE LOCATION OAT CHECKSUM:\n";
os << StringPrintf("0x%08x\n\n", oat_header.GetImageFileLocationOatChecksum());
os << "IMAGE FILE LOCATION OAT BEGIN:\n";
os << StringPrintf("0x%08x\n\n", oat_header.GetImageFileLocationOatDataBegin());
// Print the key-value store.
{
os << "KEY VALUE STORE:\n";
size_t index = 0;
const char* key;
const char* value;
while (oat_header.GetStoreKeyValuePairByIndex(index, &key, &value)) {
os << key << " = " << value << "\n";
index++;
}
os << "\n";
}
if (options_.absolute_addresses_) {
os << "BEGIN:\n";
os << reinterpret_cast<const void*>(oat_file_.Begin()) << "\n\n";
os << "END:\n";
os << reinterpret_cast<const void*>(oat_file_.End()) << "\n\n";
}
os << "SIZE:\n";
os << oat_file_.Size() << "\n\n";
os << std::flush;
// If set, adjust relative address to be searched
if (options_.addr2instr_ != 0) {
resolved_addr2instr_ = options_.addr2instr_ + oat_header.GetExecutableOffset();
os << "SEARCH ADDRESS (executable offset + input):\n";
os << StringPrintf("0x%08x\n\n", resolved_addr2instr_);
}
// Dumping the dex file overview is compact enough to do even if header only.
DexFileData cumulative;
for (size_t i = 0; i < oat_dex_files_.size(); i++) {
const OatFile::OatDexFile* oat_dex_file = oat_dex_files_[i];
CHECK(oat_dex_file != nullptr);
std::string error_msg;
const DexFile* const dex_file = OpenDexFile(oat_dex_file, &error_msg);
if (dex_file == nullptr) {
os << "Failed to open dex file '" << oat_dex_file->GetDexFileLocation() << "': "
<< error_msg;
continue;
}
DexFileData data(*dex_file);
os << "Dex file data for " << dex_file->GetLocation() << "\n";
data.Dump(os);
os << "\n";
cumulative.Add(data);
}
os << "Cumulative dex file data\n";
cumulative.Dump(os);
os << "\n";
if (!options_.dump_header_only_) {
VariableIndentationOutputStream vios(&os);
VdexFile::Header vdex_header = oat_file_.GetVdexFile()->GetHeader();
if (vdex_header.IsValid()) {
std::string error_msg;
std::vector<const DexFile*> dex_files;
for (size_t i = 0; i < oat_dex_files_.size(); i++) {
const DexFile* dex_file = OpenDexFile(oat_dex_files_[i], &error_msg);
if (dex_file == nullptr) {
os << "Error opening dex file: " << error_msg << std::endl;
return false;
}
dex_files.push_back(dex_file);
}
verifier::VerifierDeps deps(dex_files, oat_file_.GetVdexFile()->GetVerifierDepsData());
deps.Dump(&vios);
} else {
os << "UNRECOGNIZED vdex file, magic "
<< vdex_header.GetMagic()
<< ", version "
<< vdex_header.GetVersion()
<< "\n";
}
for (size_t i = 0; i < oat_dex_files_.size(); i++) {
const OatFile::OatDexFile* oat_dex_file = oat_dex_files_[i];
CHECK(oat_dex_file != nullptr);
// If file export selected skip file analysis
if (options_.export_dex_location_) {
if (!ExportDexFile(os, *oat_dex_file)) {
success = false;
}
} else {
if (!DumpOatDexFile(os, *oat_dex_file)) {
success = false;
}
}
}
}
{
os << "OAT FILE STATS:\n";
VariableIndentationOutputStream vios(&os);
stats_.Dump(vios);
}
os << std::flush;
return success;
}
size_t ComputeSize(const void* oat_data) {
if (reinterpret_cast<const uint8_t*>(oat_data) < oat_file_.Begin() ||
reinterpret_cast<const uint8_t*>(oat_data) > oat_file_.End()) {
return 0; // Address not in oat file
}
uintptr_t begin_offset = reinterpret_cast<uintptr_t>(oat_data) -
reinterpret_cast<uintptr_t>(oat_file_.Begin());
auto it = offsets_.upper_bound(begin_offset);
CHECK(it != offsets_.end());
uintptr_t end_offset = *it;
return end_offset - begin_offset;
}
InstructionSet GetOatInstructionSet() {
return oat_file_.GetOatHeader().GetInstructionSet();
}
const void* GetQuickOatCode(ArtMethod* m) REQUIRES_SHARED(Locks::mutator_lock_) {
for (size_t i = 0; i < oat_dex_files_.size(); i++) {
const OatFile::OatDexFile* oat_dex_file = oat_dex_files_[i];
CHECK(oat_dex_file != nullptr);
std::string error_msg;
const DexFile* const dex_file = OpenDexFile(oat_dex_file, &error_msg);
if (dex_file == nullptr) {
LOG(WARNING) << "Failed to open dex file '" << oat_dex_file->GetDexFileLocation()
<< "': " << error_msg;
} else {
const char* descriptor = m->GetDeclaringClassDescriptor();
const DexFile::ClassDef* class_def =
OatDexFile::FindClassDef(*dex_file, descriptor, ComputeModifiedUtf8Hash(descriptor));
if (class_def != nullptr) {
uint16_t class_def_index = dex_file->GetIndexForClassDef(*class_def);
const OatFile::OatClass oat_class = oat_dex_file->GetOatClass(class_def_index);
size_t method_index = m->GetMethodIndex();
return oat_class.GetOatMethod(method_index).GetQuickCode();
}
}
}
return nullptr;
}
struct Stats {
enum ByteKind {
kByteKindCode,
kByteKindQuickMethodHeader,
kByteKindCodeInfoLocationCatalog,
kByteKindCodeInfoDexRegisterMap,
kByteKindCodeInfoEncoding,
kByteKindCodeInfoInvokeInfo,
kByteKindCodeInfoStackMasks,
kByteKindCodeInfoRegisterMasks,
kByteKindStackMapNativePc,
kByteKindStackMapDexPc,
kByteKindStackMapDexRegisterMap,
kByteKindStackMapInlineInfoIndex,
kByteKindStackMapRegisterMaskIndex,
kByteKindStackMapStackMaskIndex,
kByteKindInlineInfoMethodIndexIdx,
kByteKindInlineInfoDexPc,
kByteKindInlineInfoExtraData,
kByteKindInlineInfoDexRegisterMap,
kByteKindInlineInfoIsLast,
kByteKindCount,
// Special ranges for std::accumulate convenience.
kByteKindStackMapFirst = kByteKindStackMapNativePc,
kByteKindStackMapLast = kByteKindStackMapStackMaskIndex,
kByteKindInlineInfoFirst = kByteKindInlineInfoMethodIndexIdx,
kByteKindInlineInfoLast = kByteKindInlineInfoIsLast,
};
int64_t bits[kByteKindCount] = {};
// Since code has deduplication, seen tracks already seen pointers to avoid double counting
// deduplicated code and tables.
std::unordered_set<const void*> seen;
// Returns true if it was newly added.
bool AddBitsIfUnique(ByteKind kind, int64_t count, const void* address) {
if (seen.insert(address).second == true) {
// True means the address was not already in the set.
AddBits(kind, count);
return true;
}
return false;
}
void AddBits(ByteKind kind, int64_t count) {
bits[kind] += count;
}
void Dump(VariableIndentationOutputStream& os) {
const int64_t sum = std::accumulate(bits, bits + kByteKindCount, 0u);
os.Stream() << "Dumping cumulative use of " << sum / kBitsPerByte << " accounted bytes\n";
if (sum > 0) {
Dump(os, "Code ", bits[kByteKindCode], sum);
Dump(os, "QuickMethodHeader ", bits[kByteKindQuickMethodHeader], sum);
Dump(os, "CodeInfoEncoding ", bits[kByteKindCodeInfoEncoding], sum);
Dump(os, "CodeInfoLocationCatalog ", bits[kByteKindCodeInfoLocationCatalog], sum);
Dump(os, "CodeInfoDexRegisterMap ", bits[kByteKindCodeInfoDexRegisterMap], sum);
Dump(os, "CodeInfoStackMasks ", bits[kByteKindCodeInfoStackMasks], sum);
Dump(os, "CodeInfoRegisterMasks ", bits[kByteKindCodeInfoRegisterMasks], sum);
Dump(os, "CodeInfoInvokeInfo ", bits[kByteKindCodeInfoInvokeInfo], sum);
// Stack map section.
const int64_t stack_map_bits = std::accumulate(bits + kByteKindStackMapFirst,
bits + kByteKindStackMapLast + 1,
0u);
Dump(os, "CodeInfoStackMap ", stack_map_bits, sum);
{
ScopedIndentation indent1(&os);
Dump(os,
"StackMapNativePc ",
bits[kByteKindStackMapNativePc],
stack_map_bits,
"stack map");
Dump(os,
"StackMapDexPcEncoding ",
bits[kByteKindStackMapDexPc],
stack_map_bits,
"stack map");
Dump(os,
"StackMapDexRegisterMap ",
bits[kByteKindStackMapDexRegisterMap],
stack_map_bits,
"stack map");
Dump(os,
"StackMapInlineInfoIndex ",
bits[kByteKindStackMapInlineInfoIndex],
stack_map_bits,
"stack map");
Dump(os,
"StackMapRegisterMaskIndex ",
bits[kByteKindStackMapRegisterMaskIndex],
stack_map_bits,
"stack map");
Dump(os,
"StackMapStackMaskIndex ",
bits[kByteKindStackMapStackMaskIndex],
stack_map_bits,
"stack map");
}
// Inline info section.
const int64_t inline_info_bits = std::accumulate(bits + kByteKindInlineInfoFirst,
bits + kByteKindInlineInfoLast + 1,
0u);
Dump(os, "CodeInfoInlineInfo ", inline_info_bits, sum);
{
ScopedIndentation indent1(&os);
Dump(os,
"InlineInfoMethodIndexIdx ",
bits[kByteKindInlineInfoMethodIndexIdx],
inline_info_bits,
"inline info");
Dump(os,
"InlineInfoDexPc ",
bits[kByteKindStackMapDexPc],
inline_info_bits,
"inline info");
Dump(os,
"InlineInfoExtraData ",
bits[kByteKindInlineInfoExtraData],
inline_info_bits,
"inline info");
Dump(os,
"InlineInfoDexRegisterMap ",
bits[kByteKindInlineInfoDexRegisterMap],
inline_info_bits,
"inline info");
Dump(os,
"InlineInfoIsLast ",
bits[kByteKindInlineInfoIsLast],
inline_info_bits,
"inline info");
}
}
os.Stream() << "\n" << std::flush;
}
private:
void Dump(VariableIndentationOutputStream& os,
const char* name,
int64_t size,
int64_t total,
const char* sum_of = "total") {
const double percent = (static_cast<double>(size) / static_cast<double>(total)) * 100;
os.Stream() << StringPrintf("%s = %8" PRId64 " (%2.0f%% of %s)\n",
name,
size / kBitsPerByte,
percent,
sum_of);
}
};
private:
void AddAllOffsets() {
// We don't know the length of the code for each method, but we need to know where to stop
// when disassembling. What we do know is that a region of code will be followed by some other
// region, so if we keep a sorted sequence of the start of each region, we can infer the length
// of a piece of code by using upper_bound to find the start of the next region.
for (size_t i = 0; i < oat_dex_files_.size(); i++) {
const OatFile::OatDexFile* oat_dex_file = oat_dex_files_[i];
CHECK(oat_dex_file != nullptr);
std::string error_msg;
const DexFile* const dex_file = OpenDexFile(oat_dex_file, &error_msg);
if (dex_file == nullptr) {
LOG(WARNING) << "Failed to open dex file '" << oat_dex_file->GetDexFileLocation()
<< "': " << error_msg;
continue;
}
offsets_.insert(reinterpret_cast<uintptr_t>(&dex_file->GetHeader()));
for (size_t class_def_index = 0;
class_def_index < dex_file->NumClassDefs();
class_def_index++) {
const DexFile::ClassDef& class_def = dex_file->GetClassDef(class_def_index);
const OatFile::OatClass oat_class = oat_dex_file->GetOatClass(class_def_index);
const uint8_t* class_data = dex_file->GetClassData(class_def);
if (class_data != nullptr) {
ClassDataItemIterator it(*dex_file, class_data);
it.SkipAllFields();
uint32_t class_method_index = 0;
while (it.HasNextDirectMethod()) {
AddOffsets(oat_class.GetOatMethod(class_method_index++));
it.Next();
}
while (it.HasNextVirtualMethod()) {
AddOffsets(oat_class.GetOatMethod(class_method_index++));
it.Next();
}
}
}
}
// If the last thing in the file is code for a method, there won't be an offset for the "next"
// thing. Instead of having a special case in the upper_bound code, let's just add an entry
// for the end of the file.
offsets_.insert(oat_file_.Size());
}
static uint32_t AlignCodeOffset(uint32_t maybe_thumb_offset) {
return maybe_thumb_offset & ~0x1; // TODO: Make this Thumb2 specific.
}
void AddOffsets(const OatFile::OatMethod& oat_method) {
uint32_t code_offset = oat_method.GetCodeOffset();
if (oat_file_.GetOatHeader().GetInstructionSet() == kThumb2) {
code_offset &= ~0x1;
}
offsets_.insert(code_offset);
offsets_.insert(oat_method.GetVmapTableOffset());
}
// Dex file data, may be for multiple different dex files.
class DexFileData {
public:
DexFileData() {}
explicit DexFileData(const DexFile& dex_file)
: num_string_ids_(dex_file.NumStringIds()),
num_method_ids_(dex_file.NumMethodIds()),
num_field_ids_(dex_file.NumFieldIds()),
num_type_ids_(dex_file.NumTypeIds()),
num_class_defs_(dex_file.NumClassDefs()) {
for (size_t class_def_index = 0; class_def_index < num_class_defs_; ++class_def_index) {
const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_def_index);
WalkClass(dex_file, class_def);
}
}
void Add(const DexFileData& other) {
AddAll(unique_string_ids_from_code_, other.unique_string_ids_from_code_);
num_string_ids_from_code_ += other.num_string_ids_from_code_;
AddAll(dex_code_item_ptrs_, other.dex_code_item_ptrs_);
dex_code_bytes_ += other.dex_code_bytes_;
num_string_ids_ += other.num_string_ids_;
num_method_ids_ += other.num_method_ids_;
num_field_ids_ += other.num_field_ids_;
num_type_ids_ += other.num_type_ids_;
num_class_defs_ += other.num_class_defs_;
}
void Dump(std::ostream& os) {
os << "Num string ids: " << num_string_ids_ << "\n";
os << "Num method ids: " << num_method_ids_ << "\n";
os << "Num field ids: " << num_field_ids_ << "\n";
os << "Num type ids: " << num_type_ids_ << "\n";
os << "Num class defs: " << num_class_defs_ << "\n";
os << "Unique strings loaded from dex code: " << unique_string_ids_from_code_.size() << "\n";
os << "Total strings loaded from dex code: " << num_string_ids_from_code_ << "\n";
os << "Number of unique dex code items: " << dex_code_item_ptrs_.size() << "\n";
os << "Total number of dex code bytes: " << dex_code_bytes_ << "\n";
}
private:
// All of the elements from one container to another.
template <typename Dest, typename Src>
static void AddAll(Dest& dest, const Src& src) {
dest.insert(src.begin(), src.end());
}
void WalkClass(const DexFile& dex_file, const DexFile::ClassDef& class_def) {
const uint8_t* class_data = dex_file.GetClassData(class_def);
if (class_data == nullptr) { // empty class such as a marker interface?
return;
}
ClassDataItemIterator it(dex_file, class_data);
it.SkipAllFields();
while (it.HasNextDirectMethod()) {
WalkCodeItem(dex_file, it.GetMethodCodeItem());
it.Next();
}
while (it.HasNextVirtualMethod()) {
WalkCodeItem(dex_file, it.GetMethodCodeItem());
it.Next();
}
DCHECK(!it.HasNext());
}
void WalkCodeItem(const DexFile& dex_file, const DexFile::CodeItem* code_item) {
if (code_item == nullptr) {
return;
}
const size_t code_item_size = code_item->insns_size_in_code_units_;
const uint16_t* code_ptr = code_item->insns_;
const uint16_t* code_end = code_item->insns_ + code_item_size;
// If we inserted a new dex code item pointer, add to total code bytes.
if (dex_code_item_ptrs_.insert(code_ptr).second) {
dex_code_bytes_ += code_item_size * sizeof(code_ptr[0]);
}
while (code_ptr < code_end) {
const Instruction* inst = Instruction::At(code_ptr);
switch (inst->Opcode()) {
case Instruction::CONST_STRING: {
const dex::StringIndex string_index(inst->VRegB_21c());
unique_string_ids_from_code_.insert(StringReference(&dex_file, string_index));
++num_string_ids_from_code_;
break;
}
case Instruction::CONST_STRING_JUMBO: {
const dex::StringIndex string_index(inst->VRegB_31c());
unique_string_ids_from_code_.insert(StringReference(&dex_file, string_index));
++num_string_ids_from_code_;
break;
}
default:
break;
}
code_ptr += inst->SizeInCodeUnits();
}
}
// Unique string ids loaded from dex code.
std::set<StringReference, StringReferenceComparator> unique_string_ids_from_code_;
// Total string ids loaded from dex code.
size_t num_string_ids_from_code_ = 0;
// Unique code pointers.
std::set<const void*> dex_code_item_ptrs_;
// Total "unique" dex code bytes.
size_t dex_code_bytes_ = 0;
// Other dex ids.
size_t num_string_ids_ = 0;
size_t num_method_ids_ = 0;
size_t num_field_ids_ = 0;
size_t num_type_ids_ = 0;
size_t num_class_defs_ = 0;
};
bool DumpOatDexFile(std::ostream& os, const OatFile::OatDexFile& oat_dex_file) {
bool success = true;
bool stop_analysis = false;
os << "OatDexFile:\n";
os << StringPrintf("location: %s\n", oat_dex_file.GetDexFileLocation().c_str());
os << StringPrintf("checksum: 0x%08x\n", oat_dex_file.GetDexFileLocationChecksum());
const uint8_t* const oat_file_begin = oat_dex_file.GetOatFile()->Begin();
const uint8_t* const vdex_file_begin = oat_dex_file.GetOatFile()->DexBegin();
// Print data range of the dex file embedded inside the corresponding vdex file.
const uint8_t* const dex_file_pointer = oat_dex_file.GetDexFilePointer();
uint32_t dex_offset = dchecked_integral_cast<uint32_t>(dex_file_pointer - vdex_file_begin);
os << StringPrintf("dex-file: 0x%08x..0x%08x\n",
dex_offset,
dchecked_integral_cast<uint32_t>(dex_offset + oat_dex_file.FileSize() - 1));
// Create the dex file early. A lot of print-out things depend on it.
std::string error_msg;
const DexFile* const dex_file = OpenDexFile(&oat_dex_file, &error_msg);
if (dex_file == nullptr) {
os << "NOT FOUND: " << error_msg << "\n\n";
os << std::flush;
return false;
}
// Print lookup table, if it exists.
if (oat_dex_file.GetLookupTableData() != nullptr) {
uint32_t table_offset = dchecked_integral_cast<uint32_t>(
oat_dex_file.GetLookupTableData() - oat_file_begin);
uint32_t table_size = TypeLookupTable::RawDataLength(dex_file->NumClassDefs());
os << StringPrintf("type-table: 0x%08x..0x%08x\n",
table_offset,
table_offset + table_size - 1);
}
VariableIndentationOutputStream vios(&os);
ScopedIndentation indent1(&vios);
for (size_t class_def_index = 0;
class_def_index < dex_file->NumClassDefs();
class_def_index++) {
const DexFile::ClassDef& class_def = dex_file->GetClassDef(class_def_index);
const char* descriptor = dex_file->GetClassDescriptor(class_def);
// TODO: Support regex
if (DescriptorToDot(descriptor).find(options_.class_filter_) == std::string::npos) {
continue;
}
uint32_t oat_class_offset = oat_dex_file.GetOatClassOffset(class_def_index);
const OatFile::OatClass oat_class = oat_dex_file.GetOatClass(class_def_index);
os << StringPrintf("%zd: %s (offset=0x%08x) (type_idx=%d)",
class_def_index, descriptor, oat_class_offset, class_def.class_idx_.index_)
<< " (" << oat_class.GetStatus() << ")"
<< " (" << oat_class.GetType() << ")\n";
// TODO: include bitmap here if type is kOatClassSomeCompiled?
if (options_.list_classes_) {
continue;
}
if (!DumpOatClass(&vios, oat_class, *dex_file, class_def, &stop_analysis)) {
success = false;
}
if (stop_analysis) {
os << std::flush;
return success;
}
}
os << "\n";
os << std::flush;
return success;
}
bool ExportDexFile(std::ostream& os, const OatFile::OatDexFile& oat_dex_file) {
std::string error_msg;
std::string dex_file_location = oat_dex_file.GetDexFileLocation();
const DexFile* const dex_file = OpenDexFile(&oat_dex_file, &error_msg);
if (dex_file == nullptr) {
os << "Failed to open dex file '" << dex_file_location << "': " << error_msg;
return false;
}
size_t fsize = oat_dex_file.FileSize();
// Some quick checks just in case
if (fsize == 0 || fsize < sizeof(DexFile::Header)) {
os << "Invalid dex file\n";
return false;
}
// Verify output directory exists
if (!OS::DirectoryExists(options_.export_dex_location_)) {
// TODO: Extend OS::DirectoryExists if symlink support is required
os << options_.export_dex_location_ << " output directory not found or symlink\n";
return false;
}
// Beautify path names
if (dex_file_location.size() > PATH_MAX || dex_file_location.size() <= 0) {
return false;
}
std::string dex_orig_name;
size_t dex_orig_pos = dex_file_location.rfind('/');
if (dex_orig_pos == std::string::npos)
dex_orig_name = dex_file_location;
else
dex_orig_name = dex_file_location.substr(dex_orig_pos + 1);
// A more elegant approach to efficiently name user installed apps is welcome
if (dex_orig_name.size() == 8 &&
dex_orig_name.compare("base.apk") == 0 &&
dex_orig_pos != std::string::npos) {
dex_file_location.erase(dex_orig_pos, strlen("base.apk") + 1);
size_t apk_orig_pos = dex_file_location.rfind('/');
if (apk_orig_pos != std::string::npos) {
dex_orig_name = dex_file_location.substr(++apk_orig_pos);
}
}
std::string out_dex_path(options_.export_dex_location_);
if (out_dex_path.back() != '/') {
out_dex_path.append("/");
}
out_dex_path.append(dex_orig_name);
out_dex_path.append("_export.dex");
if (out_dex_path.length() > PATH_MAX) {
return false;
}
std::unique_ptr<File> file(OS::CreateEmptyFile(out_dex_path.c_str()));
if (file.get() == nullptr) {
os << "Failed to open output dex file " << out_dex_path;
return false;
}
if (!file->WriteFully(dex_file->Begin(), fsize)) {
os << "Failed to write dex file";
file->Erase();
return false;
}
if (file->FlushCloseOrErase() != 0) {
os << "Flush and close failed";
return false;
}
os << StringPrintf("Dex file exported at %s (%zd bytes)\n", out_dex_path.c_str(), fsize);
os << std::flush;
return true;
}
bool DumpOatClass(VariableIndentationOutputStream* vios,
const OatFile::OatClass& oat_class, const DexFile& dex_file,
const DexFile::ClassDef& class_def, bool* stop_analysis) {
bool success = true;
bool addr_found = false;
const uint8_t* class_data = dex_file.GetClassData(class_def);
if (class_data == nullptr) { // empty class such as a marker interface?
vios->Stream() << std::flush;
return success;
}
ClassDataItemIterator it(dex_file, class_data);
it.SkipAllFields();
uint32_t class_method_index = 0;
while (it.HasNextDirectMethod()) {
if (!DumpOatMethod(vios, class_def, class_method_index, oat_class, dex_file,
it.GetMemberIndex(), it.GetMethodCodeItem(),
it.GetRawMemberAccessFlags(), &addr_found)) {
success = false;
}
if (addr_found) {
*stop_analysis = true;
return success;
}
class_method_index++;
it.Next();
}
while (it.HasNextVirtualMethod()) {
if (!DumpOatMethod(vios, class_def, class_method_index, oat_class, dex_file,
it.GetMemberIndex(), it.GetMethodCodeItem(),
it.GetRawMemberAccessFlags(), &addr_found)) {
success = false;
}
if (addr_found) {
*stop_analysis = true;
return success;
}
class_method_index++;
it.Next();
}
DCHECK(!it.HasNext());
vios->Stream() << std::flush;
return success;
}
static constexpr uint32_t kPrologueBytes = 16;
// When this was picked, the largest arm method was 55,256 bytes and arm64 was 50,412 bytes.
static constexpr uint32_t kMaxCodeSize = 100 * 1000;
bool DumpOatMethod(VariableIndentationOutputStream* vios,
const DexFile::ClassDef& class_def,
uint32_t class_method_index,
const OatFile::OatClass& oat_class, const DexFile& dex_file,
uint32_t dex_method_idx, const DexFile::CodeItem* code_item,
uint32_t method_access_flags, bool* addr_found) {
bool success = true;
// TODO: Support regex
std::string method_name = dex_file.GetMethodName(dex_file.GetMethodId(dex_method_idx));
if (method_name.find(options_.method_filter_) == std::string::npos) {
return success;
}
std::string pretty_method = dex_file.PrettyMethod(dex_method_idx, true);
vios->Stream() << StringPrintf("%d: %s (dex_method_idx=%d)\n",
class_method_index, pretty_method.c_str(),
dex_method_idx);
if (options_.list_methods_) return success;
uint32_t oat_method_offsets_offset = oat_class.GetOatMethodOffsetsOffset(class_method_index);
const OatMethodOffsets* oat_method_offsets = oat_class.GetOatMethodOffsets(class_method_index);
const OatFile::OatMethod oat_method = oat_class.GetOatMethod(class_method_index);
uint32_t code_offset = oat_method.GetCodeOffset();
uint32_t code_size = oat_method.GetQuickCodeSize();
if (resolved_addr2instr_ != 0) {
if (resolved_addr2instr_ > code_offset + code_size) {
return success;
} else {
*addr_found = true; // stop analyzing file at next iteration
}
}
// Everything below is indented at least once.
ScopedIndentation indent1(vios);
{
vios->Stream() << "DEX CODE:\n";
ScopedIndentation indent2(vios);
DumpDexCode(vios->Stream(), dex_file, code_item);
}
std::unique_ptr<StackHandleScope<1>> hs;
std::unique_ptr<verifier::MethodVerifier> verifier;
if (Runtime::Current() != nullptr) {
// We need to have the handle scope stay live until after the verifier since the verifier has
// a handle to the dex cache from hs.
hs.reset(new StackHandleScope<1>(Thread::Current()));
vios->Stream() << "VERIFIER TYPE ANALYSIS:\n";
ScopedIndentation indent2(vios);
verifier.reset(DumpVerifier(vios, hs.get(),
dex_method_idx, &dex_file, class_def, code_item,
method_access_flags));
}
{
vios->Stream() << "OatMethodOffsets ";
if (options_.absolute_addresses_) {
vios->Stream() << StringPrintf("%p ", oat_method_offsets);
}
vios->Stream() << StringPrintf("(offset=0x%08x)\n", oat_method_offsets_offset);
if (oat_method_offsets_offset > oat_file_.Size()) {
vios->Stream() << StringPrintf(
"WARNING: oat method offsets offset 0x%08x is past end of file 0x%08zx.\n",
oat_method_offsets_offset, oat_file_.Size());
// If we can't read OatMethodOffsets, the rest of the data is dangerous to read.
vios->Stream() << std::flush;
return false;
}
ScopedIndentation indent2(vios);
vios->Stream() << StringPrintf("code_offset: 0x%08x ", code_offset);
uint32_t aligned_code_begin = AlignCodeOffset(oat_method.GetCodeOffset());
if (aligned_code_begin > oat_file_.Size()) {
vios->Stream() << StringPrintf("WARNING: "
"code offset 0x%08x is past end of file 0x%08zx.\n",
aligned_code_begin, oat_file_.Size());
success = false;
}
vios->Stream() << "\n";
}
{
vios->Stream() << "OatQuickMethodHeader ";
uint32_t method_header_offset = oat_method.GetOatQuickMethodHeaderOffset();
const OatQuickMethodHeader* method_header = oat_method.GetOatQuickMethodHeader();
stats_.AddBitsIfUnique(Stats::kByteKindQuickMethodHeader,
sizeof(*method_header) * kBitsPerByte,
method_header);
if (options_.absolute_addresses_) {
vios->Stream() << StringPrintf("%p ", method_header);
}
vios->Stream() << StringPrintf("(offset=0x%08x)\n", method_header_offset);
if (method_header_offset > oat_file_.Size()) {
vios->Stream() << StringPrintf(
"WARNING: oat quick method header offset 0x%08x is past end of file 0x%08zx.\n",
method_header_offset, oat_file_.Size());
// If we can't read the OatQuickMethodHeader, the rest of the data is dangerous to read.
vios->Stream() << std::flush;
return false;
}
ScopedIndentation indent2(vios);
vios->Stream() << "vmap_table: ";
if (options_.absolute_addresses_) {
vios->Stream() << StringPrintf("%p ", oat_method.GetVmapTable());
}
uint32_t vmap_table_offset = method_header ==
nullptr ? 0 : method_header->GetVmapTableOffset();
vios->Stream() << StringPrintf("(offset=0x%08x)\n", vmap_table_offset);
size_t vmap_table_offset_limit =
(kIsVdexEnabled && IsMethodGeneratedByDexToDexCompiler(oat_method, code_item))
? oat_file_.GetVdexFile()->Size()
: method_header->GetCode() - oat_file_.Begin();
if (vmap_table_offset >= vmap_table_offset_limit) {
vios->Stream() << StringPrintf("WARNING: "
"vmap table offset 0x%08x is past end of file 0x%08zx. "
"vmap table offset was loaded from offset 0x%08x.\n",
vmap_table_offset,
vmap_table_offset_limit,
oat_method.GetVmapTableOffsetOffset());
success = false;
} else if (options_.dump_vmap_) {
DumpVmapData(vios, oat_method, code_item);
}
}
{
vios->Stream() << "QuickMethodFrameInfo\n";
ScopedIndentation indent2(vios);
vios->Stream()
<< StringPrintf("frame_size_in_bytes: %zd\n", oat_method.GetFrameSizeInBytes());
vios->Stream() << StringPrintf("core_spill_mask: 0x%08x ", oat_method.GetCoreSpillMask());
DumpSpillMask(vios->Stream(), oat_method.GetCoreSpillMask(), false);
vios->Stream() << "\n";
vios->Stream() << StringPrintf("fp_spill_mask: 0x%08x ", oat_method.GetFpSpillMask());
DumpSpillMask(vios->Stream(), oat_method.GetFpSpillMask(), true);
vios->Stream() << "\n";
}
{
// Based on spill masks from QuickMethodFrameInfo so placed
// after it is dumped, but useful for understanding quick
// code, so dumped here.
ScopedIndentation indent2(vios);
DumpVregLocations(vios->Stream(), oat_method, code_item);
}
{
vios->Stream() << "CODE: ";
uint32_t code_size_offset = oat_method.GetQuickCodeSizeOffset();
if (code_size_offset > oat_file_.Size()) {
ScopedIndentation indent2(vios);
vios->Stream() << StringPrintf("WARNING: "
"code size offset 0x%08x is past end of file 0x%08zx.",
code_size_offset, oat_file_.Size());
success = false;
} else {
const void* code = oat_method.GetQuickCode();
uint32_t aligned_code_begin = AlignCodeOffset(code_offset);
uint64_t aligned_code_end = aligned_code_begin + code_size;
stats_.AddBitsIfUnique(Stats::kByteKindCode, code_size * kBitsPerByte, code);
if (options_.absolute_addresses_) {
vios->Stream() << StringPrintf("%p ", code);
}
vios->Stream() << StringPrintf("(code_offset=0x%08x size_offset=0x%08x size=%u)%s\n",
code_offset,
code_size_offset,
code_size,
code != nullptr ? "..." : "");
ScopedIndentation indent2(vios);
if (aligned_code_begin > oat_file_.Size()) {
vios->Stream() << StringPrintf("WARNING: "
"start of code at 0x%08x is past end of file 0x%08zx.",
aligned_code_begin, oat_file_.Size());
success = false;
} else if (aligned_code_end > oat_file_.Size()) {
vios->Stream() << StringPrintf(
"WARNING: "
"end of code at 0x%08" PRIx64 " is past end of file 0x%08zx. "
"code size is 0x%08x loaded from offset 0x%08x.\n",
aligned_code_end, oat_file_.Size(),
code_size, code_size_offset);
success = false;
if (options_.disassemble_code_) {
if (code_size_offset + kPrologueBytes <= oat_file_.Size()) {
DumpCode(vios, oat_method, code_item, true, kPrologueBytes);
}
}
} else if (code_size > kMaxCodeSize) {
vios->Stream() << StringPrintf(
"WARNING: "
"code size %d is bigger than max expected threshold of %d. "
"code size is 0x%08x loaded from offset 0x%08x.\n",
code_size, kMaxCodeSize,
code_size, code_size_offset);
success = false;
if (options_.disassemble_code_) {
if (code_size_offset + kPrologueBytes <= oat_file_.Size()) {
DumpCode(vios, oat_method, code_item, true, kPrologueBytes);
}
}
} else if (options_.disassemble_code_) {
DumpCode(vios, oat_method, code_item, !success, 0);
}
}
}
vios->Stream() << std::flush;
return success;
}
void DumpSpillMask(std::ostream& os, uint32_t spill_mask, bool is_float) {
if (spill_mask == 0) {
return;
}
os << "(";
for (size_t i = 0; i < 32; i++) {
if ((spill_mask & (1 << i)) != 0) {
if (is_float) {
os << "fr" << i;
} else {
os << "r" << i;
}
spill_mask ^= 1 << i; // clear bit
if (spill_mask != 0) {
os << ", ";
} else {
break;
}
}
}
os << ")";
}
// Display data stored at the the vmap offset of an oat method.
void DumpVmapData(VariableIndentationOutputStream* vios,
const OatFile::OatMethod& oat_method,
const DexFile::CodeItem* code_item) {
if (IsMethodGeneratedByOptimizingCompiler(oat_method, code_item)) {
// The optimizing compiler outputs its CodeInfo data in the vmap table.
const void* raw_code_info = oat_method.GetVmapTable();
if (raw_code_info != nullptr) {
CodeInfo code_info(raw_code_info);
DCHECK(code_item != nullptr);
ScopedIndentation indent1(vios);
MethodInfo method_info = oat_method.GetOatQuickMethodHeader()->GetOptimizedMethodInfo();
DumpCodeInfo(vios, code_info, oat_method, *code_item, method_info);
}
} else if (IsMethodGeneratedByDexToDexCompiler(oat_method, code_item)) {
// We don't encode the size in the table, so just emit that we have quickened
// information.
ScopedIndentation indent(vios);
vios->Stream() << "quickened data\n";
} else {
// Otherwise, there is nothing to display.
}
}
// Display a CodeInfo object emitted by the optimizing compiler.
void DumpCodeInfo(VariableIndentationOutputStream* vios,
const CodeInfo& code_info,
const OatFile::OatMethod& oat_method,
const DexFile::CodeItem& code_item,
const MethodInfo& method_info) {
code_info.Dump(vios,
oat_method.GetCodeOffset(),
code_item.registers_size_,
options_.dump_code_info_stack_maps_,
instruction_set_,
method_info);
}
static int GetOutVROffset(uint16_t out_num, InstructionSet isa) {
// According to stack model, the first out is above the Method referernce.
return static_cast<size_t>(InstructionSetPointerSize(isa)) + out_num * sizeof(uint32_t);
}
static uint32_t GetVRegOffsetFromQuickCode(const DexFile::CodeItem* code_item,
uint32_t core_spills,
uint32_t fp_spills,
size_t frame_size,
int reg,
InstructionSet isa) {
PointerSize pointer_size = InstructionSetPointerSize(isa);
if (kIsDebugBuild) {
auto* runtime = Runtime::Current();
if (runtime != nullptr) {
CHECK_EQ(runtime->GetClassLinker()->GetImagePointerSize(), pointer_size);
}
}
DCHECK_ALIGNED(frame_size, kStackAlignment);
DCHECK_NE(reg, -1);
int spill_size = POPCOUNT(core_spills) * GetBytesPerGprSpillLocation(isa)
+ POPCOUNT(fp_spills) * GetBytesPerFprSpillLocation(isa)
+ sizeof(uint32_t); // Filler.
int num_regs = code_item->registers_size_ - code_item->ins_size_;
int temp_threshold = code_item->registers_size_;
const int max_num_special_temps = 1;
if (reg == temp_threshold) {
// The current method pointer corresponds to special location on stack.
return 0;
} else if (reg >= temp_threshold + max_num_special_temps) {
/*
* Special temporaries may have custom locations and the logic above deals with that.
* However, non-special temporaries are placed relative to the outs.
*/
int temps_start = code_item->outs_size_ * sizeof(uint32_t)
+ static_cast<size_t>(pointer_size) /* art method */;
int relative_offset = (reg - (temp_threshold + max_num_special_temps)) * sizeof(uint32_t);
return temps_start + relative_offset;
} else if (reg < num_regs) {
int locals_start = frame_size - spill_size - num_regs * sizeof(uint32_t);
return locals_start + (reg * sizeof(uint32_t));
} else {
// Handle ins.
return frame_size + ((reg - num_regs) * sizeof(uint32_t))
+ static_cast<size_t>(pointer_size) /* art method */;
}
}
void DumpVregLocations(std::ostream& os, const OatFile::OatMethod& oat_method,
const DexFile::CodeItem* code_item) {
if (code_item != nullptr) {
size_t num_locals_ins = code_item->registers_size_;
size_t num_ins = code_item->ins_size_;
size_t num_locals = num_locals_ins - num_ins;
size_t num_outs = code_item->outs_size_;
os << "vr_stack_locations:";
for (size_t reg = 0; reg <= num_locals_ins; reg++) {
// For readability, delimit the different kinds of VRs.
if (reg == num_locals_ins) {
os << "\n\tmethod*:";
} else if (reg == num_locals && num_ins > 0) {
os << "\n\tins:";
} else if (reg == 0 && num_locals > 0) {
os << "\n\tlocals:";
}
uint32_t offset = GetVRegOffsetFromQuickCode(code_item,
oat_method.GetCoreSpillMask(),
oat_method.GetFpSpillMask(),
oat_method.GetFrameSizeInBytes(),
reg,
GetInstructionSet());
os << " v" << reg << "[sp + #" << offset << "]";
}
for (size_t out_reg = 0; out_reg < num_outs; out_reg++) {
if (out_reg == 0) {
os << "\n\touts:";
}
uint32_t offset = GetOutVROffset(out_reg, GetInstructionSet());
os << " v" << out_reg << "[sp + #" << offset << "]";
}
os << "\n";
}
}
void DumpDexCode(std::ostream& os, const DexFile& dex_file, const DexFile::CodeItem* code_item) {
if (code_item != nullptr) {
size_t i = 0;
while (i < code_item->insns_size_in_code_units_) {
const Instruction* instruction = Instruction::At(&code_item->insns_[i]);
os << StringPrintf("0x%04zx: ", i) << instruction->DumpHexLE(5)
<< StringPrintf("\t| %s\n", instruction->DumpString(&dex_file).c_str());
i += instruction->SizeInCodeUnits();
}
}
}
// Has `oat_method` -- corresponding to the Dex `code_item` -- been compiled by
// the optimizing compiler?
static bool IsMethodGeneratedByOptimizingCompiler(const OatFile::OatMethod& oat_method,
const DexFile::CodeItem* code_item) {
// If the native GC map is null and the Dex `code_item` is not
// null, then this method has been compiled with the optimizing
// compiler.
return oat_method.GetQuickCode() != nullptr &&
oat_method.GetVmapTable() != nullptr &&
code_item != nullptr;
}
// Has `oat_method` -- corresponding to the Dex `code_item` -- been compiled by
// the dextodex compiler?
static bool IsMethodGeneratedByDexToDexCompiler(const OatFile::OatMethod& oat_method,
const DexFile::CodeItem* code_item) {
// If the quick code is null, the Dex `code_item` is not
// null, and the vmap table is not null, then this method has been compiled
// with the dextodex compiler.
return oat_method.GetQuickCode() == nullptr &&
oat_method.GetVmapTable() != nullptr &&
code_item != nullptr;
}
verifier::MethodVerifier* DumpVerifier(VariableIndentationOutputStream* vios,
StackHandleScope<1>* hs,
uint32_t dex_method_idx,
const DexFile* dex_file,
const DexFile::ClassDef& class_def,
const DexFile::CodeItem* code_item,
uint32_t method_access_flags) {
if ((method_access_flags & kAccNative) == 0) {
ScopedObjectAccess soa(Thread::Current());
Runtime* const runtime = Runtime::Current();
Handle<mirror::DexCache> dex_cache(
hs->NewHandle(runtime->GetClassLinker()->RegisterDexFile(*dex_file, nullptr)));
CHECK(dex_cache != nullptr);
DCHECK(options_.class_loader_ != nullptr);
return verifier::MethodVerifier::VerifyMethodAndDump(
soa.Self(), vios, dex_method_idx, dex_file, dex_cache, *options_.class_loader_,
class_def, code_item, nullptr, method_access_flags);
}
return nullptr;
}
// The StackMapsHelper provides the stack maps in the native PC order.
// For identical native PCs, the order from the CodeInfo is preserved.
class StackMapsHelper {
public:
explicit StackMapsHelper(const uint8_t* raw_code_info, InstructionSet instruction_set)
: code_info_(raw_code_info),
encoding_(code_info_.ExtractEncoding()),
number_of_stack_maps_(code_info_.GetNumberOfStackMaps(encoding_)),
indexes_(),
offset_(static_cast<uint32_t>(-1)),
stack_map_index_(0u),
instruction_set_(instruction_set) {
if (number_of_stack_maps_ != 0u) {
// Check if native PCs are ordered.
bool ordered = true;
StackMap last = code_info_.GetStackMapAt(0u, encoding_);
for (size_t i = 1; i != number_of_stack_maps_; ++i) {
StackMap current = code_info_.GetStackMapAt(i, encoding_);
if (last.GetNativePcOffset(encoding_.stack_map.encoding, instruction_set) >
current.GetNativePcOffset(encoding_.stack_map.encoding, instruction_set)) {
ordered = false;
break;
}
last = current;
}
if (!ordered) {
// Create indirection indexes for access in native PC order. We do not optimize
// for the fact that there can currently be only two separately ordered ranges,
// namely normal stack maps and catch-point stack maps.
indexes_.resize(number_of_stack_maps_);
std::iota(indexes_.begin(), indexes_.end(), 0u);
std::sort(indexes_.begin(),
indexes_.end(),
[this](size_t lhs, size_t rhs) {
StackMap left = code_info_.GetStackMapAt(lhs, encoding_);
uint32_t left_pc = left.GetNativePcOffset(encoding_.stack_map.encoding,
instruction_set_);
StackMap right = code_info_.GetStackMapAt(rhs, encoding_);
uint32_t right_pc = right.GetNativePcOffset(encoding_.stack_map.encoding,
instruction_set_);
// If the PCs are the same, compare indexes to preserve the original order.
return (left_pc < right_pc) || (left_pc == right_pc && lhs < rhs);
});
}
offset_ = GetStackMapAt(0).GetNativePcOffset(encoding_.stack_map.encoding,
instruction_set_);
}
}
const CodeInfo& GetCodeInfo() const {
return code_info_;
}
const CodeInfoEncoding& GetEncoding() const {
return encoding_;
}
uint32_t GetOffset() const {
return offset_;
}
StackMap GetStackMap() const {
return GetStackMapAt(stack_map_index_);
}
void Next() {
++stack_map_index_;
offset_ = (stack_map_index_ == number_of_stack_maps_)
? static_cast<uint32_t>(-1)
: GetStackMapAt(stack_map_index_).GetNativePcOffset(encoding_.stack_map.encoding,
instruction_set_);
}
private:
StackMap GetStackMapAt(size_t i) const {
if (!indexes_.empty()) {
i = indexes_[i];
}
DCHECK_LT(i, number_of_stack_maps_);
return code_info_.GetStackMapAt(i, encoding_);
}
const CodeInfo code_info_;
const CodeInfoEncoding encoding_;
const size_t number_of_stack_maps_;
dchecked_vector<size_t> indexes_; // Used if stack map native PCs are not ordered.
uint32_t offset_;
size_t stack_map_index_;
const InstructionSet instruction_set_;
};
void DumpCode(VariableIndentationOutputStream* vios,
const OatFile::OatMethod& oat_method, const DexFile::CodeItem* code_item,
bool bad_input, size_t code_size) {
const void* quick_code = oat_method.GetQuickCode();
if (code_size == 0) {
code_size = oat_method.GetQuickCodeSize();
}
if (code_size == 0 || quick_code == nullptr) {
vios->Stream() << "NO CODE!\n";
return;
} else if (!bad_input && IsMethodGeneratedByOptimizingCompiler(oat_method, code_item)) {
// The optimizing compiler outputs its CodeInfo data in the vmap table.
StackMapsHelper helper(oat_method.GetVmapTable(), instruction_set_);
MethodInfo method_info(oat_method.GetOatQuickMethodHeader()->GetOptimizedMethodInfo());
{
CodeInfoEncoding encoding(helper.GetEncoding());
StackMapEncoding stack_map_encoding(encoding.stack_map.encoding);
const size_t num_stack_maps = encoding.stack_map.num_entries;
if (stats_.AddBitsIfUnique(Stats::kByteKindCodeInfoEncoding,
encoding.HeaderSize() * kBitsPerByte,
oat_method.GetVmapTable())) {
// Stack maps
stats_.AddBits(
Stats::kByteKindStackMapNativePc,
stack_map_encoding.GetNativePcEncoding().BitSize() * num_stack_maps);
stats_.AddBits(
Stats::kByteKindStackMapDexPc,
stack_map_encoding.GetDexPcEncoding().BitSize() * num_stack_maps);
stats_.AddBits(
Stats::kByteKindStackMapDexRegisterMap,
stack_map_encoding.GetDexRegisterMapEncoding().BitSize() * num_stack_maps);
stats_.AddBits(
Stats::kByteKindStackMapInlineInfoIndex,
stack_map_encoding.GetInlineInfoEncoding().BitSize() * num_stack_maps);
stats_.AddBits(
Stats::kByteKindStackMapRegisterMaskIndex,
stack_map_encoding.GetRegisterMaskIndexEncoding().BitSize() * num_stack_maps);
stats_.AddBits(
Stats::kByteKindStackMapStackMaskIndex,
stack_map_encoding.GetStackMaskIndexEncoding().BitSize() * num_stack_maps);
// Stack masks
stats_.AddBits(
Stats::kByteKindCodeInfoStackMasks,
encoding.stack_mask.encoding.BitSize() * encoding.stack_mask.num_entries);
// Register masks
stats_.AddBits(
Stats::kByteKindCodeInfoRegisterMasks,
encoding.register_mask.encoding.BitSize() * encoding.register_mask.num_entries);
// Invoke infos
if (encoding.invoke_info.num_entries > 0u) {
stats_.AddBits(
Stats::kByteKindCodeInfoInvokeInfo,
encoding.invoke_info.encoding.BitSize() * encoding.invoke_info.num_entries);
}
// Location catalog
const size_t location_catalog_bytes =
helper.GetCodeInfo().GetDexRegisterLocationCatalogSize(encoding);
stats_.AddBits(Stats::kByteKindCodeInfoLocationCatalog,
kBitsPerByte * location_catalog_bytes);
// Dex register bytes.
const size_t dex_register_bytes =
helper.GetCodeInfo().GetDexRegisterMapsSize(encoding, code_item->registers_size_);
stats_.AddBits(
Stats::kByteKindCodeInfoDexRegisterMap,
kBitsPerByte * dex_register_bytes);
// Inline infos.
const size_t num_inline_infos = encoding.inline_info.num_entries;
if (num_inline_infos > 0u) {
stats_.AddBits(
Stats::kByteKindInlineInfoMethodIndexIdx,
encoding.inline_info.encoding.GetMethodIndexIdxEncoding().BitSize() *
num_inline_infos);
stats_.AddBits(
Stats::kByteKindInlineInfoDexPc,
encoding.inline_info.encoding.GetDexPcEncoding().BitSize() * num_inline_infos);
stats_.AddBits(
Stats::kByteKindInlineInfoExtraData,
encoding.inline_info.encoding.GetExtraDataEncoding().BitSize() * num_inline_infos);
stats_.AddBits(
Stats::kByteKindInlineInfoDexRegisterMap,
encoding.inline_info.encoding.GetDexRegisterMapEncoding().BitSize() *
num_inline_infos);
stats_.AddBits(Stats::kByteKindInlineInfoIsLast, num_inline_infos);
}
}
}
const uint8_t* quick_native_pc = reinterpret_cast<const uint8_t*>(quick_code);
size_t offset = 0;
while (offset < code_size) {
offset += disassembler_->Dump(vios->Stream(), quick_native_pc + offset);
if (offset == helper.GetOffset()) {
ScopedIndentation indent1(vios);
StackMap stack_map = helper.GetStackMap();
DCHECK(stack_map.IsValid());
stack_map.Dump(vios,
helper.GetCodeInfo(),
helper.GetEncoding(),
method_info,
oat_method.GetCodeOffset(),
code_item->registers_size_,
instruction_set_);
do {
helper.Next();
// There may be multiple stack maps at a given PC. We display only the first one.
} while (offset == helper.GetOffset());
}
DCHECK_LT(offset, helper.GetOffset());
}
} else {
const uint8_t* quick_native_pc = reinterpret_cast<const uint8_t*>(quick_code);
size_t offset = 0;
while (offset < code_size) {
offset += disassembler_->Dump(vios->Stream(), quick_native_pc + offset);
}
}
}
const OatFile& oat_file_;
const std::vector<const OatFile::OatDexFile*> oat_dex_files_;
const OatDumperOptions& options_;
uint32_t resolved_addr2instr_;
const InstructionSet instruction_set_;
std::set<uintptr_t> offsets_;
Disassembler* disassembler_;
Stats stats_;
};
class ImageDumper {
public:
ImageDumper(std::ostream* os,
gc::space::ImageSpace& image_space,
const ImageHeader& image_header,
OatDumperOptions* oat_dumper_options)
: os_(os),
vios_(os),
indent1_(&vios_),
image_space_(image_space),
image_header_(image_header),
oat_dumper_options_(oat_dumper_options) {}
bool Dump() REQUIRES_SHARED(Locks::mutator_lock_) {
std::ostream& os = *os_;
std::ostream& indent_os = vios_.Stream();
os << "MAGIC: " << image_header_.GetMagic() << "\n\n";
os << "IMAGE LOCATION: " << image_space_.GetImageLocation() << "\n\n";
os << "IMAGE BEGIN: " << reinterpret_cast<void*>(image_header_.GetImageBegin()) << "\n\n";
os << "IMAGE SIZE: " << image_header_.GetImageSize() << "\n\n";
for (size_t i = 0; i < ImageHeader::kSectionCount; ++i) {
auto section = static_cast<ImageHeader::ImageSections>(i);
os << "IMAGE SECTION " << section << ": " << image_header_.GetImageSection(section) << "\n\n";
}
os << "OAT CHECKSUM: " << StringPrintf("0x%08x\n\n", image_header_.GetOatChecksum());
os << "OAT FILE BEGIN:" << reinterpret_cast<void*>(image_header_.GetOatFileBegin()) << "\n\n";
os << "OAT DATA BEGIN:" << reinterpret_cast<void*>(image_header_.GetOatDataBegin()) << "\n\n";
os << "OAT DATA END:" << reinterpret_cast<void*>(image_header_.GetOatDataEnd()) << "\n\n";
os << "OAT FILE END:" << reinterpret_cast<void*>(image_header_.GetOatFileEnd()) << "\n\n";
os << "PATCH DELTA:" << image_header_.GetPatchDelta() << "\n\n";
os << "COMPILE PIC: " << (image_header_.CompilePic() ? "yes" : "no") << "\n\n";
{
os << "ROOTS: " << reinterpret_cast<void*>(image_header_.GetImageRoots()) << "\n";
static_assert(arraysize(image_roots_descriptions_) ==
static_cast<size_t>(ImageHeader::kImageRootsMax), "sizes must match");
DCHECK_LE(image_header_.GetImageRoots()->GetLength(), ImageHeader::kImageRootsMax);
for (int32_t i = 0, size = image_header_.GetImageRoots()->GetLength(); i != size; ++i) {
ImageHeader::ImageRoot image_root = static_cast<ImageHeader::ImageRoot>(i);
const char* image_root_description = image_roots_descriptions_[i];
mirror::Object* image_root_object = image_header_.GetImageRoot(image_root);
indent_os << StringPrintf("%s: %p\n", image_root_description, image_root_object);
if (image_root_object != nullptr && image_root_object->IsObjectArray()) {
mirror::ObjectArray<mirror::Object>* image_root_object_array
= image_root_object->AsObjectArray<mirror::Object>();
ScopedIndentation indent2(&vios_);
for (int j = 0; j < image_root_object_array->GetLength(); j++) {
mirror::Object* value = image_root_object_array->Get(j);
size_t run = 0;
for (int32_t k = j + 1; k < image_root_object_array->GetLength(); k++) {
if (value == image_root_object_array->Get(k)) {
run++;
} else {
break;
}
}
if (run == 0) {
indent_os << StringPrintf("%d: ", j);
} else {
indent_os << StringPrintf("%d to %zd: ", j, j + run);
j = j + run;
}
if (value != nullptr) {
PrettyObjectValue(indent_os, value->GetClass(), value);
} else {
indent_os << j << ": null\n";
}
}
}
}
}
{
os << "METHOD ROOTS\n";
static_assert(arraysize(image_methods_descriptions_) ==
static_cast<size_t>(ImageHeader::kImageMethodsCount), "sizes must match");
for (int i = 0; i < ImageHeader::kImageMethodsCount; i++) {
auto image_root = static_cast<ImageHeader::ImageMethod>(i);
const char* description = image_methods_descriptions_[i];
auto* image_method = image_header_.GetImageMethod(image_root);
indent_os << StringPrintf("%s: %p\n", description, image_method);
}
}
os << "\n";
Runtime* const runtime = Runtime::Current();
ClassLinker* class_linker = runtime->GetClassLinker();
std::string image_filename = image_space_.GetImageFilename();
std::string oat_location = ImageHeader::GetOatLocationFromImageLocation(image_filename);
os << "OAT LOCATION: " << oat_location;
os << "\n";
std::string error_msg;
const OatFile* oat_file = image_space_.GetOatFile();
if (oat_file == nullptr) {
oat_file = runtime->GetOatFileManager().FindOpenedOatFileFromOatLocation(oat_location);
}
if (oat_file == nullptr) {
oat_file = OatFile::Open(oat_location,
oat_location,
nullptr,
nullptr,
false,
/*low_4gb*/false,
nullptr,
&error_msg);
}
if (oat_file == nullptr) {
os << "OAT FILE NOT FOUND: " << error_msg << "\n";
return EXIT_FAILURE;
}
os << "\n";
stats_.oat_file_bytes = oat_file->Size();
oat_dumper_.reset(new OatDumper(*oat_file, *oat_dumper_options_));
for (const OatFile::OatDexFile* oat_dex_file : oat_file->GetOatDexFiles()) {
CHECK(oat_dex_file != nullptr);
stats_.oat_dex_file_sizes.push_back(std::make_pair(oat_dex_file->GetDexFileLocation(),
oat_dex_file->FileSize()));
}
os << "OBJECTS:\n" << std::flush;
// Loop through the image space and dump its objects.
gc::Heap* heap = runtime->GetHeap();
Thread* self = Thread::Current();
{
{
WriterMutexLock mu(self, *Locks::heap_bitmap_lock_);
heap->FlushAllocStack();
}
// Since FlushAllocStack() above resets the (active) allocation
// stack. Need to revoke the thread-local allocation stacks that
// point into it.
ScopedThreadSuspension sts(self, kNative);
ScopedSuspendAll ssa(__FUNCTION__);
heap->RevokeAllThreadLocalAllocationStacks(self);
}
{
// Mark dex caches.
dex_caches_.clear();
{
ReaderMutexLock mu(self, *Locks::dex_lock_);
for (const ClassLinker::DexCacheData& data : class_linker->GetDexCachesData()) {
ObjPtr<mirror::DexCache> dex_cache =
ObjPtr<mirror::DexCache>::DownCast(self->DecodeJObject(data.weak_root));
if (dex_cache != nullptr) {
dex_caches_.insert(dex_cache.Ptr());
}
}
}
auto dump_visitor = [&](mirror::Object* obj) REQUIRES_SHARED(Locks::mutator_lock_) {
DumpObject(obj);
};
ReaderMutexLock mu(self, *Locks::heap_bitmap_lock_);
// Dump the normal objects before ArtMethods.
image_space_.GetLiveBitmap()->Walk(dump_visitor);
indent_os << "\n";
// TODO: Dump fields.
// Dump methods after.
DumpArtMethodVisitor visitor(this);
image_header_.VisitPackedArtMethods(&visitor,
image_space_.Begin(),
image_header_.GetPointerSize());
// Dump the large objects separately.
heap->GetLargeObjectsSpace()->GetLiveBitmap()->Walk(dump_visitor);
indent_os << "\n";
}
os << "STATS:\n" << std::flush;
std::unique_ptr<File> file(OS::OpenFileForReading(image_filename.c_str()));
size_t data_size = image_header_.GetDataSize(); // stored size in file.
if (file == nullptr) {
LOG(WARNING) << "Failed to find image in " << image_filename;
} else {
stats_.file_bytes = file->GetLength();
// If the image is compressed, adjust to decompressed size.
size_t uncompressed_size = image_header_.GetImageSize() - sizeof(ImageHeader);
if (image_header_.GetStorageMode() == ImageHeader::kStorageModeUncompressed) {
DCHECK_EQ(uncompressed_size, data_size) << "Sizes should match for uncompressed image";
}
stats_.file_bytes += uncompressed_size - data_size;
}
size_t header_bytes = sizeof(ImageHeader);
const auto& object_section = image_header_.GetImageSection(ImageHeader::kSectionObjects);
const auto& field_section = image_header_.GetImageSection(ImageHeader::kSectionArtFields);
const auto& method_section = image_header_.GetMethodsSection();
const auto& dex_cache_arrays_section = image_header_.GetImageSection(
ImageHeader::kSectionDexCacheArrays);
const auto& intern_section = image_header_.GetImageSection(
ImageHeader::kSectionInternedStrings);
const auto& class_table_section = image_header_.GetImageSection(
ImageHeader::kSectionClassTable);
const auto& bitmap_section = image_header_.GetImageSection(ImageHeader::kSectionImageBitmap);
stats_.header_bytes = header_bytes;
// Objects are kObjectAlignment-aligned.
// CHECK_EQ(RoundUp(header_bytes, kObjectAlignment), object_section.Offset());
if (object_section.Offset() > header_bytes) {
stats_.alignment_bytes += object_section.Offset() - header_bytes;
}
// Field section is 4-byte aligned.
constexpr size_t kFieldSectionAlignment = 4U;
uint32_t end_objects = object_section.Offset() + object_section.Size();
CHECK_EQ(RoundUp(end_objects, kFieldSectionAlignment), field_section.Offset());
stats_.alignment_bytes += field_section.Offset() - end_objects;
// Method section is 4/8 byte aligned depending on target. Just check for 4-byte alignment.
uint32_t end_fields = field_section.Offset() + field_section.Size();
CHECK_ALIGNED(method_section.Offset(), 4);
stats_.alignment_bytes += method_section.Offset() - end_fields;
// Dex cache arrays section is aligned depending on the target. Just check for 4-byte alignment.
uint32_t end_methods = method_section.Offset() + method_section.Size();
CHECK_ALIGNED(dex_cache_arrays_section.Offset(), 4);
stats_.alignment_bytes += dex_cache_arrays_section.Offset() - end_methods;
// Intern table is 8-byte aligned.
uint32_t end_caches = dex_cache_arrays_section.Offset() + dex_cache_arrays_section.Size();
CHECK_EQ(RoundUp(end_caches, 8U), intern_section.Offset());
stats_.alignment_bytes += intern_section.Offset() - end_caches;
// Add space between intern table and class table.
uint32_t end_intern = intern_section.Offset() + intern_section.Size();
stats_.alignment_bytes += class_table_section.Offset() - end_intern;
// Add space between end of image data and bitmap. Expect the bitmap to be page-aligned.
const size_t bitmap_offset = sizeof(ImageHeader) + data_size;
CHECK_ALIGNED(bitmap_section.Offset(), kPageSize);
stats_.alignment_bytes += RoundUp(bitmap_offset, kPageSize) - bitmap_offset;
stats_.bitmap_bytes += bitmap_section.Size();
stats_.art_field_bytes += field_section.Size();
stats_.art_method_bytes += method_section.Size();
stats_.dex_cache_arrays_bytes += dex_cache_arrays_section.Size();
stats_.interned_strings_bytes += intern_section.Size();
stats_.class_table_bytes += class_table_section.Size();
stats_.Dump(os, indent_os);
os << "\n";
os << std::flush;
return oat_dumper_->Dump(os);
}
private:
class DumpArtMethodVisitor : public ArtMethodVisitor {
public:
explicit DumpArtMethodVisitor(ImageDumper* image_dumper) : image_dumper_(image_dumper) {}
virtual void Visit(ArtMethod* method) OVERRIDE REQUIRES_SHARED(Locks::mutator_lock_) {
std::ostream& indent_os = image_dumper_->vios_.Stream();
indent_os << method << " " << " ArtMethod: " << ArtMethod::PrettyMethod(method) << "\n";
image_dumper_->DumpMethod(method, indent_os);
indent_os << "\n";
}
private:
ImageDumper* const image_dumper_;
};
static void PrettyObjectValue(std::ostream& os,
ObjPtr<mirror::Class> type,
ObjPtr<mirror::Object> value)
REQUIRES_SHARED(Locks::mutator_lock_) {
CHECK(type != nullptr);
if (value == nullptr) {
os << StringPrintf("null %s\n", type->PrettyDescriptor().c_str());
} else if (type->IsStringClass()) {
mirror::String* string = value->AsString();
os << StringPrintf("%p String: %s\n", string,
PrintableString(string->ToModifiedUtf8().c_str()).c_str());
} else if (type->IsClassClass()) {
mirror::Class* klass = value->AsClass();
os << StringPrintf("%p Class: %s\n", klass, mirror::Class::PrettyDescriptor(klass).c_str());
} else {
os << StringPrintf("%p %s\n", value.Ptr(), type->PrettyDescriptor().c_str());
}
}
static void PrintField(std::ostream& os, ArtField* field, ObjPtr<mirror::Object> obj)
REQUIRES_SHARED(Locks::mutator_lock_) {
os << StringPrintf("%s: ", field->GetName());
switch (field->GetTypeAsPrimitiveType()) {
case Primitive::kPrimLong:
os << StringPrintf("%" PRId64 " (0x%" PRIx64 ")\n", field->Get64(obj), field->Get64(obj));
break;
case Primitive::kPrimDouble:
os << StringPrintf("%f (%a)\n", field->GetDouble(obj), field->GetDouble(obj));
break;
case Primitive::kPrimFloat:
os << StringPrintf("%f (%a)\n", field->GetFloat(obj), field->GetFloat(obj));
break;
case Primitive::kPrimInt:
os << StringPrintf("%d (0x%x)\n", field->Get32(obj), field->Get32(obj));
break;
case Primitive::kPrimChar:
os << StringPrintf("%u (0x%x)\n", field->GetChar(obj), field->GetChar(obj));
break;
case Primitive::kPrimShort:
os << StringPrintf("%d (0x%x)\n", field->GetShort(obj), field->GetShort(obj));
break;
case Primitive::kPrimBoolean:
os << StringPrintf("%s (0x%x)\n", field->GetBoolean(obj) ? "true" : "false",
field->GetBoolean(obj));
break;
case Primitive::kPrimByte:
os << StringPrintf("%d (0x%x)\n", field->GetByte(obj), field->GetByte(obj));
break;
case Primitive::kPrimNot: {
// Get the value, don't compute the type unless it is non-null as we don't want
// to cause class loading.
ObjPtr<mirror::Object> value = field->GetObj(obj);
if (value == nullptr) {
os << StringPrintf("null %s\n", PrettyDescriptor(field->GetTypeDescriptor()).c_str());
} else {
// Grab the field type without causing resolution.
ObjPtr<mirror::Class> field_type = field->GetType<false>();
if (field_type != nullptr) {
PrettyObjectValue(os, field_type, value);
} else {
os << StringPrintf("%p %s\n",
value.Ptr(),
PrettyDescriptor(field->GetTypeDescriptor()).c_str());
}
}
break;
}
default:
os << "unexpected field type: " << field->GetTypeDescriptor() << "\n";
break;
}
}
static void DumpFields(std::ostream& os, mirror::Object* obj, mirror::Class* klass)
REQUIRES_SHARED(Locks::mutator_lock_) {
mirror::Class* super = klass->GetSuperClass();
if (super != nullptr) {
DumpFields(os, obj, super);
}
for (ArtField& field : klass->GetIFields()) {
PrintField(os, &field, obj);
}
}
bool InDumpSpace(const mirror::Object* object) {
return image_space_.Contains(object);
}
const void* GetQuickOatCodeBegin(ArtMethod* m) REQUIRES_SHARED(Locks::mutator_lock_) {
const void* quick_code = m->GetEntryPointFromQuickCompiledCodePtrSize(
image_header_.GetPointerSize());
if (Runtime::Current()->GetClassLinker()->IsQuickResolutionStub(quick_code)) {
quick_code = oat_dumper_->GetQuickOatCode(m);
}
if (oat_dumper_->GetInstructionSet() == kThumb2) {
quick_code = reinterpret_cast<void*>(reinterpret_cast<uintptr_t>(quick_code) & ~0x1);
}
return quick_code;
}
uint32_t GetQuickOatCodeSize(ArtMethod* m)
REQUIRES_SHARED(Locks::mutator_lock_) {
const uint32_t* oat_code_begin = reinterpret_cast<const uint32_t*>(GetQuickOatCodeBegin(m));
if (oat_code_begin == nullptr) {
return 0;
}
return oat_code_begin[-1];
}
const void* GetQuickOatCodeEnd(ArtMethod* m)
REQUIRES_SHARED(Locks::mutator_lock_) {
const uint8_t* oat_code_begin = reinterpret_cast<const uint8_t*>(GetQuickOatCodeBegin(m));
if (oat_code_begin == nullptr) {
return nullptr;
}
return oat_code_begin + GetQuickOatCodeSize(m);
}
void DumpObject(mirror::Object* obj) REQUIRES_SHARED(Locks::mutator_lock_) {
DCHECK(obj != nullptr);
if (!InDumpSpace(obj)) {
return;
}
size_t object_bytes = obj->SizeOf();
size_t alignment_bytes = RoundUp(object_bytes, kObjectAlignment) - object_bytes;
stats_.object_bytes += object_bytes;
stats_.alignment_bytes += alignment_bytes;
std::ostream& os = vios_.Stream();
mirror::Class* obj_class = obj->GetClass();
if (obj_class->IsArrayClass()) {
os << StringPrintf("%p: %s length:%d\n", obj, obj_class->PrettyDescriptor().c_str(),
obj->AsArray()->GetLength());
} else if (obj->IsClass()) {
mirror::Class* klass = obj->AsClass();
os << StringPrintf("%p: java.lang.Class \"%s\" (", obj,
mirror::Class::PrettyDescriptor(klass).c_str())
<< klass->GetStatus() << ")\n";
} else if (obj_class->IsStringClass()) {
os << StringPrintf("%p: java.lang.String %s\n", obj,
PrintableString(obj->AsString()->ToModifiedUtf8().c_str()).c_str());
} else {
os << StringPrintf("%p: %s\n", obj, obj_class->PrettyDescriptor().c_str());
}
ScopedIndentation indent1(&vios_);
DumpFields(os, obj, obj_class);
const PointerSize image_pointer_size = image_header_.GetPointerSize();
if (obj->IsObjectArray()) {
auto* obj_array = obj->AsObjectArray<mirror::Object>();
for (int32_t i = 0, length = obj_array->GetLength(); i < length; i++) {
mirror::Object* value = obj_array->Get(i);
size_t run = 0;
for (int32_t j = i + 1; j < length; j++) {
if (value == obj_array->Get(j)) {
run++;
} else {
break;
}
}
if (run == 0) {
os << StringPrintf("%d: ", i);
} else {
os << StringPrintf("%d to %zd: ", i, i + run);
i = i + run;
}
mirror::Class* value_class =
(value == nullptr) ? obj_class->GetComponentType() : value->GetClass();
PrettyObjectValue(os, value_class, value);
}
} else if (obj->IsClass()) {
mirror::Class* klass = obj->AsClass();
if (klass->NumStaticFields() != 0) {
os << "STATICS:\n";
ScopedIndentation indent2(&vios_);
for (ArtField& field : klass->GetSFields()) {
PrintField(os, &field, field.GetDeclaringClass());
}
}
} else {
auto it = dex_caches_.find(obj);
if (it != dex_caches_.end()) {
auto* dex_cache = down_cast<mirror::DexCache*>(obj);
const auto& field_section = image_header_.GetImageSection(
ImageHeader::kSectionArtFields);
const auto& method_section = image_header_.GetMethodsSection();
size_t num_methods = dex_cache->NumResolvedMethods();
if (num_methods != 0u) {
os << "Methods (size=" << num_methods << "):\n";
ScopedIndentation indent2(&vios_);
auto* resolved_methods = dex_cache->GetResolvedMethods();
for (size_t i = 0, length = dex_cache->NumResolvedMethods(); i < length; ++i) {
auto* elem = mirror::DexCache::GetElementPtrSize(resolved_methods,
i,
image_pointer_size);
size_t run = 0;
for (size_t j = i + 1;
j != length && elem == mirror::DexCache::GetElementPtrSize(resolved_methods,
j,
image_pointer_size);
++j) {
++run;
}
if (run == 0) {
os << StringPrintf("%zd: ", i);
} else {
os << StringPrintf("%zd to %zd: ", i, i + run);
i = i + run;
}
std::string msg;
if (elem == nullptr) {
msg = "null";
} else if (method_section.Contains(
reinterpret_cast<uint8_t*>(elem) - image_space_.Begin())) {
msg = reinterpret_cast<ArtMethod*>(elem)->PrettyMethod();
} else {
msg = "<not in method section>";
}
os << StringPrintf("%p %s\n", elem, msg.c_str());
}
}
size_t num_fields = dex_cache->NumResolvedFields();
if (num_fields != 0u) {
os << "Fields (size=" << num_fields << "):\n";
ScopedIndentation indent2(&vios_);
auto* resolved_fields = dex_cache->GetResolvedFields();
for (size_t i = 0, length = dex_cache->NumResolvedFields(); i < length; ++i) {
auto* elem = mirror::DexCache::GetNativePairPtrSize(
resolved_fields, i, image_pointer_size).object;
size_t run = 0;
for (size_t j = i + 1;
j != length &&
elem == mirror::DexCache::GetNativePairPtrSize(
resolved_fields, j, image_pointer_size).object;
++j) {
++run;
}
if (run == 0) {
os << StringPrintf("%zd: ", i);
} else {
os << StringPrintf("%zd to %zd: ", i, i + run);
i = i + run;
}
std::string msg;
if (elem == nullptr) {
msg = "null";
} else if (field_section.Contains(
reinterpret_cast<uint8_t*>(elem) - image_space_.Begin())) {
msg = reinterpret_cast<ArtField*>(elem)->PrettyField();
} else {
msg = "<not in field section>";
}
os << StringPrintf("%p %s\n", elem, msg.c_str());
}
}
size_t num_types = dex_cache->NumResolvedTypes();
if (num_types != 0u) {
os << "Types (size=" << num_types << "):\n";
ScopedIndentation indent2(&vios_);
auto* resolved_types = dex_cache->GetResolvedTypes();
for (size_t i = 0; i < num_types; ++i) {
auto pair = resolved_types[i].load(std::memory_order_relaxed);
size_t run = 0;
for (size_t j = i + 1; j != num_types; ++j) {
auto other_pair = resolved_types[j].load(std::memory_order_relaxed);
if (pair.index != other_pair.index ||
pair.object.Read() != other_pair.object.Read()) {
break;
}
++run;
}
if (run == 0) {
os << StringPrintf("%zd: ", i);
} else {
os << StringPrintf("%zd to %zd: ", i, i + run);
i = i + run;
}
std::string msg;
auto* elem = pair.object.Read();
if (elem == nullptr) {
msg = "null";
} else {
msg = elem->PrettyClass();
}
os << StringPrintf("%p %u %s\n", elem, pair.index, msg.c_str());
}
}
}
}
std::string temp;
stats_.Update(obj_class->GetDescriptor(&temp), object_bytes);
}
void DumpMethod(ArtMethod* method, std::ostream& indent_os)
REQUIRES_SHARED(Locks::mutator_lock_) {
DCHECK(method != nullptr);
const void* quick_oat_code_begin = GetQuickOatCodeBegin(method);
const void* quick_oat_code_end = GetQuickOatCodeEnd(method);
const PointerSize pointer_size = image_header_.GetPointerSize();
OatQuickMethodHeader* method_header = reinterpret_cast<OatQuickMethodHeader*>(
reinterpret_cast<uintptr_t>(quick_oat_code_begin) - sizeof(OatQuickMethodHeader));
if (method->IsNative()) {
bool first_occurrence;
uint32_t quick_oat_code_size = GetQuickOatCodeSize(method);
ComputeOatSize(quick_oat_code_begin, &first_occurrence);
if (first_occurrence) {
stats_.native_to_managed_code_bytes += quick_oat_code_size;
}
if (quick_oat_code_begin != method->GetEntryPointFromQuickCompiledCodePtrSize(
image_header_.GetPointerSize())) {
indent_os << StringPrintf("OAT CODE: %p\n", quick_oat_code_begin);
}
} else if (method->IsAbstract() || method->IsClassInitializer()) {
// Don't print information for these.
} else if (method->IsRuntimeMethod()) {
ImtConflictTable* table = method->GetImtConflictTable(image_header_.GetPointerSize());
if (table != nullptr) {
indent_os << "IMT conflict table " << table << " method: ";
for (size_t i = 0, count = table->NumEntries(pointer_size); i < count; ++i) {
indent_os << ArtMethod::PrettyMethod(table->GetImplementationMethod(i, pointer_size))
<< " ";
}
}
} else {
const DexFile::CodeItem* code_item = method->GetCodeItem();
size_t dex_instruction_bytes = code_item->insns_size_in_code_units_ * 2;
stats_.dex_instruction_bytes += dex_instruction_bytes;
bool first_occurrence;
size_t vmap_table_bytes = 0u;
if (!method_header->IsOptimized()) {
// Method compiled with the optimizing compiler have no vmap table.
vmap_table_bytes = ComputeOatSize(method_header->GetVmapTable(), &first_occurrence);
if (first_occurrence) {
stats_.vmap_table_bytes += vmap_table_bytes;
}
}
uint32_t quick_oat_code_size = GetQuickOatCodeSize(method);
ComputeOatSize(quick_oat_code_begin, &first_occurrence);
if (first_occurrence) {
stats_.managed_code_bytes += quick_oat_code_size;
if (method->IsConstructor()) {
if (method->IsStatic()) {
stats_.class_initializer_code_bytes += quick_oat_code_size;
} else if (dex_instruction_bytes > kLargeConstructorDexBytes) {
stats_.large_initializer_code_bytes += quick_oat_code_size;
}
} else if (dex_instruction_bytes > kLargeMethodDexBytes) {
stats_.large_method_code_bytes += quick_oat_code_size;
}
}
stats_.managed_code_bytes_ignoring_deduplication += quick_oat_code_size;
uint32_t method_access_flags = method->GetAccessFlags();
indent_os << StringPrintf("OAT CODE: %p-%p\n", quick_oat_code_begin, quick_oat_code_end);
indent_os << StringPrintf("SIZE: Dex Instructions=%zd StackMaps=%zd AccessFlags=0x%x\n",
dex_instruction_bytes,
vmap_table_bytes,
method_access_flags);
size_t total_size = dex_instruction_bytes +
vmap_table_bytes + quick_oat_code_size + ArtMethod::Size(image_header_.GetPointerSize());
double expansion =
static_cast<double>(quick_oat_code_size) / static_cast<double>(dex_instruction_bytes);
stats_.ComputeOutliers(total_size, expansion, method);
}
}
std::set<const void*> already_seen_;
// Compute the size of the given data within the oat file and whether this is the first time
// this data has been requested
size_t ComputeOatSize(const void* oat_data, bool* first_occurrence) {
if (already_seen_.count(oat_data) == 0) {
*first_occurrence = true;
already_seen_.insert(oat_data);
} else {
*first_occurrence = false;
}
return oat_dumper_->ComputeSize(oat_data);
}
public:
struct Stats {
size_t oat_file_bytes;
size_t file_bytes;
size_t header_bytes;
size_t object_bytes;
size_t art_field_bytes;
size_t art_method_bytes;
size_t dex_cache_arrays_bytes;
size_t interned_strings_bytes;
size_t class_table_bytes;
size_t bitmap_bytes;
size_t alignment_bytes;
size_t managed_code_bytes;
size_t managed_code_bytes_ignoring_deduplication;
size_t native_to_managed_code_bytes;
size_t class_initializer_code_bytes;
size_t large_initializer_code_bytes;
size_t large_method_code_bytes;
size_t vmap_table_bytes;
size_t dex_instruction_bytes;
std::vector<ArtMethod*> method_outlier;
std::vector<size_t> method_outlier_size;
std::vector<double> method_outlier_expansion;
std::vector<std::pair<std::string, size_t>> oat_dex_file_sizes;
Stats()
: oat_file_bytes(0),
file_bytes(0),
header_bytes(0),
object_bytes(0),
art_field_bytes(0),
art_method_bytes(0),
dex_cache_arrays_bytes(0),
interned_strings_bytes(0),
class_table_bytes(0),
bitmap_bytes(0),
alignment_bytes(0),
managed_code_bytes(0),
managed_code_bytes_ignoring_deduplication(0),
native_to_managed_code_bytes(0),
class_initializer_code_bytes(0),
large_initializer_code_bytes(0),
large_method_code_bytes(0),
vmap_table_bytes(0),
dex_instruction_bytes(0) {}
struct SizeAndCount {
SizeAndCount(size_t bytes_in, size_t count_in) : bytes(bytes_in), count(count_in) {}
size_t bytes;
size_t count;
};
typedef SafeMap<std::string, SizeAndCount> SizeAndCountTable;
SizeAndCountTable sizes_and_counts;
void Update(const char* descriptor, size_t object_bytes_in) {
SizeAndCountTable::iterator it = sizes_and_counts.find(descriptor);
if (it != sizes_and_counts.end()) {
it->second.bytes += object_bytes_in;
it->second.count += 1;
} else {
sizes_and_counts.Put(descriptor, SizeAndCount(object_bytes_in, 1));
}
}
double PercentOfOatBytes(size_t size) {
return (static_cast<double>(size) / static_cast<double>(oat_file_bytes)) * 100;
}
double PercentOfFileBytes(size_t size) {
return (static_cast<double>(size) / static_cast<double>(file_bytes)) * 100;
}
double PercentOfObjectBytes(size_t size) {
return (static_cast<double>(size) / static_cast<double>(object_bytes)) * 100;
}
void ComputeOutliers(size_t total_size, double expansion, ArtMethod* method) {
method_outlier_size.push_back(total_size);
method_outlier_expansion.push_back(expansion);
method_outlier.push_back(method);
}
void DumpOutliers(std::ostream& os)
REQUIRES_SHARED(Locks::mutator_lock_) {
size_t sum_of_sizes = 0;
size_t sum_of_sizes_squared = 0;
size_t sum_of_expansion = 0;
size_t sum_of_expansion_squared = 0;
size_t n = method_outlier_size.size();
if (n <= 1) {
return;
}
for (size_t i = 0; i < n; i++) {
size_t cur_size = method_outlier_size[i];
sum_of_sizes += cur_size;
sum_of_sizes_squared += cur_size * cur_size;
double cur_expansion = method_outlier_expansion[i];
sum_of_expansion += cur_expansion;
sum_of_expansion_squared += cur_expansion * cur_expansion;
}
size_t size_mean = sum_of_sizes / n;
size_t size_variance = (sum_of_sizes_squared - sum_of_sizes * size_mean) / (n - 1);
double expansion_mean = sum_of_expansion / n;
double expansion_variance =
(sum_of_expansion_squared - sum_of_expansion * expansion_mean) / (n - 1);
// Dump methods whose size is a certain number of standard deviations from the mean
size_t dumped_values = 0;
size_t skipped_values = 0;
for (size_t i = 100; i > 0; i--) { // i is the current number of standard deviations
size_t cur_size_variance = i * i * size_variance;
bool first = true;
for (size_t j = 0; j < n; j++) {
size_t cur_size = method_outlier_size[j];
if (cur_size > size_mean) {
size_t cur_var = cur_size - size_mean;
cur_var = cur_var * cur_var;
if (cur_var > cur_size_variance) {
if (dumped_values > 20) {
if (i == 1) {
skipped_values++;
} else {
i = 2; // jump to counting for 1 standard deviation
break;
}
} else {
if (first) {
os << "\nBig methods (size > " << i << " standard deviations the norm):\n";
first = false;
}
os << ArtMethod::PrettyMethod(method_outlier[j]) << " requires storage of "
<< PrettySize(cur_size) << "\n";
method_outlier_size[j] = 0; // don't consider this method again
dumped_values++;
}
}
}
}
}
if (skipped_values > 0) {
os << "... skipped " << skipped_values
<< " methods with size > 1 standard deviation from the norm\n";
}
os << std::flush;
// Dump methods whose expansion is a certain number of standard deviations from the mean
dumped_values = 0;
skipped_values = 0;
for (size_t i = 10; i > 0; i--) { // i is the current number of standard deviations
double cur_expansion_variance = i * i * expansion_variance;
bool first = true;
for (size_t j = 0; j < n; j++) {
double cur_expansion = method_outlier_expansion[j];
if (cur_expansion > expansion_mean) {
size_t cur_var = cur_expansion - expansion_mean;
cur_var = cur_var * cur_var;
if (cur_var > cur_expansion_variance) {
if (dumped_values > 20) {
if (i == 1) {
skipped_values++;
} else {
i = 2; // jump to counting for 1 standard deviation
break;
}
} else {
if (first) {
os << "\nLarge expansion methods (size > " << i
<< " standard deviations the norm):\n";
first = false;
}
os << ArtMethod::PrettyMethod(method_outlier[j]) << " expanded code by "
<< cur_expansion << "\n";
method_outlier_expansion[j] = 0.0; // don't consider this method again
dumped_values++;
}
}
}
}
}
if (skipped_values > 0) {
os << "... skipped " << skipped_values
<< " methods with expansion > 1 standard deviation from the norm\n";
}
os << "\n" << std::flush;
}
void Dump(std::ostream& os, std::ostream& indent_os)
REQUIRES_SHARED(Locks::mutator_lock_) {
{
os << "art_file_bytes = " << PrettySize(file_bytes) << "\n\n"
<< "art_file_bytes = header_bytes + object_bytes + alignment_bytes\n";
indent_os << StringPrintf("header_bytes = %8zd (%2.0f%% of art file bytes)\n"
"object_bytes = %8zd (%2.0f%% of art file bytes)\n"
"art_field_bytes = %8zd (%2.0f%% of art file bytes)\n"
"art_method_bytes = %8zd (%2.0f%% of art file bytes)\n"
"dex_cache_arrays_bytes = %8zd (%2.0f%% of art file bytes)\n"
"interned_string_bytes = %8zd (%2.0f%% of art file bytes)\n"
"class_table_bytes = %8zd (%2.0f%% of art file bytes)\n"
"bitmap_bytes = %8zd (%2.0f%% of art file bytes)\n"
"alignment_bytes = %8zd (%2.0f%% of art file bytes)\n\n",
header_bytes, PercentOfFileBytes(header_bytes),
object_bytes, PercentOfFileBytes(object_bytes),
art_field_bytes, PercentOfFileBytes(art_field_bytes),
art_method_bytes, PercentOfFileBytes(art_method_bytes),
dex_cache_arrays_bytes,
PercentOfFileBytes(dex_cache_arrays_bytes),
interned_strings_bytes,
PercentOfFileBytes(interned_strings_bytes),
class_table_bytes, PercentOfFileBytes(class_table_bytes),
bitmap_bytes, PercentOfFileBytes(bitmap_bytes),
alignment_bytes, PercentOfFileBytes(alignment_bytes))
<< std::flush;
CHECK_EQ(file_bytes,
header_bytes + object_bytes + art_field_bytes + art_method_bytes +
dex_cache_arrays_bytes + interned_strings_bytes + class_table_bytes +
bitmap_bytes + alignment_bytes);
}
os << "object_bytes breakdown:\n";
size_t object_bytes_total = 0;
for (const auto& sizes_and_count : sizes_and_counts) {
const std::string& descriptor(sizes_and_count.first);
double average = static_cast<double>(sizes_and_count.second.bytes) /
static_cast<double>(sizes_and_count.second.count);
double percent = PercentOfObjectBytes(sizes_and_count.second.bytes);
os << StringPrintf("%32s %8zd bytes %6zd instances "
"(%4.0f bytes/instance) %2.0f%% of object_bytes\n",
descriptor.c_str(), sizes_and_count.second.bytes,
sizes_and_count.second.count, average, percent);
object_bytes_total += sizes_and_count.second.bytes;
}
os << "\n" << std::flush;
CHECK_EQ(object_bytes, object_bytes_total);
os << StringPrintf("oat_file_bytes = %8zd\n"
"managed_code_bytes = %8zd (%2.0f%% of oat file bytes)\n"
"native_to_managed_code_bytes = %8zd (%2.0f%% of oat file bytes)\n\n"
"class_initializer_code_bytes = %8zd (%2.0f%% of oat file bytes)\n"
"large_initializer_code_bytes = %8zd (%2.0f%% of oat file bytes)\n"
"large_method_code_bytes = %8zd (%2.0f%% of oat file bytes)\n\n",
oat_file_bytes,
managed_code_bytes,
PercentOfOatBytes(managed_code_bytes),
native_to_managed_code_bytes,
PercentOfOatBytes(native_to_managed_code_bytes),
class_initializer_code_bytes,
PercentOfOatBytes(class_initializer_code_bytes),
large_initializer_code_bytes,
PercentOfOatBytes(large_initializer_code_bytes),
large_method_code_bytes,
PercentOfOatBytes(large_method_code_bytes))
<< "DexFile sizes:\n";
for (const std::pair<std::string, size_t>& oat_dex_file_size : oat_dex_file_sizes) {
os << StringPrintf("%s = %zd (%2.0f%% of oat file bytes)\n",
oat_dex_file_size.first.c_str(), oat_dex_file_size.second,
PercentOfOatBytes(oat_dex_file_size.second));
}
os << "\n" << StringPrintf("vmap_table_bytes = %7zd (%2.0f%% of oat file bytes)\n\n",
vmap_table_bytes, PercentOfOatBytes(vmap_table_bytes))
<< std::flush;
os << StringPrintf("dex_instruction_bytes = %zd\n", dex_instruction_bytes)
<< StringPrintf("managed_code_bytes expansion = %.2f (ignoring deduplication %.2f)\n\n",
static_cast<double>(managed_code_bytes) /
static_cast<double>(dex_instruction_bytes),
static_cast<double>(managed_code_bytes_ignoring_deduplication) /
static_cast<double>(dex_instruction_bytes))
<< std::flush;
DumpOutliers(os);
}
} stats_;
private:
enum {
// Number of bytes for a constructor to be considered large. Based on the 1000 basic block
// threshold, we assume 2 bytes per instruction and 2 instructions per block.
kLargeConstructorDexBytes = 4000,
// Number of bytes for a method to be considered large. Based on the 4000 basic block
// threshold, we assume 2 bytes per instruction and 2 instructions per block.
kLargeMethodDexBytes = 16000
};
// For performance, use the *os_ directly for anything that doesn't need indentation
// and prepare an indentation stream with default indentation 1.
std::ostream* os_;
VariableIndentationOutputStream vios_;
ScopedIndentation indent1_;
gc::space::ImageSpace& image_space_;
const ImageHeader& image_header_;
std::unique_ptr<OatDumper> oat_dumper_;
OatDumperOptions* oat_dumper_options_;
std::set<mirror::Object*> dex_caches_;
DISALLOW_COPY_AND_ASSIGN(ImageDumper);
};
static int DumpImage(gc::space::ImageSpace* image_space,
OatDumperOptions* options,
std::ostream* os) REQUIRES_SHARED(Locks::mutator_lock_) {
const ImageHeader& image_header = image_space->GetImageHeader();
if (!image_header.IsValid()) {
fprintf(stderr, "Invalid image header %s\n", image_space->GetImageLocation().c_str());
return EXIT_FAILURE;
}
ImageDumper image_dumper(os, *image_space, image_header, options);
if (!image_dumper.Dump()) {
return EXIT_FAILURE;
}
return EXIT_SUCCESS;
}
static int DumpImages(Runtime* runtime, OatDumperOptions* options, std::ostream* os) {
// Dumping the image, no explicit class loader.
ScopedNullHandle<mirror::ClassLoader> null_class_loader;
options->class_loader_ = &null_class_loader;
ScopedObjectAccess soa(Thread::Current());
if (options->app_image_ != nullptr) {
if (options->app_oat_ == nullptr) {
LOG(ERROR) << "Can not dump app image without app oat file";
return EXIT_FAILURE;
}
// We can't know if the app image is 32 bits yet, but it contains pointers into the oat file.
// We need to map the oat file in the low 4gb or else the fixup wont be able to fit oat file
// pointers into 32 bit pointer sized ArtMethods.
std::string error_msg;
std::unique_ptr<OatFile> oat_file(OatFile::Open(options->app_oat_,
options->app_oat_,
nullptr,
nullptr,
false,
/*low_4gb*/true,
nullptr,
&error_msg));
if (oat_file == nullptr) {
LOG(ERROR) << "Failed to open oat file " << options->app_oat_ << " with error " << error_msg;
return EXIT_FAILURE;
}
std::unique_ptr<gc::space::ImageSpace> space(
gc::space::ImageSpace::CreateFromAppImage(options->app_image_, oat_file.get(), &error_msg));
if (space == nullptr) {
LOG(ERROR) << "Failed to open app image " << options->app_image_ << " with error "
<< error_msg;
}
// Open dex files for the image.
std::vector<std::unique_ptr<const DexFile>> dex_files;
if (!runtime->GetClassLinker()->OpenImageDexFiles(space.get(), &dex_files, &error_msg)) {
LOG(ERROR) << "Failed to open app image dex files " << options->app_image_ << " with error "
<< error_msg;
}
// Dump the actual image.
int result = DumpImage(space.get(), options, os);
if (result != EXIT_SUCCESS) {
return result;
}
// Fall through to dump the boot images.
}
gc::Heap* heap = runtime->GetHeap();
CHECK(heap->HasBootImageSpace()) << "No image spaces";
for (gc::space::ImageSpace* image_space : heap->GetBootImageSpaces()) {
int result = DumpImage(image_space, options, os);
if (result != EXIT_SUCCESS) {
return result;
}
}
return EXIT_SUCCESS;
}
static jobject InstallOatFile(Runtime* runtime,
std::unique_ptr<OatFile> oat_file,
std::vector<const DexFile*>* class_path)
REQUIRES_SHARED(Locks::mutator_lock_) {
Thread* self = Thread::Current();
CHECK(self != nullptr);
// Need well-known-classes.
WellKnownClasses::Init(self->GetJniEnv());
// Need to register dex files to get a working dex cache.
OatFile* oat_file_ptr = oat_file.get();
ClassLinker* class_linker = runtime->GetClassLinker();
runtime->GetOatFileManager().RegisterOatFile(std::move(oat_file));
for (const OatFile::OatDexFile* odf : oat_file_ptr->GetOatDexFiles()) {
std::string error_msg;
const DexFile* const dex_file = OpenDexFile(odf, &error_msg);
CHECK(dex_file != nullptr) << error_msg;
ObjPtr<mirror::DexCache> dex_cache =
class_linker->RegisterDexFile(*dex_file, nullptr);
CHECK(dex_cache != nullptr);
class_path->push_back(dex_file);
}
// Need a class loader. Fake that we're a compiler.
// Note: this will run initializers through the unstarted runtime, so make sure it's
// initialized.
interpreter::UnstartedRuntime::Initialize();
jobject class_loader = class_linker->CreatePathClassLoader(self, *class_path);
return class_loader;
}
static int DumpOatWithRuntime(Runtime* runtime,
std::unique_ptr<OatFile> oat_file,
OatDumperOptions* options,
std::ostream* os) {
CHECK(runtime != nullptr && oat_file != nullptr && options != nullptr);
ScopedObjectAccess soa(Thread::Current());
OatFile* oat_file_ptr = oat_file.get();
std::vector<const DexFile*> class_path;
jobject class_loader = InstallOatFile(runtime, std::move(oat_file), &class_path);
// Use the class loader while dumping.
StackHandleScope<1> scope(soa.Self());
Handle<mirror::ClassLoader> loader_handle = scope.NewHandle(
soa.Decode<mirror::ClassLoader>(class_loader));
options->class_loader_ = &loader_handle;
OatDumper oat_dumper(*oat_file_ptr, *options);
bool success = oat_dumper.Dump(*os);
return (success) ? EXIT_SUCCESS : EXIT_FAILURE;
}
static int DumpOatWithoutRuntime(OatFile* oat_file, OatDumperOptions* options, std::ostream* os) {
CHECK(oat_file != nullptr && options != nullptr);
// No image = no class loader.
ScopedNullHandle<mirror::ClassLoader> null_class_loader;
options->class_loader_ = &null_class_loader;
OatDumper oat_dumper(*oat_file, *options);
bool success = oat_dumper.Dump(*os);
return (success) ? EXIT_SUCCESS : EXIT_FAILURE;
}
static int DumpOat(Runtime* runtime, const char* oat_filename, OatDumperOptions* options,
std::ostream* os) {
std::string error_msg;
std::unique_ptr<OatFile> oat_file(OatFile::Open(oat_filename,
oat_filename,
nullptr,
nullptr,
false,
/*low_4gb*/false,
nullptr,
&error_msg));
if (oat_file == nullptr) {
fprintf(stderr, "Failed to open oat file from '%s': %s\n", oat_filename, error_msg.c_str());
return EXIT_FAILURE;
}
if (runtime != nullptr) {
return DumpOatWithRuntime(runtime, std::move(oat_file), options, os);
} else {
return DumpOatWithoutRuntime(oat_file.get(), options, os);
}
}
static int SymbolizeOat(const char* oat_filename, std::string& output_name, bool no_bits) {
std::string error_msg;
std::unique_ptr<OatFile> oat_file(OatFile::Open(oat_filename,
oat_filename,
nullptr,
nullptr,
false,
/*low_4gb*/false,
nullptr,
&error_msg));
if (oat_file == nullptr) {
fprintf(stderr, "Failed to open oat file from '%s': %s\n", oat_filename, error_msg.c_str());
return EXIT_FAILURE;
}
bool result;
// Try to produce an ELF file of the same type. This is finicky, as we have used 32-bit ELF
// files for 64-bit code in the past.
if (Is64BitInstructionSet(oat_file->GetOatHeader().GetInstructionSet())) {
OatSymbolizer<ElfTypes64> oat_symbolizer(oat_file.get(), output_name, no_bits);
result = oat_symbolizer.Symbolize();
} else {
OatSymbolizer<ElfTypes32> oat_symbolizer(oat_file.get(), output_name, no_bits);
result = oat_symbolizer.Symbolize();
}
if (!result) {
fprintf(stderr, "Failed to symbolize\n");
return EXIT_FAILURE;
}
return EXIT_SUCCESS;
}
class IMTDumper {
public:
static bool Dump(Runtime* runtime,
const std::string& imt_file,
bool dump_imt_stats,
const char* oat_filename) {
Thread* self = Thread::Current();
ScopedObjectAccess soa(self);
StackHandleScope<1> scope(self);
MutableHandle<mirror::ClassLoader> class_loader = scope.NewHandle<mirror::ClassLoader>(nullptr);
std::vector<const DexFile*> class_path;
if (oat_filename != nullptr) {
std::string error_msg;
std::unique_ptr<OatFile> oat_file(OatFile::Open(oat_filename,
oat_filename,
nullptr,
nullptr,
false,
/*low_4gb*/false,
nullptr,
&error_msg));
if (oat_file == nullptr) {
fprintf(stderr, "Failed to open oat file from '%s': %s\n", oat_filename, error_msg.c_str());
return false;
}
class_loader.Assign(soa.Decode<mirror::ClassLoader>(
InstallOatFile(runtime, std::move(oat_file), &class_path)));
} else {
class_loader.Assign(nullptr); // Boot classloader. Just here for explicit documentation.
class_path = runtime->GetClassLinker()->GetBootClassPath();
}
if (!imt_file.empty()) {
return DumpImt(runtime, imt_file, class_loader);
}
if (dump_imt_stats) {
return DumpImtStats(runtime, class_path, class_loader);
}
LOG(FATAL) << "Should not reach here";
UNREACHABLE();
}
private:
static bool DumpImt(Runtime* runtime,
const std::string& imt_file,
Handle<mirror::ClassLoader> h_class_loader)
REQUIRES_SHARED(Locks::mutator_lock_) {
std::vector<std::string> lines = ReadCommentedInputFromFile(imt_file);
std::unordered_set<std::string> prepared;
for (const std::string& line : lines) {
// A line should be either a class descriptor, in which case we will dump the complete IMT,
// or a class descriptor and an interface method, in which case we will lookup the method,
// determine its IMT slot, and check the class' IMT.
size_t first_space = line.find(' ');
if (first_space == std::string::npos) {
DumpIMTForClass(runtime, line, h_class_loader, &prepared);
} else {
DumpIMTForMethod(runtime,
line.substr(0, first_space),
line.substr(first_space + 1, std::string::npos),
h_class_loader,
&prepared);
}
std::cerr << std::endl;
}
return true;
}
static bool DumpImtStats(Runtime* runtime,
const std::vector<const DexFile*>& dex_files,
Handle<mirror::ClassLoader> h_class_loader)
REQUIRES_SHARED(Locks::mutator_lock_) {
size_t without_imt = 0;
size_t with_imt = 0;
std::map<size_t, size_t> histogram;
ClassLinker* class_linker = runtime->GetClassLinker();
const PointerSize pointer_size = class_linker->GetImagePointerSize();
std::unordered_set<std::string> prepared;
Thread* self = Thread::Current();
StackHandleScope<1> scope(self);
MutableHandle<mirror::Class> h_klass(scope.NewHandle<mirror::Class>(nullptr));
for (const DexFile* dex_file : dex_files) {
for (uint32_t class_def_index = 0;
class_def_index != dex_file->NumClassDefs();
++class_def_index) {
const DexFile::ClassDef& class_def = dex_file->GetClassDef(class_def_index);
const char* descriptor = dex_file->GetClassDescriptor(class_def);
h_klass.Assign(class_linker->FindClass(self, descriptor, h_class_loader));
if (h_klass == nullptr) {
std::cerr << "Warning: could not load " << descriptor << std::endl;
continue;
}
if (HasNoIMT(runtime, h_klass, pointer_size, &prepared)) {
without_imt++;
continue;
}
ImTable* im_table = PrepareAndGetImTable(runtime, h_klass, pointer_size, &prepared);
if (im_table == nullptr) {
// Should not happen, but accept.
without_imt++;
continue;
}
with_imt++;
for (size_t imt_index = 0; imt_index != ImTable::kSize; ++imt_index) {
ArtMethod* ptr = im_table->Get(imt_index, pointer_size);
if (ptr->IsRuntimeMethod()) {
if (ptr->IsImtUnimplementedMethod()) {
histogram[0]++;
} else {
ImtConflictTable* current_table = ptr->GetImtConflictTable(pointer_size);
histogram[current_table->NumEntries(pointer_size)]++;
}
} else {
histogram[1]++;
}
}
}
}
std::cerr << "IMT stats:"
<< std::endl << std::endl;
std::cerr << " " << with_imt << " classes with IMT."
<< std::endl << std::endl;
std::cerr << " " << without_imt << " classes without IMT (or copy from Object)."
<< std::endl << std::endl;
double sum_one = 0;
size_t count_one = 0;
std::cerr << " " << "IMT histogram" << std::endl;
for (auto& bucket : histogram) {
std::cerr << " " << bucket.first << " " << bucket.second << std::endl;
if (bucket.first > 0) {
sum_one += bucket.second * bucket.first;
count_one += bucket.second;
}
}
double count_zero = count_one + histogram[0];
std::cerr << " Stats:" << std::endl;
std::cerr << " Average depth (including empty): " << (sum_one / count_zero) << std::endl;
std::cerr << " Average depth (excluding empty): " << (sum_one / count_one) << std::endl;
return true;
}
// Return whether the given class has no IMT (or the one shared with java.lang.Object).
static bool HasNoIMT(Runtime* runtime,
Handle<mirror::Class> klass,
const PointerSize pointer_size,
std::unordered_set<std::string>* prepared)
REQUIRES_SHARED(Locks::mutator_lock_) {
if (klass->IsObjectClass() || !klass->ShouldHaveImt()) {
return true;
}
if (klass->GetImt(pointer_size) == nullptr) {
PrepareClass(runtime, klass, prepared);
}
mirror::Class* object_class = mirror::Class::GetJavaLangClass()->GetSuperClass();
DCHECK(object_class->IsObjectClass());
bool result = klass->GetImt(pointer_size) == object_class->GetImt(pointer_size);
if (klass->GetIfTable()->Count() == 0) {
DCHECK(result);
}
return result;
}
static void PrintTable(ImtConflictTable* table, PointerSize pointer_size)
REQUIRES_SHARED(Locks::mutator_lock_) {
if (table == nullptr) {
std::cerr << " <No IMT?>" << std::endl;
return;
}
size_t table_index = 0;
for (;;) {
ArtMethod* ptr = table->GetInterfaceMethod(table_index, pointer_size);
if (ptr == nullptr) {
return;
}
table_index++;
std::cerr << " " << ptr->PrettyMethod(true) << std::endl;
}
}
static ImTable* PrepareAndGetImTable(Runtime* runtime,
Thread* self,
Handle<mirror::ClassLoader> h_loader,
const std::string& class_name,
const PointerSize pointer_size,
mirror::Class** klass_out,
std::unordered_set<std::string>* prepared)
REQUIRES_SHARED(Locks::mutator_lock_) {
if (class_name.empty()) {
return nullptr;
}
std::string descriptor;
if (class_name[0] == 'L') {
descriptor = class_name;
} else {
descriptor = DotToDescriptor(class_name.c_str());
}
mirror::Class* klass = runtime->GetClassLinker()->FindClass(self, descriptor.c_str(), h_loader);
if (klass == nullptr) {
self->ClearException();
std::cerr << "Did not find " << class_name << std::endl;
*klass_out = nullptr;
return nullptr;
}
StackHandleScope<1> scope(Thread::Current());
Handle<mirror::Class> h_klass = scope.NewHandle<mirror::Class>(klass);
ImTable* ret = PrepareAndGetImTable(runtime, h_klass, pointer_size, prepared);
*klass_out = h_klass.Get();
return ret;
}
static ImTable* PrepareAndGetImTable(Runtime* runtime,
Handle<mirror::Class> h_klass,
const PointerSize pointer_size,
std::unordered_set<std::string>* prepared)
REQUIRES_SHARED(Locks::mutator_lock_) {
PrepareClass(runtime, h_klass, prepared);
return h_klass->GetImt(pointer_size);
}
static void DumpIMTForClass(Runtime* runtime,
const std::string& class_name,
Handle<mirror::ClassLoader> h_loader,
std::unordered_set<std::string>* prepared)
REQUIRES_SHARED(Locks::mutator_lock_) {
const PointerSize pointer_size = runtime->GetClassLinker()->GetImagePointerSize();
mirror::Class* klass;
ImTable* imt = PrepareAndGetImTable(runtime,
Thread::Current(),
h_loader,
class_name,
pointer_size,
&klass,
prepared);
if (imt == nullptr) {
return;
}
std::cerr << class_name << std::endl << " IMT:" << std::endl;
for (size_t index = 0; index < ImTable::kSize; ++index) {
std::cerr << " " << index << ":" << std::endl;
ArtMethod* ptr = imt->Get(index, pointer_size);
if (ptr->IsRuntimeMethod()) {
if (ptr->IsImtUnimplementedMethod()) {
std::cerr << " <empty>" << std::endl;
} else {
ImtConflictTable* current_table = ptr->GetImtConflictTable(pointer_size);
PrintTable(current_table, pointer_size);
}
} else {
std::cerr << " " << ptr->PrettyMethod(true) << std::endl;
}
}
std::cerr << " Interfaces:" << std::endl;
// Run through iftable, find methods that slot here, see if they fit.
mirror::IfTable* if_table = klass->GetIfTable();
for (size_t i = 0, num_interfaces = klass->GetIfTableCount(); i < num_interfaces; ++i) {
mirror::Class* iface = if_table->GetInterface(i);
std::string iface_name;
std::cerr << " " << iface->GetDescriptor(&iface_name) << std::endl;
for (ArtMethod& iface_method : iface->GetVirtualMethods(pointer_size)) {
uint32_t class_hash, name_hash, signature_hash;
ImTable::GetImtHashComponents(&iface_method, &class_hash, &name_hash, &signature_hash);
uint32_t imt_slot = ImTable::GetImtIndex(&iface_method);
std::cerr << " " << iface_method.PrettyMethod(true)
<< " slot=" << imt_slot
<< std::hex
<< " class_hash=0x" << class_hash
<< " name_hash=0x" << name_hash
<< " signature_hash=0x" << signature_hash
<< std::dec
<< std::endl;
}
}
}
static void DumpIMTForMethod(Runtime* runtime,
const std::string& class_name,
const std::string& method,
Handle<mirror::ClassLoader> h_loader,
std::unordered_set<std::string>* prepared)
REQUIRES_SHARED(Locks::mutator_lock_) {
const PointerSize pointer_size = runtime->GetClassLinker()->GetImagePointerSize();
mirror::Class* klass;
ImTable* imt = PrepareAndGetImTable(runtime,
Thread::Current(),
h_loader,
class_name,
pointer_size,
&klass,
prepared);
if (imt == nullptr) {
return;
}
std::cerr << class_name << " <" << method << ">" << std::endl;
for (size_t index = 0; index < ImTable::kSize; ++index) {
ArtMethod* ptr = imt->Get(index, pointer_size);
if (ptr->IsRuntimeMethod()) {
if (ptr->IsImtUnimplementedMethod()) {
continue;
}
ImtConflictTable* current_table = ptr->GetImtConflictTable(pointer_size);
if (current_table == nullptr) {
continue;
}
size_t table_index = 0;
for (;;) {
ArtMethod* ptr2 = current_table->GetInterfaceMethod(table_index, pointer_size);
if (ptr2 == nullptr) {
break;
}
table_index++;
std::string p_name = ptr2->PrettyMethod(true);
if (android::base::StartsWith(p_name, method.c_str())) {
std::cerr << " Slot "
<< index
<< " ("
<< current_table->NumEntries(pointer_size)
<< ")"
<< std::endl;
PrintTable(current_table, pointer_size);
return;
}
}
} else {
std::string p_name = ptr->PrettyMethod(true);
if (android::base::StartsWith(p_name, method.c_str())) {
std::cerr << " Slot " << index << " (1)" << std::endl;
std::cerr << " " << p_name << std::endl;
} else {
// Run through iftable, find methods that slot here, see if they fit.
mirror::IfTable* if_table = klass->GetIfTable();
for (size_t i = 0, num_interfaces = klass->GetIfTableCount(); i < num_interfaces; ++i) {
mirror::Class* iface = if_table->GetInterface(i);
size_t num_methods = iface->NumDeclaredVirtualMethods();
if (num_methods > 0) {
for (ArtMethod& iface_method : iface->GetMethods(pointer_size)) {
if (ImTable::GetImtIndex(&iface_method) == index) {
std::string i_name = iface_method.PrettyMethod(true);
if (android::base::StartsWith(i_name, method.c_str())) {
std::cerr << " Slot " << index << " (1)" << std::endl;
std::cerr << " " << p_name << " (" << i_name << ")" << std::endl;
}
}
}
}
}
}
}
}
}
// Read lines from the given stream, dropping comments and empty lines
static std::vector<std::string> ReadCommentedInputStream(std::istream& in_stream) {
std::vector<std::string> output;
while (in_stream.good()) {
std::string dot;
std::getline(in_stream, dot);
if (android::base::StartsWith(dot, "#") || dot.empty()) {
continue;
}
output.push_back(dot);
}
return output;
}
// Read lines from the given file, dropping comments and empty lines.
static std::vector<std::string> ReadCommentedInputFromFile(const std::string& input_filename) {
std::unique_ptr<std::ifstream> input_file(new std::ifstream(input_filename, std::ifstream::in));
if (input_file.get() == nullptr) {
LOG(ERROR) << "Failed to open input file " << input_filename;
return std::vector<std::string>();
}
std::vector<std::string> result = ReadCommentedInputStream(*input_file);
input_file->close();
return result;
}
// Prepare a class, i.e., ensure it has a filled IMT. Will do so recursively for superclasses,
// and note in the given set that the work was done.
static void PrepareClass(Runtime* runtime,
Handle<mirror::Class> h_klass,
std::unordered_set<std::string>* done)
REQUIRES_SHARED(Locks::mutator_lock_) {
if (!h_klass->ShouldHaveImt()) {
return;
}
std::string name;
name = h_klass->GetDescriptor(&name);
if (done->find(name) != done->end()) {
return;
}
done->insert(name);
if (h_klass->HasSuperClass()) {
StackHandleScope<1> h(Thread::Current());
PrepareClass(runtime, h.NewHandle<mirror::Class>(h_klass->GetSuperClass()), done);
}
if (!h_klass->IsTemp()) {
runtime->GetClassLinker()->FillIMTAndConflictTables(h_klass.Get());
}
}
};
struct OatdumpArgs : public CmdlineArgs {
protected:
using Base = CmdlineArgs;
virtual ParseStatus ParseCustom(const StringPiece& option,
std::string* error_msg) OVERRIDE {
{
ParseStatus base_parse = Base::ParseCustom(option, error_msg);
if (base_parse != kParseUnknownArgument) {
return base_parse;
}
}
if (option.starts_with("--oat-file=")) {
oat_filename_ = option.substr(strlen("--oat-file=")).data();
} else if (option.starts_with("--image=")) {
image_location_ = option.substr(strlen("--image=")).data();
} else if (option == "--no-dump:vmap") {
dump_vmap_ = false;
} else if (option =="--dump:code_info_stack_maps") {
dump_code_info_stack_maps_ = true;
} else if (option == "--no-disassemble") {
disassemble_code_ = false;
} else if (option =="--header-only") {
dump_header_only_ = true;
} else if (option.starts_with("--symbolize=")) {
oat_filename_ = option.substr(strlen("--symbolize=")).data();
symbolize_ = true;
} else if (option.starts_with("--only-keep-debug")) {
only_keep_debug_ = true;
} else if (option.starts_with("--class-filter=")) {
class_filter_ = option.substr(strlen("--class-filter=")).data();
} else if (option.starts_with("--method-filter=")) {
method_filter_ = option.substr(strlen("--method-filter=")).data();
} else if (option.starts_with("--list-classes")) {
list_classes_ = true;
} else if (option.starts_with("--list-methods")) {
list_methods_ = true;
} else if (option.starts_with("--export-dex-to=")) {
export_dex_location_ = option.substr(strlen("--export-dex-to=")).data();
} else if (option.starts_with("--addr2instr=")) {
if (!ParseUint(option.substr(strlen("--addr2instr=")).data(), &addr2instr_)) {
*error_msg = "Address conversion failed";
return kParseError;
}
} else if (option.starts_with("--app-image=")) {
app_image_ = option.substr(strlen("--app-image=")).data();
} else if (option.starts_with("--app-oat=")) {
app_oat_ = option.substr(strlen("--app-oat=")).data();
} else if (option.starts_with("--dump-imt=")) {
imt_dump_ = option.substr(strlen("--dump-imt=")).data();
} else if (option == "--dump-imt-stats") {
imt_stat_dump_ = true;
} else {
return kParseUnknownArgument;
}
return kParseOk;
}
virtual ParseStatus ParseChecks(std::string* error_msg) OVERRIDE {
// Infer boot image location from the image location if possible.
if (boot_image_location_ == nullptr) {
boot_image_location_ = image_location_;
}
// Perform the parent checks.
ParseStatus parent_checks = Base::ParseChecks(error_msg);
if (parent_checks != kParseOk) {
return parent_checks;
}
// Perform our own checks.
if (image_location_ == nullptr && oat_filename_ == nullptr) {
*error_msg = "Either --image or --oat-file must be specified";
return kParseError;
} else if (image_location_ != nullptr && oat_filename_ != nullptr) {
*error_msg = "Either --image or --oat-file must be specified but not both";
return kParseError;
}
return kParseOk;
}
virtual std::string GetUsage() const {
std::string usage;
usage +=
"Usage: oatdump [options] ...\n"
" Example: oatdump --image=$ANDROID_PRODUCT_OUT/system/framework/boot.art\n"
" Example: adb shell oatdump --image=/system/framework/boot.art\n"
"\n"
// Either oat-file or image is required.
" --oat-file=<file.oat>: specifies an input oat filename.\n"
" Example: --oat-file=/system/framework/boot.oat\n"
"\n"
" --image=<file.art>: specifies an input image location.\n"
" Example: --image=/system/framework/boot.art\n"
"\n"
" --app-image=<file.art>: specifies an input app image. Must also have a specified\n"
" boot image (with --image) and app oat file (with --app-oat).\n"
" Example: --app-image=app.art\n"
"\n"
" --app-oat=<file.odex>: specifies an input app oat.\n"
" Example: --app-oat=app.odex\n"
"\n";
usage += Base::GetUsage();
usage += // Optional.
" --no-dump:vmap may be used to disable vmap dumping.\n"
" Example: --no-dump:vmap\n"
"\n"
" --dump:code_info_stack_maps enables dumping of stack maps in CodeInfo sections.\n"
" Example: --dump:code_info_stack_maps\n"
"\n"
" --no-disassemble may be used to disable disassembly.\n"
" Example: --no-disassemble\n"
"\n"
" --header-only may be used to print only the oat header.\n"
" Example: --header-only\n"
"\n"
" --list-classes may be used to list target file classes (can be used with filters).\n"
" Example: --list-classes\n"
" Example: --list-classes --class-filter=com.example.foo\n"
"\n"
" --list-methods may be used to list target file methods (can be used with filters).\n"
" Example: --list-methods\n"
" Example: --list-methods --class-filter=com.example --method-filter=foo\n"
"\n"
" --symbolize=<file.oat>: output a copy of file.oat with elf symbols included.\n"
" Example: --symbolize=/system/framework/boot.oat\n"
"\n"
" --only-keep-debug<file.oat>: Modifies the behaviour of --symbolize so that\n"
" .rodata and .text sections are omitted in the output file to save space.\n"
" Example: --symbolize=/system/framework/boot.oat --only-keep-debug\n"
"\n"
" --class-filter=<class name>: only dumps classes that contain the filter.\n"
" Example: --class-filter=com.example.foo\n"
"\n"
" --method-filter=<method name>: only dumps methods that contain the filter.\n"
" Example: --method-filter=foo\n"
"\n"
" --export-dex-to=<directory>: may be used to export oat embedded dex files.\n"
" Example: --export-dex-to=/data/local/tmp\n"
"\n"
" --addr2instr=<address>: output matching method disassembled code from relative\n"
" address (e.g. PC from crash dump)\n"
" Example: --addr2instr=0x00001a3b\n"
"\n"
" --dump-imt=<file.txt>: output IMT collisions (if any) for the given receiver\n"
" types and interface methods in the given file. The file\n"
" is read line-wise, where each line should either be a class\n"
" name or descriptor, or a class name/descriptor and a prefix\n"
" of a complete method name (separated by a whitespace).\n"
" Example: --dump-imt=imt.txt\n"
"\n"
" --dump-imt-stats: output IMT statistics for the given boot image\n"
" Example: --dump-imt-stats"
"\n";
return usage;
}
public:
const char* oat_filename_ = nullptr;
const char* class_filter_ = "";
const char* method_filter_ = "";
const char* image_location_ = nullptr;
std::string elf_filename_prefix_;
std::string imt_dump_;
bool dump_vmap_ = true;
bool dump_code_info_stack_maps_ = false;
bool disassemble_code_ = true;
bool symbolize_ = false;
bool only_keep_debug_ = false;
bool list_classes_ = false;
bool list_methods_ = false;
bool dump_header_only_ = false;
bool imt_stat_dump_ = false;
uint32_t addr2instr_ = 0;
const char* export_dex_location_ = nullptr;
const char* app_image_ = nullptr;
const char* app_oat_ = nullptr;
};
struct OatdumpMain : public CmdlineMain<OatdumpArgs> {
virtual bool NeedsRuntime() OVERRIDE {
CHECK(args_ != nullptr);
// If we are only doing the oat file, disable absolute_addresses. Keep them for image dumping.
bool absolute_addresses = (args_->oat_filename_ == nullptr);
oat_dumper_options_.reset(new OatDumperOptions(
args_->dump_vmap_,
args_->dump_code_info_stack_maps_,
args_->disassemble_code_,
absolute_addresses,
args_->class_filter_,
args_->method_filter_,
args_->list_classes_,
args_->list_methods_,
args_->dump_header_only_,
args_->export_dex_location_,
args_->app_image_,
args_->app_oat_,
args_->addr2instr_));
return (args_->boot_image_location_ != nullptr ||
args_->image_location_ != nullptr ||
!args_->imt_dump_.empty()) &&
!args_->symbolize_;
}
virtual bool ExecuteWithoutRuntime() OVERRIDE {
CHECK(args_ != nullptr);
CHECK(args_->oat_filename_ != nullptr);
MemMap::Init();
if (args_->symbolize_) {
// ELF has special kind of section called SHT_NOBITS which allows us to create
// sections which exist but their data is omitted from the ELF file to save space.
// This is what "strip --only-keep-debug" does when it creates separate ELF file
// with only debug data. We use it in similar way to exclude .rodata and .text.
bool no_bits = args_->only_keep_debug_;
return SymbolizeOat(args_->oat_filename_, args_->output_name_, no_bits) == EXIT_SUCCESS;
} else {
return DumpOat(nullptr,
args_->oat_filename_,
oat_dumper_options_.get(),
args_->os_) == EXIT_SUCCESS;
}
}
virtual bool ExecuteWithRuntime(Runtime* runtime) {
CHECK(args_ != nullptr);
if (!args_->imt_dump_.empty() || args_->imt_stat_dump_) {
return IMTDumper::Dump(runtime,
args_->imt_dump_,
args_->imt_stat_dump_,
args_->oat_filename_);
}
if (args_->oat_filename_ != nullptr) {
return DumpOat(runtime,
args_->oat_filename_,
oat_dumper_options_.get(),
args_->os_) == EXIT_SUCCESS;
}
return DumpImages(runtime, oat_dumper_options_.get(), args_->os_) == EXIT_SUCCESS;
}
std::unique_ptr<OatDumperOptions> oat_dumper_options_;
};
} // namespace art
int main(int argc, char** argv) {
art::OatdumpMain main;
return main.Main(argc, argv);
}