blob: 04cc33da7bf4df6553152edd76eddca77f797b44 [file] [log] [blame]
/*
* Copyright (C) 2018 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "src/profiling/memory/bookkeeping.h"
#include <fcntl.h>
#include <inttypes.h>
#include <sys/stat.h>
#include <sys/types.h>
#include "perfetto/base/file_utils.h"
#include "perfetto/base/logging.h"
#include "perfetto/base/scoped_file.h"
namespace perfetto {
namespace profiling {
namespace {
using ::perfetto::protos::pbzero::ProfilePacket;
// This needs to be lower than the maximum acceptable chunk size, because this
// is checked *before* writing another submessage. We conservatively assume
// submessages can be up to 100k here for a 500k chunk size.
// DropBox has a 500k chunk limit, and each chunk needs to parse as a proto.
uint32_t kPacketSizeThreshold = 400000;
}
GlobalCallstackTrie::Node* GlobalCallstackTrie::Node::GetOrCreateChild(
const Interned<Frame>& loc) {
Node* child = children_.Get(loc);
if (!child)
child = children_.Emplace(loc, this);
return child;
}
void HeapTracker::RecordMalloc(const std::vector<FrameData>& callstack,
uint64_t address,
uint64_t size,
uint64_t sequence_number,
uint64_t timestamp) {
auto it = allocations_.find(address);
if (it != allocations_.end()) {
Allocation& alloc = it->second;
PERFETTO_DCHECK(alloc.sequence_number != sequence_number);
if (alloc.sequence_number < sequence_number) {
// As we are overwriting the previous allocation, the previous allocation
// must have been freed.
//
// This makes the sequencing a bit incorrect. We are overwriting this
// allocation, so we prentend both the alloc and the free for this have
// already happened at committed_sequence_number_, while in fact the free
// might not have happened until right before this operation.
if (alloc.sequence_number > committed_sequence_number_) {
// Only count the previous allocation if it hasn't already been
// committed to avoid double counting it.
alloc.AddToCallstackAllocations();
}
alloc.SubtractFromCallstackAllocations();
GlobalCallstackTrie::Node* node = callsites_->CreateCallsite(callstack);
alloc.total_size = size;
alloc.sequence_number = sequence_number;
alloc.SetCallstackAllocations(MaybeCreateCallstackAllocations(node));
}
} else {
GlobalCallstackTrie::Node* node = callsites_->CreateCallsite(callstack);
allocations_.emplace(address,
Allocation(size, sequence_number,
MaybeCreateCallstackAllocations(node)));
}
RecordOperation(sequence_number, {address, timestamp});
}
void HeapTracker::RecordOperation(uint64_t sequence_number,
const PendingOperation& operation) {
if (sequence_number != committed_sequence_number_ + 1) {
pending_operations_.emplace(sequence_number, operation);
return;
}
CommitOperation(sequence_number, operation);
// At this point some other pending operations might be eligible to be
// committed.
auto it = pending_operations_.begin();
while (it != pending_operations_.end() &&
it->first == committed_sequence_number_ + 1) {
CommitOperation(it->first, it->second);
it = pending_operations_.erase(it);
}
}
void HeapTracker::CommitOperation(uint64_t sequence_number,
const PendingOperation& operation) {
committed_sequence_number_++;
committed_timestamp_ = operation.timestamp;
uint64_t address = operation.allocation_address;
// We will see many frees for addresses we do not know about.
auto leaf_it = allocations_.find(address);
if (leaf_it == allocations_.end())
return;
Allocation& value = leaf_it->second;
if (value.sequence_number == sequence_number) {
value.AddToCallstackAllocations();
} else if (value.sequence_number < sequence_number) {
value.SubtractFromCallstackAllocations();
allocations_.erase(leaf_it);
}
// else (value.sequence_number > sequence_number:
// This allocation has been replaced by a newer one in RecordMalloc.
// This code commits ther previous allocation's malloc (and implicit free
// that must have happened, as there is now a new allocation at the same
// address). This means that this operation, be it a malloc or a free, must
// be treated as a no-op.
}
void HeapTracker::Dump(
std::function<void(ProfilePacket::ProcessHeapSamples*)> fill_process_header,
DumpState* dump_state) {
// There are two reasons we remove the unused callstack allocations on the
// next iteration of Dump:
// * We need to remove them after the callstacks were dumped, which currently
// happens after the allocations are dumped.
// * This way, we do not destroy and recreate callstacks as frequently.
for (auto it_and_alloc : dead_callstack_allocations_) {
auto& it = it_and_alloc.first;
uint64_t allocated = it_and_alloc.second;
const CallstackAllocations& alloc = it->second;
if (alloc.allocs == 0 && alloc.allocation_count == allocated)
callstack_allocations_.erase(it);
}
dead_callstack_allocations_.clear();
if (dump_state->currently_written() > kPacketSizeThreshold)
dump_state->NewProfilePacket();
ProfilePacket::ProcessHeapSamples* proto =
dump_state->current_profile_packet->add_process_dumps();
fill_process_header(proto);
proto->set_timestamp(committed_timestamp_);
for (auto it = callstack_allocations_.begin();
it != callstack_allocations_.end(); ++it) {
if (dump_state->currently_written() > kPacketSizeThreshold) {
dump_state->NewProfilePacket();
proto = dump_state->current_profile_packet->add_process_dumps();
fill_process_header(proto);
proto->set_timestamp(committed_timestamp_);
}
const CallstackAllocations& alloc = it->second;
dump_state->callstacks_to_dump.emplace(alloc.node);
ProfilePacket::HeapSample* sample = proto->add_samples();
sample->set_callstack_id(alloc.node->id());
sample->set_self_allocated(alloc.allocated);
sample->set_self_freed(alloc.freed);
sample->set_alloc_count(alloc.allocation_count);
sample->set_free_count(alloc.free_count);
if (alloc.allocs == 0)
dead_callstack_allocations_.emplace_back(it, alloc.allocation_count);
}
}
uint64_t HeapTracker::GetSizeForTesting(const std::vector<FrameData>& stack) {
GlobalCallstackTrie::Node* node = callsites_->CreateCallsite(stack);
// Hack to make it go away again if it wasn't used before.
// This is only good because this is used for testing only.
GlobalCallstackTrie::IncrementNode(node);
GlobalCallstackTrie::DecrementNode(node);
auto it = callstack_allocations_.find(node);
if (it == callstack_allocations_.end()) {
return 0;
}
const CallstackAllocations& alloc = it->second;
return alloc.allocated - alloc.freed;
}
std::vector<Interned<Frame>> GlobalCallstackTrie::BuildCallstack(
const Node* node) const {
std::vector<Interned<Frame>> res;
while (node != &root_) {
res.emplace_back(node->location_);
node = node->parent_;
}
return res;
}
GlobalCallstackTrie::Node* GlobalCallstackTrie::CreateCallsite(
const std::vector<FrameData>& callstack) {
Node* node = &root_;
for (const FrameData& loc : callstack) {
node = node->GetOrCreateChild(InternCodeLocation(loc));
}
return node;
}
void GlobalCallstackTrie::IncrementNode(Node* node) {
while (node != nullptr) {
node->ref_count_ += 1;
node = node->parent_;
}
}
void GlobalCallstackTrie::DecrementNode(Node* node) {
PERFETTO_DCHECK(node->ref_count_ >= 1);
bool delete_prev = false;
Node* prev = nullptr;
while (node != nullptr) {
if (delete_prev)
node->children_.Remove(*prev);
node->ref_count_ -= 1;
delete_prev = node->ref_count_ == 0;
prev = node;
node = node->parent_;
}
}
Interned<Frame> GlobalCallstackTrie::InternCodeLocation(const FrameData& loc) {
Mapping map(string_interner_.Intern(loc.build_id));
map.offset = loc.frame.map_elf_start_offset;
map.start = loc.frame.map_start;
map.end = loc.frame.map_end;
map.load_bias = loc.frame.map_load_bias;
base::StringSplitter sp(loc.frame.map_name, '/');
while (sp.Next())
map.path_components.emplace_back(string_interner_.Intern(sp.cur_token()));
Frame frame(mapping_interner_.Intern(std::move(map)),
string_interner_.Intern(loc.frame.function_name),
loc.frame.rel_pc);
return frame_interner_.Intern(frame);
}
Interned<Frame> GlobalCallstackTrie::MakeRootFrame() {
Mapping map(string_interner_.Intern(""));
Frame frame(mapping_interner_.Intern(std::move(map)),
string_interner_.Intern(""), 0);
return frame_interner_.Intern(frame);
}
void DumpState::WriteMap(const Interned<Mapping> map) {
auto map_it_and_inserted = dumped_mappings.emplace(map.id());
if (map_it_and_inserted.second) {
for (const Interned<std::string>& str : map->path_components)
WriteString(str);
WriteString(map->build_id);
if (currently_written() > kPacketSizeThreshold)
NewProfilePacket();
auto mapping = current_profile_packet->add_mappings();
mapping->set_id(map.id());
mapping->set_offset(map->offset);
mapping->set_start(map->start);
mapping->set_end(map->end);
mapping->set_load_bias(map->load_bias);
mapping->set_build_id(map->build_id.id());
for (const Interned<std::string>& str : map->path_components)
mapping->add_path_string_ids(str.id());
}
}
void DumpState::WriteFrame(Interned<Frame> frame) {
WriteMap(frame->mapping);
WriteString(frame->function_name);
bool inserted;
std::tie(std::ignore, inserted) = dumped_frames.emplace(frame.id());
if (inserted) {
if (currently_written() > kPacketSizeThreshold)
NewProfilePacket();
auto frame_proto = current_profile_packet->add_frames();
frame_proto->set_id(frame.id());
frame_proto->set_function_name_id(frame->function_name.id());
frame_proto->set_mapping_id(frame->mapping.id());
frame_proto->set_rel_pc(frame->rel_pc);
}
}
void DumpState::WriteString(const Interned<std::string>& str) {
bool inserted;
std::tie(std::ignore, inserted) = dumped_strings.emplace(str.id());
if (inserted) {
if (currently_written() > kPacketSizeThreshold)
NewProfilePacket();
auto interned_string = current_profile_packet->add_strings();
interned_string->set_id(str.id());
interned_string->set_str(reinterpret_cast<const uint8_t*>(str->c_str()),
str->size());
}
}
} // namespace profiling
} // namespace perfetto