// Copyright (c) 2010 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include #include #include // for CreateToolhelp32Snapshot() #include #include "tools/memory_watcher/memory_watcher.h" #include "base/file_util.h" #include "base/logging.h" #include "base/metrics/stats_counters.h" #include "base/strings/string_util.h" #include "base/strings/utf_string_conversions.h" #include "base/synchronization/lock.h" #include "tools/memory_watcher/call_stack.h" #include "tools/memory_watcher/preamble_patcher.h" static base::StatsCounter mem_in_use("MemoryInUse.Bytes"); static base::StatsCounter mem_in_use_blocks("MemoryInUse.Blocks"); static base::StatsCounter mem_in_use_allocs("MemoryInUse.Allocs"); static base::StatsCounter mem_in_use_frees("MemoryInUse.Frees"); // --------------------------------------------------------------------- MemoryWatcher::MemoryWatcher() : file_(NULL), hooked_(false), active_thread_id_(0) { MemoryHook::Initialize(); CallStack::Initialize(); block_map_ = new CallStackMap(); // Register last - only after we're ready for notifications! Hook(); } MemoryWatcher::~MemoryWatcher() { Unhook(); CloseLogFile(); // Pointers in the block_map are part of the MemoryHook heap. Be sure // to delete the map before closing the heap. delete block_map_; } void MemoryWatcher::Hook() { DCHECK(!hooked_); MemoryHook::RegisterWatcher(this); hooked_ = true; } void MemoryWatcher::Unhook() { if (hooked_) { MemoryHook::UnregisterWatcher(this); hooked_ = false; } } void MemoryWatcher::OpenLogFile() { DCHECK(file_ == NULL); file_name_ = "memwatcher"; if (!log_name_.empty()) { file_name_ += "."; file_name_ += log_name_; } file_name_ += ".log"; char buf[16]; file_name_ += _itoa(GetCurrentProcessId(), buf, 10); std::string tmp_name(file_name_); tmp_name += ".tmp"; file_ = fopen(tmp_name.c_str(), "w+"); } void MemoryWatcher::CloseLogFile() { if (file_ != NULL) { fclose(file_); file_ = NULL; std::wstring tmp_name = ASCIIToWide(file_name_); tmp_name += L".tmp"; base::Move(base::FilePath(tmp_name), base::FilePath(ASCIIToWide(file_name_))); } } bool MemoryWatcher::LockedRecursionDetected() const { if (!active_thread_id_) return false; DWORD thread_id = GetCurrentThreadId(); // TODO(jar): Perchance we should use atomic access to member. return thread_id == active_thread_id_; } void MemoryWatcher::OnTrack(HANDLE heap, int32 id, int32 size) { // Don't track zeroes. It's a waste of time. if (size == 0) return; if (LockedRecursionDetected()) return; // AllocationStack overrides new/delete to not allocate // from the main heap. AllocationStack* stack = new AllocationStack(size); if (!stack->Valid()) return; // Recursion blocked generation of stack. { base::AutoLock lock(block_map_lock_); // Ideally, we'd like to verify that the block being added // here is not already in our list of tracked blocks. However, // the lookup in our hash table is expensive and slows us too // much. CallStackMap::iterator block_it = block_map_->find(id); if (block_it != block_map_->end()) { #if 0 // Don't do this until stack->ToString() uses ONLY our heap. active_thread_id_ = GetCurrentThreadId(); PrivateAllocatorString output; block_it->second->ToString(&output); // VLOG(1) << "First Stack size " << stack->size() << "was\n" << output; stack->ToString(&output); // VLOG(1) << "Second Stack size " << stack->size() << "was\n" << output; #endif // 0 // TODO(jar): We should delete one stack, and keep the other, perhaps // based on size. // For now, just delete the first, and keep the second? delete block_it->second; } // TODO(jar): Perchance we should use atomic access to member. active_thread_id_ = 0; // Note: Only do this AFTER exiting above scope! (*block_map_)[id] = stack; } mem_in_use.Add(size); mem_in_use_blocks.Increment(); mem_in_use_allocs.Increment(); } void MemoryWatcher::OnUntrack(HANDLE heap, int32 id, int32 size) { DCHECK_GE(size, 0); // Don't bother with these. if (size == 0) return; if (LockedRecursionDetected()) return; { base::AutoLock lock(block_map_lock_); active_thread_id_ = GetCurrentThreadId(); // First, find the block in our block_map. CallStackMap::iterator it = block_map_->find(id); if (it != block_map_->end()) { AllocationStack* stack = it->second; DCHECK(stack->size() == size); block_map_->erase(id); delete stack; } else { // Untracked item. This happens a fair amount, and it is // normal. A lot of time elapses during process startup // before the allocation routines are hooked. size = 0; // Ignore size in tallies. } // TODO(jar): Perchance we should use atomic access to member. active_thread_id_ = 0; } mem_in_use.Add(-size); mem_in_use_blocks.Decrement(); mem_in_use_frees.Increment(); } void MemoryWatcher::SetLogName(char* log_name) { if (!log_name) return; log_name_ = log_name; } // Help sort lists of stacks based on allocation cost. // Note: Sort based on allocation count is interesting too! static bool CompareCallStackIdItems(MemoryWatcher::StackTrack* left, MemoryWatcher::StackTrack* right) { return left->size > right->size; } void MemoryWatcher::DumpLeaks() { // We can only dump the leaks once. We'll cleanup the hooks here. if (!hooked_) return; Unhook(); base::AutoLock lock(block_map_lock_); active_thread_id_ = GetCurrentThreadId(); OpenLogFile(); // Aggregate contributions from each allocated block on per-stack basis. CallStackIdMap stack_map; for (CallStackMap::iterator block_it = block_map_->begin(); block_it != block_map_->end(); ++block_it) { AllocationStack* stack = block_it->second; int32 stack_hash = stack->hash(); int32 alloc_block_size = stack->size(); CallStackIdMap::iterator it = stack_map.find(stack_hash); if (it == stack_map.end()) { StackTrack tracker; tracker.count = 1; tracker.size = alloc_block_size; tracker.stack = stack; // Temporary pointer into block_map_. stack_map[stack_hash] = tracker; } else { it->second.count++; it->second.size += alloc_block_size; } } // Don't release lock yet, as block_map_ is still pointed into. // Put references to StrackTracks into array for sorting. std::vector > stack_tracks(stack_map.size()); CallStackIdMap::iterator it = stack_map.begin(); for (size_t i = 0; i < stack_tracks.size(); ++i) { stack_tracks[i] = &(it->second); ++it; } sort(stack_tracks.begin(), stack_tracks.end(), CompareCallStackIdItems); int32 total_bytes = 0; int32 total_blocks = 0; for (size_t i = 0; i < stack_tracks.size(); ++i) { StackTrack* stack_track = stack_tracks[i]; fwprintf(file_, L"%d bytes, %d allocs, #%d\n", stack_track->size, stack_track->count, i); total_bytes += stack_track->size; total_blocks += stack_track->count; CallStack* stack = stack_track->stack; PrivateAllocatorString output; stack->ToString(&output); fprintf(file_, "%s", output.c_str()); } fprintf(file_, "Total Leaks: %d\n", total_blocks); fprintf(file_, "Total Stacks: %d\n", stack_tracks.size()); fprintf(file_, "Total Bytes: %d\n", total_bytes); CloseLogFile(); }