mirror of
https://github.com/beefytech/Beef.git
synced 2025-06-22 09:38:01 +02:00
656 lines
No EOL
26 KiB
C++
656 lines
No EOL
26 KiB
C++
#if 0
|
|
|
|
/* Copyright (c) 2006, Google Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are
|
|
* met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following disclaimer
|
|
* in the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Google Inc. nor the names of its
|
|
* contributors may be used to endorse or promote products derived from
|
|
* this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* ---
|
|
* Author: Maxim Lifantsev
|
|
*/
|
|
|
|
//
|
|
// Background and key design points of MemoryRegionMap.
|
|
//
|
|
// MemoryRegionMap is a low-level module with quite atypical requirements that
|
|
// result in some degree of non-triviality of the implementation and design.
|
|
//
|
|
// MemoryRegionMap collects info about *all* memory regions created with
|
|
// mmap, munmap, mremap, sbrk.
|
|
// They key word above is 'all': all that are happening in a process
|
|
// during its lifetime frequently starting even before global object
|
|
// constructor execution.
|
|
//
|
|
// This is needed by the primary client of MemoryRegionMap:
|
|
// HeapLeakChecker uses the regions and the associated stack traces
|
|
// to figure out what part of the memory is the heap:
|
|
// if MemoryRegionMap were to miss some (early) regions, leak checking would
|
|
// stop working correctly.
|
|
//
|
|
// To accomplish the goal of functioning before/during global object
|
|
// constructor execution MemoryRegionMap is done as a singleton service
|
|
// that relies on own on-demand initialized static constructor-less data,
|
|
// and only relies on other low-level modules that can also function properly
|
|
// even before global object constructors run.
|
|
//
|
|
// Accomplishing the goal of collecting data about all mmap, munmap, mremap,
|
|
// sbrk occurrences is a more involved: conceptually to do this one needs to
|
|
// record some bits of data in particular about any mmap or sbrk call,
|
|
// but to do that one needs to allocate memory for that data at some point,
|
|
// but all memory allocations in the end themselves come from an mmap
|
|
// or sbrk call (that's how the address space of the process grows).
|
|
//
|
|
// Also note that we need to do all the above recording from
|
|
// within an mmap/sbrk hook which is sometimes/frequently is made by a memory
|
|
// allocator, including the allocator MemoryRegionMap itself must rely on.
|
|
// In the case of heap-checker usage this includes even the very first
|
|
// mmap/sbrk call happening in the program: heap-checker gets activated due to
|
|
// a link-time installed mmap/sbrk hook and it initializes MemoryRegionMap
|
|
// and asks it to record info about this very first call right from that
|
|
// very first hook invocation.
|
|
//
|
|
// MemoryRegionMap is doing its memory allocations via LowLevelAlloc:
|
|
// unlike more complex standard memory allocator, LowLevelAlloc cooperates with
|
|
// MemoryRegionMap by not holding any of its own locks while it calls mmap
|
|
// to get memory, thus we are able to call LowLevelAlloc from
|
|
// our mmap/sbrk hooks without causing a deadlock in it.
|
|
// For the same reason of deadlock prevention the locking in MemoryRegionMap
|
|
// itself is write-recursive which is an exception to Google's mutex usage.
|
|
//
|
|
// We still need to break the infinite cycle of mmap calling our hook,
|
|
// which asks LowLevelAlloc for memory to record this mmap,
|
|
// which (sometimes) causes mmap, which calls our hook, and so on.
|
|
// We do this as follows: on a recursive call of MemoryRegionMap's
|
|
// mmap/sbrk/mremap hook we record the data about the allocation in a
|
|
// static fixed-sized stack (saved_regions), when the recursion unwinds
|
|
// but before returning from the outer hook call we unwind this stack and
|
|
// move the data from saved_regions to its permanent place in the RegionSet,
|
|
// which can cause more allocations and mmap-s and recursion and unwinding,
|
|
// but the whole process ends eventually due to the fact that for the small
|
|
// allocations we are doing LowLevelAlloc reuses one mmap call and parcels out
|
|
// the memory it created to satisfy several of our allocation requests.
|
|
//
|
|
|
|
// ========================================================================= //
|
|
|
|
#include <config.h>
|
|
|
|
#ifdef HAVE_UNISTD_H
|
|
#include <unistd.h>
|
|
#endif
|
|
#ifdef HAVE_INTTYPES_H
|
|
#include <inttypes.h>
|
|
#endif
|
|
#ifdef HAVE_MMAP
|
|
#include <sys/mman.h>
|
|
#elif !defined(MAP_FAILED)
|
|
#define MAP_FAILED -1 // the only thing we need from mman.h
|
|
#endif
|
|
#ifdef HAVE_PTHREAD
|
|
#include <pthread.h> // for pthread_t, pthread_self()
|
|
#endif
|
|
#include <stddef.h>
|
|
|
|
#include <algorithm>
|
|
#include <set>
|
|
|
|
#include "memory_region_map.h"
|
|
|
|
#include "base/logging.h"
|
|
#include "base/low_level_alloc.h"
|
|
#include "malloc_hook-inl.h"
|
|
|
|
#include <gperftools/stacktrace.h>
|
|
#include <gperftools/malloc_hook.h>
|
|
|
|
// MREMAP_FIXED is a linux extension. How it's used in this file,
|
|
// setting it to 0 is equivalent to saying, "This feature isn't
|
|
// supported", which is right.
|
|
#ifndef MREMAP_FIXED
|
|
# define MREMAP_FIXED 0
|
|
#endif
|
|
|
|
using std::max;
|
|
|
|
// ========================================================================= //
|
|
|
|
int MemoryRegionMap::client_count_ = 0;
|
|
int MemoryRegionMap::max_stack_depth_ = 0;
|
|
MemoryRegionMap::RegionSet* MemoryRegionMap::regions_ = NULL;
|
|
LowLevelAlloc::Arena* MemoryRegionMap::arena_ = NULL;
|
|
SpinLock MemoryRegionMap::lock_(SpinLock::LINKER_INITIALIZED);
|
|
SpinLock MemoryRegionMap::owner_lock_( // ACQUIRED_AFTER(lock_)
|
|
SpinLock::LINKER_INITIALIZED);
|
|
int MemoryRegionMap::recursion_count_ = 0; // GUARDED_BY(owner_lock_)
|
|
pthread_t MemoryRegionMap::lock_owner_tid_; // GUARDED_BY(owner_lock_)
|
|
int64 MemoryRegionMap::map_size_ = 0;
|
|
int64 MemoryRegionMap::unmap_size_ = 0;
|
|
|
|
// ========================================================================= //
|
|
|
|
// Simple hook into execution of global object constructors,
|
|
// so that we do not call pthread_self() when it does not yet work.
|
|
static bool libpthread_initialized = false;
|
|
static bool initializer = (libpthread_initialized = true, true);
|
|
|
|
static inline bool current_thread_is(pthread_t should_be) {
|
|
// Before main() runs, there's only one thread, so we're always that thread
|
|
if (!libpthread_initialized) return true;
|
|
// this starts working only sometime well into global constructor execution:
|
|
return pthread_equal(pthread_self(), should_be);
|
|
}
|
|
|
|
// ========================================================================= //
|
|
|
|
// Constructor-less place-holder to store a RegionSet in.
|
|
union MemoryRegionMap::RegionSetRep {
|
|
char rep[sizeof(RegionSet)];
|
|
void* align_it; // do not need a better alignment for 'rep' than this
|
|
RegionSet* region_set() { return reinterpret_cast<RegionSet*>(rep); }
|
|
};
|
|
|
|
// The bytes where MemoryRegionMap::regions_ will point to.
|
|
// We use RegionSetRep with noop c-tor so that global construction
|
|
// does not interfere.
|
|
static MemoryRegionMap::RegionSetRep regions_rep;
|
|
|
|
// ========================================================================= //
|
|
|
|
// Has InsertRegionLocked been called recursively
|
|
// (or rather should we *not* use regions_ to record a hooked mmap).
|
|
static bool recursive_insert = false;
|
|
|
|
void MemoryRegionMap::Init(int max_stack_depth) {
|
|
RAW_VLOG(10, "MemoryRegionMap Init");
|
|
RAW_CHECK(max_stack_depth >= 0, "");
|
|
// Make sure we don't overflow the memory in region stacks:
|
|
RAW_CHECK(max_stack_depth <= kMaxStackDepth,
|
|
"need to increase kMaxStackDepth?");
|
|
Lock();
|
|
client_count_ += 1;
|
|
max_stack_depth_ = max(max_stack_depth_, max_stack_depth);
|
|
if (client_count_ > 1) {
|
|
// not first client: already did initialization-proper
|
|
Unlock();
|
|
RAW_VLOG(10, "MemoryRegionMap Init increment done");
|
|
return;
|
|
}
|
|
// Set our hooks and make sure they were installed:
|
|
RAW_CHECK(MallocHook::AddMmapHook(&MmapHook), "");
|
|
RAW_CHECK(MallocHook::AddMremapHook(&MremapHook), "");
|
|
RAW_CHECK(MallocHook::AddSbrkHook(&SbrkHook), "");
|
|
RAW_CHECK(MallocHook::AddMunmapHook(&MunmapHook), "");
|
|
// We need to set recursive_insert since the NewArena call itself
|
|
// will already do some allocations with mmap which our hooks will catch
|
|
// recursive_insert allows us to buffer info about these mmap calls.
|
|
// Note that Init() can be (and is) sometimes called
|
|
// already from within an mmap/sbrk hook.
|
|
recursive_insert = true;
|
|
arena_ = LowLevelAlloc::NewArena(0, LowLevelAlloc::DefaultArena());
|
|
recursive_insert = false;
|
|
HandleSavedRegionsLocked(&InsertRegionLocked); // flush the buffered ones
|
|
// Can't instead use HandleSavedRegionsLocked(&DoInsertRegionLocked) before
|
|
// recursive_insert = false; as InsertRegionLocked will also construct
|
|
// regions_ on demand for us.
|
|
Unlock();
|
|
RAW_VLOG(10, "MemoryRegionMap Init done");
|
|
}
|
|
|
|
bool MemoryRegionMap::Shutdown() {
|
|
RAW_VLOG(10, "MemoryRegionMap Shutdown");
|
|
Lock();
|
|
RAW_CHECK(client_count_ > 0, "");
|
|
client_count_ -= 1;
|
|
if (client_count_ != 0) { // not last client; need not really shutdown
|
|
Unlock();
|
|
RAW_VLOG(10, "MemoryRegionMap Shutdown decrement done");
|
|
return true;
|
|
}
|
|
RAW_CHECK(MallocHook::RemoveMmapHook(&MmapHook), "");
|
|
RAW_CHECK(MallocHook::RemoveMremapHook(&MremapHook), "");
|
|
RAW_CHECK(MallocHook::RemoveSbrkHook(&SbrkHook), "");
|
|
RAW_CHECK(MallocHook::RemoveMunmapHook(&MunmapHook), "");
|
|
if (regions_) regions_->~RegionSet();
|
|
regions_ = NULL;
|
|
bool deleted_arena = LowLevelAlloc::DeleteArena(arena_);
|
|
if (deleted_arena) {
|
|
arena_ = 0;
|
|
} else {
|
|
RAW_LOG(WARNING, "Can't delete LowLevelAlloc arena: it's being used");
|
|
}
|
|
Unlock();
|
|
RAW_VLOG(10, "MemoryRegionMap Shutdown done");
|
|
return deleted_arena;
|
|
}
|
|
|
|
// Invariants (once libpthread_initialized is true):
|
|
// * While lock_ is not held, recursion_count_ is 0 (and
|
|
// lock_owner_tid_ is the previous owner, but we don't rely on
|
|
// that).
|
|
// * recursion_count_ and lock_owner_tid_ are only written while
|
|
// both lock_ and owner_lock_ are held. They may be read under
|
|
// just owner_lock_.
|
|
// * At entry and exit of Lock() and Unlock(), the current thread
|
|
// owns lock_ iff pthread_equal(lock_owner_tid_, pthread_self())
|
|
// && recursion_count_ > 0.
|
|
void MemoryRegionMap::Lock() {
|
|
{
|
|
SpinLockHolder l(&owner_lock_);
|
|
if (recursion_count_ > 0 && current_thread_is(lock_owner_tid_)) {
|
|
RAW_CHECK(lock_.IsHeld(), "Invariants violated");
|
|
recursion_count_++;
|
|
RAW_CHECK(recursion_count_ <= 5,
|
|
"recursive lock nesting unexpectedly deep");
|
|
return;
|
|
}
|
|
}
|
|
lock_.Lock();
|
|
{
|
|
SpinLockHolder l(&owner_lock_);
|
|
RAW_CHECK(recursion_count_ == 0,
|
|
"Last Unlock didn't reset recursion_count_");
|
|
if (libpthread_initialized)
|
|
lock_owner_tid_ = pthread_self();
|
|
recursion_count_ = 1;
|
|
}
|
|
}
|
|
|
|
void MemoryRegionMap::Unlock() {
|
|
SpinLockHolder l(&owner_lock_);
|
|
RAW_CHECK(recursion_count_ > 0, "unlock when not held");
|
|
RAW_CHECK(lock_.IsHeld(),
|
|
"unlock when not held, and recursion_count_ is wrong");
|
|
RAW_CHECK(current_thread_is(lock_owner_tid_), "unlock by non-holder");
|
|
recursion_count_--;
|
|
if (recursion_count_ == 0) {
|
|
lock_.Unlock();
|
|
}
|
|
}
|
|
|
|
bool MemoryRegionMap::LockIsHeld() {
|
|
SpinLockHolder l(&owner_lock_);
|
|
return lock_.IsHeld() && current_thread_is(lock_owner_tid_);
|
|
}
|
|
|
|
const MemoryRegionMap::Region*
|
|
MemoryRegionMap::DoFindRegionLocked(uintptr_t addr) {
|
|
RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
|
|
if (regions_ != NULL) {
|
|
Region sample;
|
|
sample.SetRegionSetKey(addr);
|
|
RegionSet::iterator region = regions_->lower_bound(sample);
|
|
if (region != regions_->end()) {
|
|
RAW_CHECK(addr <= region->end_addr, "");
|
|
if (region->start_addr <= addr && addr < region->end_addr) {
|
|
return &(*region);
|
|
}
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
bool MemoryRegionMap::FindRegion(uintptr_t addr, Region* result) {
|
|
Lock();
|
|
const Region* region = DoFindRegionLocked(addr);
|
|
if (region != NULL) *result = *region; // create it as an independent copy
|
|
Unlock();
|
|
return region != NULL;
|
|
}
|
|
|
|
bool MemoryRegionMap::FindAndMarkStackRegion(uintptr_t stack_top,
|
|
Region* result) {
|
|
Lock();
|
|
const Region* region = DoFindRegionLocked(stack_top);
|
|
if (region != NULL) {
|
|
RAW_VLOG(10, "Stack at %p is inside region %p..%p",
|
|
reinterpret_cast<void*>(stack_top),
|
|
reinterpret_cast<void*>(region->start_addr),
|
|
reinterpret_cast<void*>(region->end_addr));
|
|
const_cast<Region*>(region)->set_is_stack(); // now we know
|
|
// cast is safe (set_is_stack does not change the set ordering key)
|
|
*result = *region; // create *result as an independent copy
|
|
}
|
|
Unlock();
|
|
return region != NULL;
|
|
}
|
|
|
|
MemoryRegionMap::RegionIterator MemoryRegionMap::BeginRegionLocked() {
|
|
RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
|
|
RAW_CHECK(regions_ != NULL, "");
|
|
return regions_->begin();
|
|
}
|
|
|
|
MemoryRegionMap::RegionIterator MemoryRegionMap::EndRegionLocked() {
|
|
RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
|
|
RAW_CHECK(regions_ != NULL, "");
|
|
return regions_->end();
|
|
}
|
|
|
|
inline void MemoryRegionMap::DoInsertRegionLocked(const Region& region) {
|
|
RAW_VLOG(12, "Inserting region %p..%p from %p",
|
|
reinterpret_cast<void*>(region.start_addr),
|
|
reinterpret_cast<void*>(region.end_addr),
|
|
reinterpret_cast<void*>(region.caller()));
|
|
RegionSet::const_iterator i = regions_->lower_bound(region);
|
|
if (i != regions_->end() && i->start_addr <= region.start_addr) {
|
|
RAW_DCHECK(region.end_addr <= i->end_addr, ""); // lower_bound ensures this
|
|
return; // 'region' is a subset of an already recorded region; do nothing
|
|
// We can be stricter and allow this only when *i has been created via
|
|
// an mmap with MAP_NORESERVE flag set.
|
|
}
|
|
if (DEBUG_MODE) {
|
|
RAW_CHECK(i == regions_->end() || !region.Overlaps(*i),
|
|
"Wow, overlapping memory regions");
|
|
Region sample;
|
|
sample.SetRegionSetKey(region.start_addr);
|
|
i = regions_->lower_bound(sample);
|
|
RAW_CHECK(i == regions_->end() || !region.Overlaps(*i),
|
|
"Wow, overlapping memory regions");
|
|
}
|
|
region.AssertIsConsistent(); // just making sure
|
|
// This inserts and allocates permanent storage for region
|
|
// and its call stack data: it's safe to do it now:
|
|
regions_->insert(region);
|
|
RAW_VLOG(12, "Inserted region %p..%p :",
|
|
reinterpret_cast<void*>(region.start_addr),
|
|
reinterpret_cast<void*>(region.end_addr));
|
|
if (VLOG_IS_ON(12)) LogAllLocked();
|
|
}
|
|
|
|
// These variables are local to MemoryRegionMap::InsertRegionLocked()
|
|
// and MemoryRegionMap::HandleSavedRegionsLocked()
|
|
// and are file-level to ensure that they are initialized at load time.
|
|
|
|
// Number of unprocessed region inserts.
|
|
static int saved_regions_count = 0;
|
|
|
|
// Unprocessed inserts (must be big enough to hold all allocations that can
|
|
// be caused by a InsertRegionLocked call).
|
|
// Region has no constructor, so that c-tor execution does not interfere
|
|
// with the any-time use of the static memory behind saved_regions.
|
|
static MemoryRegionMap::Region saved_regions[20];
|
|
|
|
inline void MemoryRegionMap::HandleSavedRegionsLocked(
|
|
void (*insert_func)(const Region& region)) {
|
|
while (saved_regions_count > 0) {
|
|
// Making a local-var copy of the region argument to insert_func
|
|
// including its stack (w/o doing any memory allocations) is important:
|
|
// in many cases the memory in saved_regions
|
|
// will get written-to during the (*insert_func)(r) call below.
|
|
Region r = saved_regions[--saved_regions_count];
|
|
(*insert_func)(r);
|
|
}
|
|
}
|
|
|
|
inline void MemoryRegionMap::InsertRegionLocked(const Region& region) {
|
|
RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
|
|
// We can be called recursively, because RegionSet constructor
|
|
// and DoInsertRegionLocked() (called below) can call the allocator.
|
|
// recursive_insert tells us if that's the case. When this happens,
|
|
// region insertion information is recorded in saved_regions[],
|
|
// and taken into account when the recursion unwinds.
|
|
// Do the insert:
|
|
if (recursive_insert) { // recursion: save in saved_regions
|
|
RAW_VLOG(12, "Saving recursive insert of region %p..%p from %p",
|
|
reinterpret_cast<void*>(region.start_addr),
|
|
reinterpret_cast<void*>(region.end_addr),
|
|
reinterpret_cast<void*>(region.caller()));
|
|
RAW_CHECK(saved_regions_count < arraysize(saved_regions), "");
|
|
// Copy 'region' to saved_regions[saved_regions_count]
|
|
// together with the contents of its call_stack,
|
|
// then increment saved_regions_count.
|
|
saved_regions[saved_regions_count++] = region;
|
|
} else { // not a recusrive call
|
|
if (regions_ == NULL) { // init regions_
|
|
RAW_VLOG(12, "Initializing region set");
|
|
regions_ = regions_rep.region_set();
|
|
recursive_insert = true;
|
|
new(regions_) RegionSet();
|
|
HandleSavedRegionsLocked(&DoInsertRegionLocked);
|
|
recursive_insert = false;
|
|
}
|
|
recursive_insert = true;
|
|
// Do the actual insertion work to put new regions into regions_:
|
|
DoInsertRegionLocked(region);
|
|
HandleSavedRegionsLocked(&DoInsertRegionLocked);
|
|
recursive_insert = false;
|
|
}
|
|
}
|
|
|
|
// We strip out different number of stack frames in debug mode
|
|
// because less inlining happens in that case
|
|
#ifdef NDEBUG
|
|
static const int kStripFrames = 1;
|
|
#else
|
|
static const int kStripFrames = 3;
|
|
#endif
|
|
|
|
void MemoryRegionMap::RecordRegionAddition(const void* start, size_t size) {
|
|
// Record start/end info about this memory acquisition call in a new region:
|
|
Region region;
|
|
region.Create(start, size);
|
|
// First get the call stack info into the local varible 'region':
|
|
const int depth =
|
|
max_stack_depth_ > 0
|
|
? MallocHook::GetCallerStackTrace(const_cast<void**>(region.call_stack),
|
|
max_stack_depth_, kStripFrames + 1)
|
|
: 0;
|
|
region.set_call_stack_depth(depth); // record stack info fully
|
|
RAW_VLOG(10, "New global region %p..%p from %p",
|
|
reinterpret_cast<void*>(region.start_addr),
|
|
reinterpret_cast<void*>(region.end_addr),
|
|
reinterpret_cast<void*>(region.caller()));
|
|
// Note: none of the above allocates memory.
|
|
Lock(); // recursively lock
|
|
map_size_ += size;
|
|
InsertRegionLocked(region);
|
|
// This will (eventually) allocate storage for and copy over the stack data
|
|
// from region.call_stack_data_ that is pointed by region.call_stack().
|
|
Unlock();
|
|
}
|
|
|
|
void MemoryRegionMap::RecordRegionRemoval(const void* start, size_t size) {
|
|
Lock();
|
|
if (recursive_insert) {
|
|
// First remove the removed region from saved_regions, if it's
|
|
// there, to prevent overrunning saved_regions in recursive
|
|
// map/unmap call sequences, and also from later inserting regions
|
|
// which have already been unmapped.
|
|
uintptr_t start_addr = reinterpret_cast<uintptr_t>(start);
|
|
uintptr_t end_addr = start_addr + size;
|
|
int put_pos = 0;
|
|
int old_count = saved_regions_count;
|
|
for (int i = 0; i < old_count; ++i, ++put_pos) {
|
|
Region& r = saved_regions[i];
|
|
if (r.start_addr == start_addr && r.end_addr == end_addr) {
|
|
// An exact match, so it's safe to remove.
|
|
--saved_regions_count;
|
|
--put_pos;
|
|
RAW_VLOG(10, ("Insta-Removing saved region %p..%p; "
|
|
"now have %d saved regions"),
|
|
reinterpret_cast<void*>(start_addr),
|
|
reinterpret_cast<void*>(end_addr),
|
|
saved_regions_count);
|
|
} else {
|
|
if (put_pos < i) {
|
|
saved_regions[put_pos] = saved_regions[i];
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if (regions_ == NULL) { // We must have just unset the hooks,
|
|
// but this thread was already inside the hook.
|
|
Unlock();
|
|
return;
|
|
}
|
|
if (!recursive_insert) {
|
|
HandleSavedRegionsLocked(&InsertRegionLocked);
|
|
}
|
|
// first handle adding saved regions if any
|
|
uintptr_t start_addr = reinterpret_cast<uintptr_t>(start);
|
|
uintptr_t end_addr = start_addr + size;
|
|
// subtract start_addr, end_addr from all the regions
|
|
RAW_VLOG(10, "Removing global region %p..%p; have %" PRIuS " regions",
|
|
reinterpret_cast<void*>(start_addr),
|
|
reinterpret_cast<void*>(end_addr),
|
|
regions_->size());
|
|
Region sample;
|
|
sample.SetRegionSetKey(start_addr);
|
|
// Only iterate over the regions that might overlap start_addr..end_addr:
|
|
for (RegionSet::iterator region = regions_->lower_bound(sample);
|
|
region != regions_->end() && region->start_addr < end_addr;
|
|
/*noop*/) {
|
|
RAW_VLOG(13, "Looking at region %p..%p",
|
|
reinterpret_cast<void*>(region->start_addr),
|
|
reinterpret_cast<void*>(region->end_addr));
|
|
if (start_addr <= region->start_addr &&
|
|
region->end_addr <= end_addr) { // full deletion
|
|
RAW_VLOG(12, "Deleting region %p..%p",
|
|
reinterpret_cast<void*>(region->start_addr),
|
|
reinterpret_cast<void*>(region->end_addr));
|
|
RegionSet::iterator d = region;
|
|
++region;
|
|
regions_->erase(d);
|
|
continue;
|
|
} else if (region->start_addr < start_addr &&
|
|
end_addr < region->end_addr) { // cutting-out split
|
|
RAW_VLOG(12, "Splitting region %p..%p in two",
|
|
reinterpret_cast<void*>(region->start_addr),
|
|
reinterpret_cast<void*>(region->end_addr));
|
|
// Make another region for the start portion:
|
|
// The new region has to be the start portion because we can't
|
|
// just modify region->end_addr as it's the sorting key.
|
|
Region r = *region;
|
|
r.set_end_addr(start_addr);
|
|
InsertRegionLocked(r);
|
|
// cut *region from start:
|
|
const_cast<Region&>(*region).set_start_addr(end_addr);
|
|
} else if (end_addr > region->start_addr &&
|
|
start_addr <= region->start_addr) { // cut from start
|
|
RAW_VLOG(12, "Start-chopping region %p..%p",
|
|
reinterpret_cast<void*>(region->start_addr),
|
|
reinterpret_cast<void*>(region->end_addr));
|
|
const_cast<Region&>(*region).set_start_addr(end_addr);
|
|
} else if (start_addr > region->start_addr &&
|
|
start_addr < region->end_addr) { // cut from end
|
|
RAW_VLOG(12, "End-chopping region %p..%p",
|
|
reinterpret_cast<void*>(region->start_addr),
|
|
reinterpret_cast<void*>(region->end_addr));
|
|
// Can't just modify region->end_addr (it's the sorting key):
|
|
Region r = *region;
|
|
r.set_end_addr(start_addr);
|
|
RegionSet::iterator d = region;
|
|
++region;
|
|
// It's safe to erase before inserting since r is independent of *d:
|
|
// r contains an own copy of the call stack:
|
|
regions_->erase(d);
|
|
InsertRegionLocked(r);
|
|
continue;
|
|
}
|
|
++region;
|
|
}
|
|
RAW_VLOG(12, "Removed region %p..%p; have %" PRIuS " regions",
|
|
reinterpret_cast<void*>(start_addr),
|
|
reinterpret_cast<void*>(end_addr),
|
|
regions_->size());
|
|
if (VLOG_IS_ON(12)) LogAllLocked();
|
|
unmap_size_ += size;
|
|
Unlock();
|
|
}
|
|
|
|
void MemoryRegionMap::MmapHook(const void* result,
|
|
const void* start, size_t size,
|
|
int prot, int flags,
|
|
int fd, off_t offset) {
|
|
// TODO(maxim): replace all 0x%" PRIxS " by %p when RAW_VLOG uses a safe
|
|
// snprintf reimplementation that does not malloc to pretty-print NULL
|
|
RAW_VLOG(10, "MMap = 0x%" PRIxPTR " of %" PRIuS " at %" PRIu64 " "
|
|
"prot %d flags %d fd %d offs %" PRId64,
|
|
reinterpret_cast<uintptr_t>(result), size,
|
|
reinterpret_cast<uint64>(start), prot, flags, fd,
|
|
static_cast<int64>(offset));
|
|
if (result != reinterpret_cast<void*>(MAP_FAILED) && size != 0) {
|
|
RecordRegionAddition(result, size);
|
|
}
|
|
}
|
|
|
|
void MemoryRegionMap::MunmapHook(const void* ptr, size_t size) {
|
|
RAW_VLOG(10, "MUnmap of %p %" PRIuS "", ptr, size);
|
|
if (size != 0) {
|
|
RecordRegionRemoval(ptr, size);
|
|
}
|
|
}
|
|
|
|
void MemoryRegionMap::MremapHook(const void* result,
|
|
const void* old_addr, size_t old_size,
|
|
size_t new_size, int flags,
|
|
const void* new_addr) {
|
|
RAW_VLOG(10, "MRemap = 0x%" PRIxPTR " of 0x%" PRIxPTR " %" PRIuS " "
|
|
"to %" PRIuS " flags %d new_addr=0x%" PRIxPTR,
|
|
(uintptr_t)result, (uintptr_t)old_addr,
|
|
old_size, new_size, flags,
|
|
flags & MREMAP_FIXED ? (uintptr_t)new_addr : 0);
|
|
if (result != reinterpret_cast<void*>(-1)) {
|
|
RecordRegionRemoval(old_addr, old_size);
|
|
RecordRegionAddition(result, new_size);
|
|
}
|
|
}
|
|
|
|
extern "C" void* __sbrk(ptrdiff_t increment); // defined in libc
|
|
|
|
void MemoryRegionMap::SbrkHook(const void* result, ptrdiff_t increment) {
|
|
RAW_VLOG(10, "Sbrk = 0x%" PRIxPTR " of %" PRIdS "", (uintptr_t)result, increment);
|
|
if (result != reinterpret_cast<void*>(-1)) {
|
|
if (increment > 0) {
|
|
void* new_end = sbrk(0);
|
|
RecordRegionAddition(result, reinterpret_cast<uintptr_t>(new_end) -
|
|
reinterpret_cast<uintptr_t>(result));
|
|
} else if (increment < 0) {
|
|
void* new_end = sbrk(0);
|
|
RecordRegionRemoval(new_end, reinterpret_cast<uintptr_t>(result) -
|
|
reinterpret_cast<uintptr_t>(new_end));
|
|
}
|
|
}
|
|
}
|
|
|
|
void MemoryRegionMap::LogAllLocked() {
|
|
RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
|
|
RAW_LOG(INFO, "List of regions:");
|
|
uintptr_t previous = 0;
|
|
for (RegionSet::const_iterator r = regions_->begin();
|
|
r != regions_->end(); ++r) {
|
|
RAW_LOG(INFO, "Memory region 0x%" PRIxPTR "..0x%" PRIxPTR " "
|
|
"from 0x%" PRIxPTR " stack=%d",
|
|
r->start_addr, r->end_addr, r->caller(), r->is_stack);
|
|
RAW_CHECK(previous < r->end_addr, "wow, we messed up the set order");
|
|
// this must be caused by uncontrolled recursive operations on regions_
|
|
previous = r->end_addr;
|
|
}
|
|
RAW_LOG(INFO, "End of regions list");
|
|
}
|
|
|
|
#endif |