054bdf7fc0
The class added unecessary level of indirection to the heap profiler implementation. I merged some of its implementation into HeapProfiler and deleted the rest. This refactoring is also a prerequisite for keeping HeapObjectsMap alive even when all snapshots are deleted. BUG=None LOG=N R=alph@chromium.org, mstarzinger@chromium.org Review URL: https://codereview.chromium.org/101393002 git-svn-id: http://v8.googlecode.com/svn/branches/bleeding_edge@18221 ce2b1a6d-e550-0410-aec6-3dcde31c8c00
216 lines
6.5 KiB
C++
216 lines
6.5 KiB
C++
// Copyright 2009-2010 the V8 project authors. All rights reserved.
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions are
|
|
// met:
|
|
//
|
|
// * Redistributions of source code must retain the above copyright
|
|
// notice, this list of conditions and the following disclaimer.
|
|
// * Redistributions in binary form must reproduce the above
|
|
// copyright notice, this list of conditions and the following
|
|
// disclaimer in the documentation and/or other materials provided
|
|
// with the distribution.
|
|
// * Neither the name of Google Inc. nor the names of its
|
|
// contributors may be used to endorse or promote products derived
|
|
// from this software without specific prior written permission.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
#include "v8.h"
|
|
|
|
#include "heap-profiler.h"
|
|
|
|
#include "allocation-tracker.h"
|
|
#include "heap-snapshot-generator-inl.h"
|
|
|
|
namespace v8 {
|
|
namespace internal {
|
|
|
|
HeapProfiler::HeapProfiler(Heap* heap)
|
|
: ids_(new HeapObjectsMap(heap)),
|
|
names_(new StringsStorage(heap)),
|
|
next_snapshot_uid_(1),
|
|
is_tracking_object_moves_(false) {
|
|
}
|
|
|
|
|
|
static void DeleteHeapSnapshot(HeapSnapshot** snapshot_ptr) {
|
|
delete *snapshot_ptr;
|
|
}
|
|
|
|
|
|
HeapProfiler::~HeapProfiler() {
|
|
snapshots_.Iterate(DeleteHeapSnapshot);
|
|
snapshots_.Clear();
|
|
}
|
|
|
|
|
|
void HeapProfiler::DeleteAllSnapshots() {
|
|
snapshots_.Iterate(DeleteHeapSnapshot);
|
|
snapshots_.Clear();
|
|
names_.Reset(new StringsStorage(heap()));
|
|
}
|
|
|
|
|
|
void HeapProfiler::RemoveSnapshot(HeapSnapshot* snapshot) {
|
|
snapshots_.RemoveElement(snapshot);
|
|
}
|
|
|
|
|
|
void HeapProfiler::DefineWrapperClass(
|
|
uint16_t class_id, v8::HeapProfiler::WrapperInfoCallback callback) {
|
|
ASSERT(class_id != v8::HeapProfiler::kPersistentHandleNoClassId);
|
|
if (wrapper_callbacks_.length() <= class_id) {
|
|
wrapper_callbacks_.AddBlock(
|
|
NULL, class_id - wrapper_callbacks_.length() + 1);
|
|
}
|
|
wrapper_callbacks_[class_id] = callback;
|
|
}
|
|
|
|
|
|
v8::RetainedObjectInfo* HeapProfiler::ExecuteWrapperClassCallback(
|
|
uint16_t class_id, Object** wrapper) {
|
|
if (wrapper_callbacks_.length() <= class_id) return NULL;
|
|
return wrapper_callbacks_[class_id](
|
|
class_id, Utils::ToLocal(Handle<Object>(wrapper)));
|
|
}
|
|
|
|
|
|
HeapSnapshot* HeapProfiler::TakeSnapshot(
|
|
const char* name,
|
|
v8::ActivityControl* control,
|
|
v8::HeapProfiler::ObjectNameResolver* resolver) {
|
|
HeapSnapshot* result = new HeapSnapshot(this, name, next_snapshot_uid_++);
|
|
{
|
|
HeapSnapshotGenerator generator(result, control, resolver, heap());
|
|
if (!generator.GenerateSnapshot()) {
|
|
delete result;
|
|
result = NULL;
|
|
} else {
|
|
snapshots_.Add(result);
|
|
}
|
|
}
|
|
ids_->RemoveDeadEntries();
|
|
is_tracking_object_moves_ = true;
|
|
return result;
|
|
}
|
|
|
|
|
|
HeapSnapshot* HeapProfiler::TakeSnapshot(
|
|
String* name,
|
|
v8::ActivityControl* control,
|
|
v8::HeapProfiler::ObjectNameResolver* resolver) {
|
|
return TakeSnapshot(names_->GetName(name), control, resolver);
|
|
}
|
|
|
|
|
|
void HeapProfiler::StartHeapObjectsTracking(bool track_allocations) {
|
|
ids_->UpdateHeapObjectsMap();
|
|
is_tracking_object_moves_ = true;
|
|
ASSERT(!is_tracking_allocations());
|
|
if (track_allocations) {
|
|
allocation_tracker_.Reset(new AllocationTracker(*ids_, *names_));
|
|
heap()->DisableInlineAllocation();
|
|
}
|
|
}
|
|
|
|
|
|
SnapshotObjectId HeapProfiler::PushHeapObjectsStats(OutputStream* stream) {
|
|
return ids_->PushHeapObjectsStats(stream);
|
|
}
|
|
|
|
|
|
void HeapProfiler::StopHeapObjectsTracking() {
|
|
ids_->StopHeapObjectsTracking();
|
|
if (is_tracking_allocations()) {
|
|
allocation_tracker_.Reset(NULL);
|
|
heap()->EnableInlineAllocation();
|
|
}
|
|
}
|
|
|
|
|
|
size_t HeapProfiler::GetMemorySizeUsedByProfiler() {
|
|
size_t size = sizeof(*this);
|
|
size += names_->GetUsedMemorySize();
|
|
size += ids_->GetUsedMemorySize();
|
|
size += GetMemoryUsedByList(snapshots_);
|
|
for (int i = 0; i < snapshots_.length(); ++i) {
|
|
size += snapshots_[i]->RawSnapshotSize();
|
|
}
|
|
return size;
|
|
}
|
|
|
|
|
|
int HeapProfiler::GetSnapshotsCount() {
|
|
return snapshots_.length();
|
|
}
|
|
|
|
|
|
HeapSnapshot* HeapProfiler::GetSnapshot(int index) {
|
|
return snapshots_.at(index);
|
|
}
|
|
|
|
|
|
SnapshotObjectId HeapProfiler::GetSnapshotObjectId(Handle<Object> obj) {
|
|
if (!obj->IsHeapObject())
|
|
return v8::HeapProfiler::kUnknownObjectId;
|
|
return ids_->FindEntry(HeapObject::cast(*obj)->address());
|
|
}
|
|
|
|
|
|
void HeapProfiler::ObjectMoveEvent(Address from, Address to, int size) {
|
|
ids_->MoveObject(from, to, size);
|
|
}
|
|
|
|
|
|
void HeapProfiler::AllocationEvent(Address addr, int size) {
|
|
DisallowHeapAllocation no_allocation;
|
|
if (!allocation_tracker_.is_empty()) {
|
|
allocation_tracker_->AllocationEvent(addr, size);
|
|
}
|
|
}
|
|
|
|
|
|
void HeapProfiler::UpdateObjectSizeEvent(Address addr, int size) {
|
|
ids_->UpdateObjectSize(addr, size);
|
|
}
|
|
|
|
|
|
void HeapProfiler::SetRetainedObjectInfo(UniqueId id,
|
|
RetainedObjectInfo* info) {
|
|
// TODO(yurus, marja): Don't route this information through GlobalHandles.
|
|
heap()->isolate()->global_handles()->SetRetainedObjectInfo(id, info);
|
|
}
|
|
|
|
|
|
Handle<HeapObject> HeapProfiler::FindHeapObjectById(SnapshotObjectId id) {
|
|
heap()->CollectAllGarbage(Heap::kMakeHeapIterableMask,
|
|
"HeapProfiler::FindHeapObjectById");
|
|
DisallowHeapAllocation no_allocation;
|
|
HeapObject* object = NULL;
|
|
HeapIterator iterator(heap(), HeapIterator::kFilterUnreachable);
|
|
// Make sure that object with the given id is still reachable.
|
|
for (HeapObject* obj = iterator.next();
|
|
obj != NULL;
|
|
obj = iterator.next()) {
|
|
if (ids_->FindEntry(obj->address()) == id) {
|
|
ASSERT(object == NULL);
|
|
object = obj;
|
|
// Can't break -- kFilterUnreachable requires full heap traversal.
|
|
}
|
|
}
|
|
return object != NULL ? Handle<HeapObject>(object) : Handle<HeapObject>();
|
|
}
|
|
|
|
|
|
} } // namespace v8::internal
|