2010-03-19 09:46:53 +00:00
|
|
|
// Copyright 2010 the V8 project authors. All rights reserved.
|
|
|
|
// Redistribution and use in source and binary forms, with or without
|
|
|
|
// modification, are permitted provided that the following conditions are
|
|
|
|
// met:
|
|
|
|
//
|
|
|
|
// * Redistributions of source code must retain the above copyright
|
|
|
|
// notice, this list of conditions and the following disclaimer.
|
|
|
|
// * Redistributions in binary form must reproduce the above
|
|
|
|
// copyright notice, this list of conditions and the following
|
|
|
|
// disclaimer in the documentation and/or other materials provided
|
|
|
|
// with the distribution.
|
|
|
|
// * Neither the name of Google Inc. nor the names of its
|
|
|
|
// contributors may be used to endorse or promote products derived
|
|
|
|
// from this software without specific prior written permission.
|
|
|
|
//
|
|
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
|
2010-03-30 11:38:39 +00:00
|
|
|
#ifdef ENABLE_CPP_PROFILES_PROCESSOR
|
|
|
|
|
2010-03-19 09:46:53 +00:00
|
|
|
#include "v8.h"
|
|
|
|
|
|
|
|
#include "cpu-profiler-inl.h"
|
|
|
|
|
|
|
|
namespace v8 {
|
|
|
|
namespace internal {
|
|
|
|
|
|
|
|
static const int kEventsBufferSize = 256*KB;
|
|
|
|
static const int kTickSamplesBufferChunkSize = 64*KB;
|
|
|
|
static const int kTickSamplesBufferChunksCount = 16;
|
|
|
|
|
|
|
|
|
|
|
|
ProfilerEventsProcessor::ProfilerEventsProcessor(ProfileGenerator* generator)
|
|
|
|
: generator_(generator),
|
|
|
|
running_(false),
|
|
|
|
events_buffer_(kEventsBufferSize),
|
|
|
|
ticks_buffer_(sizeof(TickSampleEventRecord),
|
|
|
|
kTickSamplesBufferChunkSize,
|
|
|
|
kTickSamplesBufferChunksCount),
|
|
|
|
enqueue_order_(0) { }
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::CodeCreateEvent(Logger::LogEventsAndTags tag,
|
|
|
|
String* name,
|
|
|
|
String* resource_name,
|
|
|
|
int line_number,
|
|
|
|
Address start,
|
|
|
|
unsigned size) {
|
|
|
|
CodeEventsContainer evt_rec;
|
|
|
|
CodeCreateEventRecord* rec = &evt_rec.CodeCreateEventRecord_;
|
|
|
|
rec->type = CodeEventRecord::CODE_CREATION;
|
|
|
|
rec->order = ++enqueue_order_;
|
|
|
|
rec->start = start;
|
|
|
|
rec->entry = generator_->NewCodeEntry(tag, name, resource_name, line_number);
|
|
|
|
rec->size = size;
|
|
|
|
events_buffer_.Enqueue(evt_rec);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::CodeCreateEvent(Logger::LogEventsAndTags tag,
|
|
|
|
const char* name,
|
|
|
|
Address start,
|
|
|
|
unsigned size) {
|
|
|
|
CodeEventsContainer evt_rec;
|
|
|
|
CodeCreateEventRecord* rec = &evt_rec.CodeCreateEventRecord_;
|
|
|
|
rec->type = CodeEventRecord::CODE_CREATION;
|
|
|
|
rec->order = ++enqueue_order_;
|
|
|
|
rec->start = start;
|
|
|
|
rec->entry = generator_->NewCodeEntry(tag, name);
|
|
|
|
rec->size = size;
|
|
|
|
events_buffer_.Enqueue(evt_rec);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::CodeCreateEvent(Logger::LogEventsAndTags tag,
|
|
|
|
int args_count,
|
|
|
|
Address start,
|
|
|
|
unsigned size) {
|
|
|
|
CodeEventsContainer evt_rec;
|
|
|
|
CodeCreateEventRecord* rec = &evt_rec.CodeCreateEventRecord_;
|
|
|
|
rec->type = CodeEventRecord::CODE_CREATION;
|
|
|
|
rec->order = ++enqueue_order_;
|
|
|
|
rec->start = start;
|
|
|
|
rec->entry = generator_->NewCodeEntry(tag, args_count);
|
|
|
|
rec->size = size;
|
|
|
|
events_buffer_.Enqueue(evt_rec);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::CodeMoveEvent(Address from, Address to) {
|
|
|
|
CodeEventsContainer evt_rec;
|
|
|
|
CodeMoveEventRecord* rec = &evt_rec.CodeMoveEventRecord_;
|
|
|
|
rec->type = CodeEventRecord::CODE_MOVE;
|
|
|
|
rec->order = ++enqueue_order_;
|
|
|
|
rec->from = from;
|
|
|
|
rec->to = to;
|
|
|
|
events_buffer_.Enqueue(evt_rec);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::CodeDeleteEvent(Address from) {
|
|
|
|
CodeEventsContainer evt_rec;
|
|
|
|
CodeDeleteEventRecord* rec = &evt_rec.CodeDeleteEventRecord_;
|
|
|
|
rec->type = CodeEventRecord::CODE_DELETE;
|
|
|
|
rec->order = ++enqueue_order_;
|
|
|
|
rec->start = from;
|
|
|
|
events_buffer_.Enqueue(evt_rec);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::FunctionCreateEvent(Address alias,
|
|
|
|
Address start) {
|
|
|
|
CodeEventsContainer evt_rec;
|
|
|
|
CodeAliasEventRecord* rec = &evt_rec.CodeAliasEventRecord_;
|
|
|
|
rec->type = CodeEventRecord::CODE_ALIAS;
|
|
|
|
rec->order = ++enqueue_order_;
|
|
|
|
rec->alias = alias;
|
|
|
|
rec->start = start;
|
|
|
|
events_buffer_.Enqueue(evt_rec);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::FunctionMoveEvent(Address from, Address to) {
|
|
|
|
CodeMoveEvent(from, to);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::FunctionDeleteEvent(Address from) {
|
|
|
|
CodeDeleteEvent(from);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
bool ProfilerEventsProcessor::ProcessCodeEvent(unsigned* dequeue_order) {
|
|
|
|
if (!events_buffer_.IsEmpty()) {
|
|
|
|
CodeEventsContainer record;
|
|
|
|
events_buffer_.Dequeue(&record);
|
|
|
|
switch (record.generic.type) {
|
|
|
|
#define PROFILER_TYPE_CASE(type, clss) \
|
|
|
|
case CodeEventRecord::type: \
|
|
|
|
record.clss##_.UpdateCodeMap(generator_->code_map()); \
|
|
|
|
break;
|
|
|
|
|
|
|
|
CODE_EVENTS_TYPE_LIST(PROFILER_TYPE_CASE)
|
|
|
|
|
|
|
|
#undef PROFILER_TYPE_CASE
|
|
|
|
default: return true; // Skip record.
|
|
|
|
}
|
|
|
|
*dequeue_order = record.generic.order;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
bool ProfilerEventsProcessor::ProcessTicks(unsigned dequeue_order) {
|
|
|
|
while (true) {
|
|
|
|
const TickSampleEventRecord* rec =
|
2010-03-30 11:38:39 +00:00
|
|
|
TickSampleEventRecord::cast(ticks_buffer_.StartDequeue());
|
2010-03-19 09:46:53 +00:00
|
|
|
if (rec == NULL) return false;
|
|
|
|
if (rec->order == dequeue_order) {
|
|
|
|
generator_->RecordTickSample(rec->sample);
|
|
|
|
ticks_buffer_.FinishDequeue();
|
|
|
|
} else {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ProfilerEventsProcessor::Run() {
|
|
|
|
unsigned dequeue_order = 0;
|
|
|
|
running_ = true;
|
|
|
|
|
|
|
|
while (running_) {
|
|
|
|
// Process ticks until we have any.
|
|
|
|
if (ProcessTicks(dequeue_order)) {
|
|
|
|
// All ticks of the current dequeue_order are processed,
|
|
|
|
// proceed to the next code event.
|
|
|
|
ProcessCodeEvent(&dequeue_order);
|
|
|
|
}
|
|
|
|
YieldCPU();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Process remaining tick events.
|
|
|
|
ticks_buffer_.FlushResidualRecords();
|
|
|
|
// Perform processing until we have tick events, skip remaining code events.
|
|
|
|
while (ProcessTicks(dequeue_order) && ProcessCodeEvent(&dequeue_order)) { }
|
|
|
|
}
|
|
|
|
|
|
|
|
} } // namespace v8::internal
|
2010-03-30 11:38:39 +00:00
|
|
|
|
|
|
|
#endif // ENABLE_CPP_PROFILES_PROCESSOR
|