Relocate suspended generator activations when enabling debug mode

R=yangguo@chromium.org
BUG=v8:3289
LOG=N

Review URL: https://codereview.chromium.org/260423002

git-svn-id: http://v8.googlecode.com/svn/branches/bleeding_edge@21141 ce2b1a6d-e550-0410-aec6-3dcde31c8c00
This commit is contained in:
wingo@igalia.com 2014-05-05 12:57:14 +00:00
parent b66afe30fe
commit 9a9943b564
5 changed files with 240 additions and 72 deletions

View File

@ -1881,6 +1881,59 @@ static void CollectActiveFunctionsFromThread(
}
// Figure out how many bytes of "pc_offset" correspond to actual code by
// subtracting off the bytes that correspond to constant/veneer pools. See
// Assembler::CheckConstPool() and Assembler::CheckVeneerPool(). Note that this
// is only useful for architectures using constant pools or veneer pools.
static int ComputeCodeOffsetFromPcOffset(Code *code, int pc_offset) {
ASSERT_EQ(code->kind(), Code::FUNCTION);
ASSERT(!code->has_debug_break_slots());
ASSERT_LE(0, pc_offset);
ASSERT_LT(pc_offset, code->instruction_end() - code->instruction_start());
int mask = RelocInfo::ModeMask(RelocInfo::CONST_POOL) |
RelocInfo::ModeMask(RelocInfo::VENEER_POOL);
byte *pc = code->instruction_start() + pc_offset;
int code_offset = pc_offset;
for (RelocIterator it(code, mask); !it.done(); it.next()) {
RelocInfo* info = it.rinfo();
if (info->pc() >= pc) break;
ASSERT(RelocInfo::IsConstPool(info->rmode()));
code_offset -= static_cast<int>(info->data());
ASSERT_LE(0, code_offset);
}
return code_offset;
}
// The inverse of ComputeCodeOffsetFromPcOffset.
static int ComputePcOffsetFromCodeOffset(Code *code, int code_offset) {
ASSERT_EQ(code->kind(), Code::FUNCTION);
int mask = RelocInfo::ModeMask(RelocInfo::DEBUG_BREAK_SLOT) |
RelocInfo::ModeMask(RelocInfo::CONST_POOL) |
RelocInfo::ModeMask(RelocInfo::VENEER_POOL);
int reloc = 0;
for (RelocIterator it(code, mask); !it.done(); it.next()) {
RelocInfo* info = it.rinfo();
if (info->pc() - code->instruction_start() - reloc >= code_offset) break;
if (RelocInfo::IsDebugBreakSlot(info->rmode())) {
reloc += Assembler::kDebugBreakSlotLength;
} else {
ASSERT(RelocInfo::IsConstPool(info->rmode()));
reloc += static_cast<int>(info->data());
}
}
int pc_offset = code_offset + reloc;
ASSERT_LT(code->instruction_start() + pc_offset, code->instruction_end());
return pc_offset;
}
static void RedirectActivationsToRecompiledCodeOnThread(
Isolate* isolate,
ThreadLocalTop* top) {
@ -1902,51 +1955,12 @@ static void RedirectActivationsToRecompiledCodeOnThread(
continue;
}
// Iterate over the RelocInfo in the original code to compute the sum of the
// constant pools and veneer pools sizes. (See Assembler::CheckConstPool()
// and Assembler::CheckVeneerPool())
// Note that this is only useful for architectures using constant pools or
// veneer pools.
int pool_mask = RelocInfo::ModeMask(RelocInfo::CONST_POOL) |
RelocInfo::ModeMask(RelocInfo::VENEER_POOL);
int frame_pool_size = 0;
for (RelocIterator it(*frame_code, pool_mask); !it.done(); it.next()) {
RelocInfo* info = it.rinfo();
if (info->pc() >= frame->pc()) break;
frame_pool_size += static_cast<int>(info->data());
}
intptr_t frame_offset =
frame->pc() - frame_code->instruction_start() - frame_pool_size;
// Iterate over the RelocInfo for new code to find the number of bytes
// generated for debug slots and constant pools.
int debug_break_slot_bytes = 0;
int new_code_pool_size = 0;
int mask = RelocInfo::ModeMask(RelocInfo::DEBUG_BREAK_SLOT) |
RelocInfo::ModeMask(RelocInfo::CONST_POOL) |
RelocInfo::ModeMask(RelocInfo::VENEER_POOL);
for (RelocIterator it(*new_code, mask); !it.done(); it.next()) {
// Check if the pc in the new code with debug break
// slots is before this slot.
RelocInfo* info = it.rinfo();
intptr_t new_offset = info->pc() - new_code->instruction_start() -
new_code_pool_size - debug_break_slot_bytes;
if (new_offset >= frame_offset) {
break;
}
if (RelocInfo::IsDebugBreakSlot(info->rmode())) {
debug_break_slot_bytes += Assembler::kDebugBreakSlotLength;
} else {
ASSERT(RelocInfo::IsConstPool(info->rmode()));
// The size of the pools is encoded in the data.
new_code_pool_size += static_cast<int>(info->data());
}
}
int old_pc_offset = frame->pc() - frame_code->instruction_start();
int code_offset = ComputeCodeOffsetFromPcOffset(*frame_code, old_pc_offset);
int new_pc_offset = ComputePcOffsetFromCodeOffset(*new_code, code_offset);
// Compute the equivalent pc in the new code.
byte* new_pc = new_code->instruction_start() + frame_offset +
debug_break_slot_bytes + new_code_pool_size;
byte* new_pc = new_code->instruction_start() + new_pc_offset;
if (FLAG_trace_deopt) {
PrintF("Replacing code %08" V8PRIxPTR " - %08" V8PRIxPTR " (%d) "
@ -2002,6 +2016,55 @@ class ActiveFunctionsRedirector : public ThreadVisitor {
};
class ForceDebuggerActive {
public:
explicit ForceDebuggerActive(Isolate *isolate) {
isolate_ = isolate;
old_state_ = isolate->debugger()->force_debugger_active();
isolate_->debugger()->set_force_debugger_active(true);
}
~ForceDebuggerActive() {
isolate_->debugger()->set_force_debugger_active(old_state_);
}
private:
Isolate *isolate_;
bool old_state_;
DISALLOW_COPY_AND_ASSIGN(ForceDebuggerActive);
};
void Debug::MaybeRecompileFunctionForDebugging(Handle<JSFunction> function) {
ASSERT_EQ(Code::FUNCTION, function->code()->kind());
ASSERT_EQ(function->code(), function->shared()->code());
if (function->code()->has_debug_break_slots()) return;
ForceDebuggerActive force_debugger_active(isolate_);
MaybeHandle<Code> code = Compiler::GetCodeForDebugging(function);
// Recompilation can fail. In that case leave the code as it was.
if (!code.is_null())
function->ReplaceCode(*code.ToHandleChecked());
ASSERT_EQ(function->code(), function->shared()->code());
}
void Debug::RecompileAndRelocateSuspendedGenerators(
const List<Handle<JSGeneratorObject> > &generators) {
for (int i = 0; i < generators.length(); i++) {
Handle<JSFunction> fun(generators[i]->function());
MaybeRecompileFunctionForDebugging(fun);
int code_offset = generators[i]->continuation();
int pc_offset = ComputePcOffsetFromCodeOffset(fun->code(), code_offset);
generators[i]->set_continuation(pc_offset);
}
}
void Debug::PrepareForBreakPoints() {
// If preparing for the first break point make sure to deoptimize all
// functions as debugging does not work with optimized code.
@ -2021,6 +2084,21 @@ void Debug::PrepareForBreakPoints() {
// is used both in GC and non-GC code.
List<Handle<JSFunction> > active_functions(100);
// A list of all suspended generators.
List<Handle<JSGeneratorObject> > suspended_generators;
// A list of all generator functions. We need to recompile all functions,
// but we don't know until after visiting the whole heap which generator
// functions have suspended activations and which do not. As in the case of
// functions with activations on the stack, we need to be careful with
// generator functions with suspended activations because although they
// should be recompiled, recompilation can fail, and we need to avoid
// leaving the heap in an inconsistent state.
//
// We could perhaps avoid this list and instead re-use the GC metadata
// links.
List<Handle<JSFunction> > generator_functions;
{
// We are going to iterate heap to find all functions without
// debug break slots.
@ -2028,6 +2106,9 @@ void Debug::PrepareForBreakPoints() {
heap->CollectAllGarbage(Heap::kMakeHeapIterableMask,
"preparing for breakpoints");
// Collecting the generators should not alter iterability of the heap.
ASSERT(heap->IsHeapIterable());
// Ensure no GC in this scope as we are going to use gc_metadata
// field in the Code object to mark active functions.
DisallowHeapAllocation no_allocation;
@ -2058,6 +2139,11 @@ void Debug::PrepareForBreakPoints() {
if (function->IsBuiltin()) continue;
if (shared->code()->gc_metadata() == active_code_marker) continue;
if (shared->is_generator()) {
generator_functions.Add(Handle<JSFunction>(function, isolate_));
continue;
}
Code::Kind kind = function->code()->kind();
if (kind == Code::FUNCTION &&
!function->code()->has_debug_break_slots()) {
@ -2077,6 +2163,24 @@ void Debug::PrepareForBreakPoints() {
function->shared()->set_code(*lazy_compile);
}
}
} else if (obj->IsJSGeneratorObject()) {
JSGeneratorObject* gen = JSGeneratorObject::cast(obj);
if (!gen->is_suspended()) continue;
JSFunction* fun = gen->function();
ASSERT_EQ(fun->code()->kind(), Code::FUNCTION);
if (fun->code()->has_debug_break_slots()) continue;
int pc_offset = gen->continuation();
ASSERT_LT(0, pc_offset);
int code_offset =
ComputeCodeOffsetFromPcOffset(fun->code(), pc_offset);
// This will be fixed after we recompile the functions.
gen->set_continuation(code_offset);
suspended_generators.Add(Handle<JSGeneratorObject>(gen, isolate_));
}
}
@ -2087,42 +2191,35 @@ void Debug::PrepareForBreakPoints() {
}
}
// Recompile generator functions that have suspended activations, and
// relocate those activations.
RecompileAndRelocateSuspendedGenerators(suspended_generators);
// Mark generator functions that didn't have suspended activations for lazy
// recompilation. Note that this set does not include any active functions.
for (int i = 0; i < generator_functions.length(); i++) {
Handle<JSFunction> &function = generator_functions[i];
if (function->code()->kind() != Code::FUNCTION) continue;
if (function->code()->has_debug_break_slots()) continue;
function->set_code(*lazy_compile);
function->shared()->set_code(*lazy_compile);
}
// Now recompile all functions with activation frames and and
// patch the return address to run in the new compiled code.
// patch the return address to run in the new compiled code. It could be
// that some active functions were recompiled already by the suspended
// generator recompilation pass above; a generator with suspended
// activations could also have active activations. That's fine.
for (int i = 0; i < active_functions.length(); i++) {
Handle<JSFunction> function = active_functions[i];
Handle<SharedFunctionInfo> shared(function->shared());
if (function->code()->kind() == Code::FUNCTION &&
function->code()->has_debug_break_slots()) {
// Nothing to do. Function code already had debug break slots.
continue;
}
// If recompilation is not possible just skip it.
if (shared->is_toplevel() ||
!shared->allows_lazy_compilation() ||
shared->code()->kind() == Code::BUILTIN) {
continue;
}
if (shared->is_toplevel()) continue;
if (!shared->allows_lazy_compilation()) continue;
if (shared->code()->kind() == Code::BUILTIN) continue;
// Make sure that the shared full code is compiled with debug
// break slots.
if (!shared->code()->has_debug_break_slots()) {
// Try to compile the full code with debug break slots. If it
// fails just keep the current code.
bool prev_force_debugger_active =
isolate_->debugger()->force_debugger_active();
isolate_->debugger()->set_force_debugger_active(true);
Handle<Code> code = Compiler::GetCodeForDebugging(
function).ToHandleChecked();
function->ReplaceCode(*code);
isolate_->debugger()->set_force_debugger_active(
prev_force_debugger_active);
}
// Keep function code in sync with shared function info.
function->set_code(shared->code());
MaybeRecompileFunctionForDebugging(function);
}
RedirectActivationsToRecompiledCodeOnThread(isolate_,

View File

@ -511,6 +511,10 @@ class Debug {
Handle<Object> CheckBreakPoints(Handle<Object> break_point);
bool CheckBreakPoint(Handle<Object> break_point_object);
void MaybeRecompileFunctionForDebugging(Handle<JSFunction> function);
void RecompileAndRelocateSuspendedGenerators(
const List<Handle<JSGeneratorObject> > &suspended_generators);
// Global handle to debug context where all the debugger JavaScript code is
// loaded.
Handle<Context> debug_context_;

View File

@ -5711,6 +5711,11 @@ SMI_ACCESSORS(JSGeneratorObject, continuation, kContinuationOffset)
ACCESSORS(JSGeneratorObject, operand_stack, FixedArray, kOperandStackOffset)
SMI_ACCESSORS(JSGeneratorObject, stack_handler_index, kStackHandlerIndexOffset)
bool JSGeneratorObject::is_suspended() {
ASSERT_LT(kGeneratorExecuting, kGeneratorClosed);
ASSERT_EQ(kGeneratorClosed, 0);
return continuation() > 0;
}
JSGeneratorObject* JSGeneratorObject::cast(Object* obj) {
ASSERT(obj->IsJSGeneratorObject());

View File

@ -7461,6 +7461,7 @@ class JSGeneratorObject: public JSObject {
// cannot be resumed.
inline int continuation();
inline void set_continuation(int continuation);
inline bool is_suspended();
// [operand_stack]: Saved operand stack.
DECL_ACCESSORS(operand_stack, FixedArray)

View File

@ -0,0 +1,61 @@
// Copyright 2014 the V8 project authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
// Flags: --expose-debug-as debug --harmony-generators
var Debug = debug.Debug;
function assertIteratorResult(value, done, result) {
assertEquals({value: value, done: done}, result);
}
function RunTest(formals_and_body, args, value1, value2) {
// A null listener. It isn't important what the listener does.
function listener(event, exec_state, event_data, data) {
}
// Create the generator function outside a debugging context. It will probably
// be lazily compiled.
var gen = (function*(){}).constructor.apply(null, formals_and_body);
// Instantiate the generator object.
var obj = gen.apply(null, args);
// Advance to the first yield.
assertIteratorResult(value1, false, obj.next());
// Add a breakpoint on line 3 (the second yield).
var bp = Debug.setBreakPoint(gen, 3);
// Enable the debugger, which should force recompilation of the generator
// function and relocation of the suspended generator activation.
Debug.setListener(listener);
// Check that the generator resumes and suspends properly.
assertIteratorResult(value2, false, obj.next());
// Disable debugger -- should not force recompilation.
Debug.clearBreakPoint(bp);
Debug.setListener(null);
// Run to completion.
assertIteratorResult(undefined, true, obj.next());
}
function prog(a, b, c) {
return a + ';\n' + 'yield ' + b + ';\n' + 'yield ' + c;
}
// Simple empty local scope.
RunTest([prog('', '1', '2')], [], 1, 2);
RunTest([prog('for (;;) break', '1', '2')], [], 1, 2);
RunTest([prog('while (0) foo()', '1', '2')], [], 1, 2);
RunTest(['a', prog('var x = 3', 'a', 'x')], [1], 1, 3);
RunTest(['a', prog('', '1', '2')], [42], 1, 2);
RunTest(['a', prog('for (;;) break', '1', '2')], [42], 1, 2);