// Copyright 2015 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #ifndef V8_INTERPRETER_INTERPRETER_ASSEMBLER_H_ #define V8_INTERPRETER_INTERPRETER_ASSEMBLER_H_ #include "src/builtins/builtins.h" #include "src/codegen/code-stub-assembler.h" #include "src/common/globals.h" #include "src/interpreter/bytecode-register.h" #include "src/interpreter/bytecodes.h" #include "src/runtime/runtime.h" #include "src/utils/allocation.h" namespace v8 { namespace internal { namespace interpreter { class V8_EXPORT_PRIVATE InterpreterAssembler : public CodeStubAssembler { public: InterpreterAssembler(compiler::CodeAssemblerState* state, Bytecode bytecode, OperandScale operand_scale); ~InterpreterAssembler(); // Returns the 32-bit unsigned count immediate for bytecode operand // |operand_index| in the current bytecode. TNode BytecodeOperandCount(int operand_index); // Returns the 32-bit unsigned flag for bytecode operand |operand_index| // in the current bytecode. TNode BytecodeOperandFlag(int operand_index); // Returns the 32-bit zero-extended index immediate for bytecode operand // |operand_index| in the current bytecode. TNode BytecodeOperandIdxInt32(int operand_index); // Returns the word zero-extended index immediate for bytecode operand // |operand_index| in the current bytecode. TNode BytecodeOperandIdx(int operand_index); // Returns the smi index immediate for bytecode operand |operand_index| // in the current bytecode. TNode BytecodeOperandIdxSmi(int operand_index); // Returns the TaggedIndex immediate for bytecode operand |operand_index| // in the current bytecode. TNode BytecodeOperandIdxTaggedIndex(int operand_index); // Returns the 32-bit unsigned immediate for bytecode operand |operand_index| // in the current bytecode. TNode BytecodeOperandUImm(int operand_index); // Returns the word-size unsigned immediate for bytecode operand // |operand_index| in the current bytecode. TNode BytecodeOperandUImmWord(int operand_index); // Returns the unsigned smi immediate for bytecode operand |operand_index| in // the current bytecode. TNode BytecodeOperandUImmSmi(int operand_index); // Returns the 32-bit signed immediate for bytecode operand |operand_index| // in the current bytecode. TNode BytecodeOperandImm(int operand_index); // Returns the word-size signed immediate for bytecode operand |operand_index| // in the current bytecode. TNode BytecodeOperandImmIntPtr(int operand_index); // Returns the smi immediate for bytecode operand |operand_index| in the // current bytecode. TNode BytecodeOperandImmSmi(int operand_index); // Returns the 32-bit unsigned runtime id immediate for bytecode operand // |operand_index| in the current bytecode. TNode BytecodeOperandRuntimeId(int operand_index); // Returns the word zero-extended native context index immediate for bytecode // operand |operand_index| in the current bytecode. TNode BytecodeOperandNativeContextIndex(int operand_index); // Returns the 32-bit unsigned intrinsic id immediate for bytecode operand // |operand_index| in the current bytecode. TNode BytecodeOperandIntrinsicId(int operand_index); // Accumulator. TNode GetAccumulator(); void SetAccumulator(TNode value); // Context. TNode GetContext(); void SetContext(TNode value); // Context at |depth| in the context chain starting at |context|. TNode GetContextAtDepth(TNode context, TNode depth); // Goto the given |target| if the context chain starting at |context| has any // extensions up to the given |depth|. void GotoIfHasContextExtensionUpToDepth(TNode context, TNode depth, Label* target); // A RegListNodePair provides an abstraction over lists of registers. class RegListNodePair { public: RegListNodePair(TNode base_reg_location, TNode reg_count) : base_reg_location_(base_reg_location), reg_count_(reg_count) {} TNode reg_count() const { return reg_count_; } TNode base_reg_location() const { return base_reg_location_; } private: TNode base_reg_location_; TNode reg_count_; }; // Backup/restore register file to/from a fixed array of the correct length. // There is an asymmetry between suspend/export and resume/import. // - Suspend copies arguments and registers to the generator. // - Resume copies only the registers from the generator, the arguments // are copied by the ResumeGenerator trampoline. TNode ExportParametersAndRegisterFile( TNode array, const RegListNodePair& registers, TNode formal_parameter_count); TNode ImportRegisterFile(TNode array, const RegListNodePair& registers, TNode formal_parameter_count); // Loads from and stores to the interpreter register file. TNode LoadRegister(Register reg); TNode LoadAndUntagRegister(Register reg); TNode LoadRegisterAtOperandIndex(int operand_index); std::pair, TNode> LoadRegisterPairAtOperandIndex( int operand_index); void StoreRegister(TNode value, Register reg); void StoreRegisterAtOperandIndex(TNode value, int operand_index); void StoreRegisterPairAtOperandIndex(TNode value1, TNode value2, int operand_index); void StoreRegisterTripleAtOperandIndex(TNode value1, TNode value2, TNode value3, int operand_index); RegListNodePair GetRegisterListAtOperandIndex(int operand_index); TNode LoadRegisterFromRegisterList(const RegListNodePair& reg_list, int index); TNode RegisterLocationInRegisterList(const RegListNodePair& reg_list, int index); // Load constant at the index specified in operand |operand_index| from the // constant pool. TNode LoadConstantPoolEntryAtOperandIndex(int operand_index); // Load and untag constant at the index specified in operand |operand_index| // from the constant pool. TNode LoadAndUntagConstantPoolEntryAtOperandIndex(int operand_index); // Load constant at |index| in the constant pool. TNode LoadConstantPoolEntry(TNode index); // Load and untag constant at |index| in the constant pool. TNode LoadAndUntagConstantPoolEntry(TNode index); // Load the FeedbackVector for the current function. The retuned node could be // undefined. TNode LoadFeedbackVector(); // Call JSFunction or Callable |function| with |args| arguments, possibly // including the receiver depending on |receiver_mode|. After the call returns // directly dispatches to the next bytecode. void CallJSAndDispatch(TNode function, TNode context, const RegListNodePair& args, ConvertReceiverMode receiver_mode); // Call JSFunction or Callable |function| with |arg_count| arguments (not // including receiver) passed as |args|, possibly including the receiver // depending on |receiver_mode|. After the call returns directly dispatches to // the next bytecode. template void CallJSAndDispatch(TNode function, TNode context, TNode arg_count, ConvertReceiverMode receiver_mode, TArgs... args); // Call JSFunction or Callable |function| with |args| // arguments (not including receiver), and the final argument being spread. // After the call returns directly dispatches to the next bytecode. void CallJSWithSpreadAndDispatch(TNode function, TNode context, const RegListNodePair& args, TNode slot_id, TNode maybe_feedback_vector); // Call constructor |target| with |args| arguments (not including receiver). // The |new_target| is the same as the |target| for the new keyword, but // differs for the super keyword. TNode Construct(TNode target, TNode context, TNode new_target, const RegListNodePair& args, TNode slot_id, TNode maybe_feedback_vector); // Call constructor |target| with |args| arguments (not including // receiver). The last argument is always a spread. The |new_target| is the // same as the |target| for the new keyword, but differs for the super // keyword. TNode ConstructWithSpread(TNode target, TNode context, TNode new_target, const RegListNodePair& args, TNode slot_id, TNode maybe_feedback_vector); // Call runtime function with |args| arguments which will return |return_size| // number of values. compiler::Node* CallRuntimeN(TNode function_id, TNode context, const RegListNodePair& args, int return_size = 1); // Jump forward relative to the current bytecode by the |jump_offset|. void Jump(TNode jump_offset); // Jump backward relative to the current bytecode by the |jump_offset|. void JumpBackward(TNode jump_offset); // Jump forward relative to the current bytecode by |jump_offset| if the // word values |lhs| and |rhs| are equal. void JumpIfTaggedEqual(TNode lhs, TNode rhs, TNode jump_offset); // Jump forward relative to the current bytecode by |jump_offset| if the // word values |lhs| and |rhs| are not equal. void JumpIfTaggedNotEqual(TNode lhs, TNode rhs, TNode jump_offset); // Updates the profiler interrupt budget for a return. void UpdateInterruptBudgetOnReturn(); // Returns the OSR nesting level from the bytecode header. TNode LoadOsrNestingLevel(); // Dispatch to the bytecode. void Dispatch(); // Dispatch bytecode as wide operand variant. void DispatchWide(OperandScale operand_scale); // Dispatch to |target_bytecode| at |new_bytecode_offset|. // |target_bytecode| should be equivalent to loading from the offset. void DispatchToBytecode(TNode target_bytecode, TNode new_bytecode_offset); // Abort with the given abort reason. void Abort(AbortReason abort_reason); void AbortIfWordNotEqual(TNode lhs, TNode rhs, AbortReason abort_reason); // Abort if |register_count| is invalid for given register file array. void AbortIfRegisterCountInvalid( TNode parameters_and_registers, TNode formal_parameter_count, TNode register_count); // Dispatch to frame dropper trampoline if necessary. void MaybeDropFrames(TNode context); // Returns the offset from the BytecodeArrayPointer of the current bytecode. TNode BytecodeOffset(); protected: Bytecode bytecode() const { return bytecode_; } static bool TargetSupportsUnalignedAccess(); void ToNumberOrNumeric(Object::Conversion mode); private: // Returns a pointer to the current function's BytecodeArray object. TNode BytecodeArrayTaggedPointer(); // Returns a pointer to first entry in the interpreter dispatch table. TNode DispatchTablePointer(); // Returns the accumulator value without checking whether bytecode // uses it. This is intended to be used only in dispatch and in // tracing as these need to bypass accumulator use validity checks. TNode GetAccumulatorUnchecked(); // Returns the frame pointer for the interpreted frame of the function being // interpreted. TNode GetInterpretedFramePointer(); // Operations on registers. TNode RegisterLocation(Register reg); TNode RegisterLocation(TNode reg_index); TNode NextRegister(TNode reg_index); TNode LoadRegister(TNode reg_index); void StoreRegister(TNode value, TNode reg_index); // Saves and restores interpreter bytecode offset to the interpreter stack // frame when performing a call. void CallPrologue(); void CallEpilogue(); // Increment the dispatch counter for the (current, next) bytecode pair. void TraceBytecodeDispatch(TNode target_bytecode); // Traces the current bytecode by calling |function_id|. void TraceBytecode(Runtime::FunctionId function_id); // Updates the bytecode array's interrupt budget by a 32-bit unsigned |weight| // and calls Runtime::kInterrupt if counter reaches zero. If |backward|, then // the interrupt budget is decremented, otherwise it is incremented. void UpdateInterruptBudget(TNode weight, bool backward); // Returns the offset of register |index| relative to RegisterFilePointer(). TNode RegisterFrameOffset(TNode index); // Returns the offset of an operand relative to the current bytecode offset. TNode OperandOffset(int operand_index); // Returns a value built from an sequence of bytes in the bytecode // array starting at |relative_offset| from the current bytecode. // The |result_type| determines the size and signedness. of the // value read. This method should only be used on architectures that // do not support unaligned memory accesses. TNode BytecodeOperandReadUnaligned( int relative_offset, MachineType result_type, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); // Returns zero- or sign-extended to word32 value of the operand. TNode BytecodeOperandUnsignedByte( int operand_index, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); TNode BytecodeOperandSignedByte( int operand_index, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); TNode BytecodeOperandUnsignedShort( int operand_index, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); TNode BytecodeOperandSignedShort( int operand_index, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); TNode BytecodeOperandUnsignedQuad( int operand_index, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); TNode BytecodeOperandSignedQuad( int operand_index, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); // Returns zero- or sign-extended to word32 value of the operand of // given size. TNode BytecodeSignedOperand( int operand_index, OperandSize operand_size, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); TNode BytecodeUnsignedOperand( int operand_index, OperandSize operand_size, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); // Returns the word-size sign-extended register index for bytecode operand // |operand_index| in the current bytecode. Value is not poisoned on // speculation since the value loaded from the register is poisoned instead. TNode BytecodeOperandReg( int operand_index, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); // Returns the word zero-extended index immediate for bytecode operand // |operand_index| in the current bytecode for use when loading a . TNode BytecodeOperandConstantPoolIdx( int operand_index, LoadSensitivity needs_poisoning = LoadSensitivity::kCritical); // Jump relative to the current bytecode by the |jump_offset|. If |backward|, // then jump backward (subtract the offset), otherwise jump forward (add the // offset). Helper function for Jump and JumpBackward. void Jump(TNode jump_offset, bool backward); // Jump forward relative to the current bytecode by |jump_offset| if the // |condition| is true. Helper function for JumpIfTaggedEqual and // JumpIfTaggedNotEqual. void JumpConditional(TNode condition, TNode jump_offset); // Save the bytecode offset to the interpreter frame. void SaveBytecodeOffset(); // Reload the bytecode offset from the interpreter frame. TNode ReloadBytecodeOffset(); // Updates and returns BytecodeOffset() advanced by the current bytecode's // size. Traces the exit of the current bytecode. TNode Advance(); // Updates and returns BytecodeOffset() advanced by delta bytecodes. // Traces the exit of the current bytecode. TNode Advance(int delta); TNode Advance(TNode delta, bool backward = false); // Load the bytecode at |bytecode_offset|. TNode LoadBytecode(TNode bytecode_offset); // Look ahead for Star and inline it in a branch. Returns a new target // bytecode node for dispatch. TNode StarDispatchLookahead(TNode target_bytecode); // Build code for Star at the current BytecodeOffset() and Advance() to the // next dispatch offset. void InlineStar(); // Dispatch to the bytecode handler with code entry point |handler_entry|. void DispatchToBytecodeHandlerEntry(TNode handler_entry, TNode bytecode_offset); int CurrentBytecodeSize() const; OperandScale operand_scale() const { return operand_scale_; } Bytecode bytecode_; OperandScale operand_scale_; CodeStubAssembler::TVariable interpreted_frame_pointer_; CodeStubAssembler::TVariable bytecode_array_; CodeStubAssembler::TVariable bytecode_offset_; CodeStubAssembler::TVariable dispatch_table_; CodeStubAssembler::TVariable accumulator_; AccumulatorUse accumulator_use_; bool made_call_; bool reloaded_frame_ptr_; bool bytecode_array_valid_; DISALLOW_COPY_AND_ASSIGN(InterpreterAssembler); }; } // namespace interpreter } // namespace internal } // namespace v8 #endif // V8_INTERPRETER_INTERPRETER_ASSEMBLER_H_