// Copyright 2012 the V8 project authors. All rights reserved. // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following // disclaimer in the documentation and/or other materials provided // with the distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived // from this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. #ifndef V8_CODE_STUBS_H_ #define V8_CODE_STUBS_H_ #include "allocation.h" #include "assembler.h" #include "globals.h" #include "codegen.h" namespace v8 { namespace internal { // List of code stubs used on all platforms. #define CODE_STUB_LIST_ALL_PLATFORMS(V) \ V(CallFunction) \ V(CallConstruct) \ V(UnaryOp) \ V(BinaryOp) \ V(StringAdd) \ V(SubString) \ V(StringCompare) \ V(Compare) \ V(CompareIC) \ V(CompareNilIC) \ V(MathPow) \ V(StringLength) \ V(FunctionPrototype) \ V(StoreArrayLength) \ V(RecordWrite) \ V(StoreBufferOverflow) \ V(RegExpExec) \ V(TranscendentalCache) \ V(Instanceof) \ V(ConvertToDouble) \ V(WriteInt32ToHeapNumber) \ V(StackCheck) \ V(Interrupt) \ V(FastNewClosure) \ V(FastNewContext) \ V(FastNewBlockContext) \ V(FastCloneShallowArray) \ V(FastCloneShallowObject) \ V(ToBoolean) \ V(ToNumber) \ V(ArgumentsAccess) \ V(RegExpConstructResult) \ V(NumberToString) \ V(CEntry) \ V(JSEntry) \ V(KeyedLoadElement) \ V(ArrayNoArgumentConstructor) \ V(ArraySingleArgumentConstructor) \ V(ArrayNArgumentsConstructor) \ V(InternalArrayNoArgumentConstructor) \ V(InternalArraySingleArgumentConstructor) \ V(InternalArrayNArgumentsConstructor) \ V(KeyedStoreElement) \ V(DebuggerStatement) \ V(NameDictionaryLookup) \ V(ElementsTransitionAndStore) \ V(TransitionElementsKind) \ V(StoreArrayLiteralElement) \ V(StubFailureTrampoline) \ V(ArrayConstructor) \ V(InternalArrayConstructor) \ V(ProfileEntryHook) \ /* IC Handler stubs */ \ V(LoadField) \ V(KeyedLoadField) // List of code stubs only used on ARM platforms. #ifdef V8_TARGET_ARCH_ARM #define CODE_STUB_LIST_ARM(V) \ V(GetProperty) \ V(SetProperty) \ V(InvokeBuiltin) \ V(RegExpCEntry) \ V(DirectCEntry) #else #define CODE_STUB_LIST_ARM(V) #endif // List of code stubs only used on MIPS platforms. #ifdef V8_TARGET_ARCH_MIPS #define CODE_STUB_LIST_MIPS(V) \ V(RegExpCEntry) \ V(DirectCEntry) #else #define CODE_STUB_LIST_MIPS(V) #endif // Combined list of code stubs. #define CODE_STUB_LIST(V) \ CODE_STUB_LIST_ALL_PLATFORMS(V) \ CODE_STUB_LIST_ARM(V) \ CODE_STUB_LIST_MIPS(V) // Mode to overwrite BinaryExpression values. enum OverwriteMode { NO_OVERWRITE, OVERWRITE_LEFT, OVERWRITE_RIGHT }; enum UnaryOverwriteMode { UNARY_OVERWRITE, UNARY_NO_OVERWRITE }; // Stub is base classes of all stubs. class CodeStub BASE_EMBEDDED { public: enum Major { #define DEF_ENUM(name) name, CODE_STUB_LIST(DEF_ENUM) #undef DEF_ENUM NoCache, // marker for stubs that do custom caching NUMBER_OF_IDS }; // Retrieve the code for the stub. Generate the code if needed. Handle GetCode(Isolate* isolate); static Major MajorKeyFromKey(uint32_t key) { return static_cast(MajorKeyBits::decode(key)); } static int MinorKeyFromKey(uint32_t key) { return MinorKeyBits::decode(key); } // Gets the major key from a code object that is a code stub or binary op IC. static Major GetMajorKey(Code* code_stub) { return static_cast(code_stub->major_key()); } static const char* MajorName(Major major_key, bool allow_unknown_keys); virtual ~CodeStub() {} bool CompilingCallsToThisStubIsGCSafe(Isolate* isolate) { bool is_pregenerated = IsPregenerated(); Code* code = NULL; CHECK(!is_pregenerated || FindCodeInCache(&code, isolate)); return is_pregenerated; } // See comment above, where Instanceof is defined. virtual bool IsPregenerated() { return false; } static void GenerateStubsAheadOfTime(Isolate* isolate); static void GenerateFPStubs(Isolate* isolate); // Some stubs put untagged junk on the stack that cannot be scanned by the // GC. This means that we must be statically sure that no GC can occur while // they are running. If that is the case they should override this to return // true, which will cause an assertion if we try to call something that can // GC or if we try to put a stack frame on top of the junk, which would not // result in a traversable stack. virtual bool SometimesSetsUpAFrame() { return true; } // Lookup the code in the (possibly custom) cache. bool FindCodeInCache(Code** code_out, Isolate* isolate); // Returns information for computing the number key. virtual Major MajorKey() = 0; virtual int MinorKey() = 0; virtual InlineCacheState GetICState() { return UNINITIALIZED; } virtual Code::ExtraICState GetExtraICState() { return Code::kNoExtraICState; } virtual Code::StubType GetStubType() { return Code::NORMAL; } virtual int GetStubFlags() { return -1; } protected: static bool CanUseFPRegisters(); // Generates the assembler code for the stub. virtual Handle GenerateCode() = 0; // Returns whether the code generated for this stub needs to be allocated as // a fixed (non-moveable) code object. virtual bool NeedsImmovableCode() { return false; } private: // Perform bookkeeping required after code generation when stub code is // initially generated. void RecordCodeGeneration(Code* code, Isolate* isolate); // Finish the code object after it has been generated. virtual void FinishCode(Handle code) { } // Activate newly generated stub. Is called after // registering stub in the stub cache. virtual void Activate(Code* code) { } // BinaryOpStub needs to override this. virtual Code::Kind GetCodeKind() const; // Add the code to a specialized cache, specific to an individual // stub type. Please note, this method must add the code object to a // roots object, otherwise we will remove the code during GC. virtual void AddToSpecialCache(Handle new_object) { } // Find code in a specialized cache, work is delegated to the specific stub. virtual bool FindCodeInSpecialCache(Code** code_out, Isolate* isolate) { return false; } // If a stub uses a special cache override this. virtual bool UseSpecialCache() { return false; } // Returns a name for logging/debugging purposes. SmartArrayPointer GetName(); virtual void PrintName(StringStream* stream); // Computes the key based on major and minor. uint32_t GetKey() { ASSERT(static_cast(MajorKey()) < NUMBER_OF_IDS); return MinorKeyBits::encode(MinorKey()) | MajorKeyBits::encode(MajorKey()); } class MajorKeyBits: public BitField {}; class MinorKeyBits: public BitField {}; // NOLINT friend class BreakPointIterator; }; class PlatformCodeStub : public CodeStub { public: // Retrieve the code for the stub. Generate the code if needed. virtual Handle GenerateCode(); virtual Code::Kind GetCodeKind() const { return Code::STUB; } protected: // Generates the assembler code for the stub. virtual void Generate(MacroAssembler* masm) = 0; }; enum StubFunctionMode { NOT_JS_FUNCTION_STUB_MODE, JS_FUNCTION_STUB_MODE }; struct CodeStubInterfaceDescriptor { CodeStubInterfaceDescriptor(); int register_param_count_; const Register* stack_parameter_count_; // if hint_stack_parameter_count_ > 0, the code stub can optimize the // return sequence. Default value is -1, which means it is ignored. int hint_stack_parameter_count_; StubFunctionMode function_mode_; Register* register_params_; Address deoptimization_handler_; int environment_length() const { if (stack_parameter_count_ != NULL) { return register_param_count_ + 1; } return register_param_count_; } bool initialized() const { return register_param_count_ >= 0; } void SetMissHandler(ExternalReference handler) { miss_handler_ = handler; has_miss_handler_ = true; } ExternalReference miss_handler() { ASSERT(has_miss_handler_); return miss_handler_; } bool has_miss_handler() { return has_miss_handler_; } private: ExternalReference miss_handler_; bool has_miss_handler_; }; // A helper to make up for the fact that type Register is not fully // defined outside of the platform directories #define DESCRIPTOR_GET_PARAMETER_REGISTER(descriptor, index) \ ((index) == (descriptor)->register_param_count_) \ ? *((descriptor)->stack_parameter_count_) \ : (descriptor)->register_params_[(index)] class HydrogenCodeStub : public CodeStub { public: enum InitializationState { UNINITIALIZED, INITIALIZED }; explicit HydrogenCodeStub(InitializationState state = INITIALIZED) { is_uninitialized_ = (state == UNINITIALIZED); } virtual Code::Kind GetCodeKind() const { return Code::STUB; } CodeStubInterfaceDescriptor* GetInterfaceDescriptor(Isolate* isolate) { return isolate->code_stub_interface_descriptor(MajorKey()); } bool IsUninitialized() { return is_uninitialized_; } template static Handle GetUninitialized(Isolate* isolate) { SubClass::GenerateAheadOfTime(isolate); return SubClass().GetCode(isolate); } virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor) = 0; // Retrieve the code for the stub. Generate the code if needed. virtual Handle GenerateCode() = 0; virtual int NotMissMinorKey() = 0; Handle GenerateLightweightMissCode(Isolate* isolate); private: class MinorKeyBits: public BitField {}; class IsMissBits: public BitField {}; void GenerateLightweightMiss(MacroAssembler* masm); virtual int MinorKey() { return IsMissBits::encode(is_uninitialized_) | MinorKeyBits::encode(NotMissMinorKey()); } bool is_uninitialized_; }; // Helper interface to prepare to/restore after making runtime calls. class RuntimeCallHelper { public: virtual ~RuntimeCallHelper() {} virtual void BeforeCall(MacroAssembler* masm) const = 0; virtual void AfterCall(MacroAssembler* masm) const = 0; protected: RuntimeCallHelper() {} private: DISALLOW_COPY_AND_ASSIGN(RuntimeCallHelper); }; } } // namespace v8::internal #if V8_TARGET_ARCH_IA32 #include "ia32/code-stubs-ia32.h" #elif V8_TARGET_ARCH_X64 #include "x64/code-stubs-x64.h" #elif V8_TARGET_ARCH_ARM #include "arm/code-stubs-arm.h" #elif V8_TARGET_ARCH_MIPS #include "mips/code-stubs-mips.h" #else #error Unsupported target architecture. #endif namespace v8 { namespace internal { // RuntimeCallHelper implementation used in stubs: enters/leaves a // newly created internal frame before/after the runtime call. class StubRuntimeCallHelper : public RuntimeCallHelper { public: StubRuntimeCallHelper() {} virtual void BeforeCall(MacroAssembler* masm) const; virtual void AfterCall(MacroAssembler* masm) const; }; // Trivial RuntimeCallHelper implementation. class NopRuntimeCallHelper : public RuntimeCallHelper { public: NopRuntimeCallHelper() {} virtual void BeforeCall(MacroAssembler* masm) const {} virtual void AfterCall(MacroAssembler* masm) const {} }; class StackCheckStub : public PlatformCodeStub { public: StackCheckStub() { } void Generate(MacroAssembler* masm); private: Major MajorKey() { return StackCheck; } int MinorKey() { return 0; } }; class InterruptStub : public PlatformCodeStub { public: InterruptStub() { } void Generate(MacroAssembler* masm); private: Major MajorKey() { return Interrupt; } int MinorKey() { return 0; } }; class ToNumberStub: public PlatformCodeStub { public: ToNumberStub() { } void Generate(MacroAssembler* masm); private: Major MajorKey() { return ToNumber; } int MinorKey() { return 0; } }; class FastNewClosureStub : public PlatformCodeStub { public: explicit FastNewClosureStub(LanguageMode language_mode, bool is_generator) : language_mode_(language_mode), is_generator_(is_generator) { } void Generate(MacroAssembler* masm); private: class StrictModeBits: public BitField {}; class IsGeneratorBits: public BitField {}; Major MajorKey() { return FastNewClosure; } int MinorKey() { return StrictModeBits::encode(language_mode_ != CLASSIC_MODE) | IsGeneratorBits::encode(is_generator_); } LanguageMode language_mode_; bool is_generator_; }; class FastNewContextStub : public PlatformCodeStub { public: static const int kMaximumSlots = 64; explicit FastNewContextStub(int slots) : slots_(slots) { ASSERT(slots_ > 0 && slots_ <= kMaximumSlots); } void Generate(MacroAssembler* masm); private: int slots_; Major MajorKey() { return FastNewContext; } int MinorKey() { return slots_; } }; class FastNewBlockContextStub : public PlatformCodeStub { public: static const int kMaximumSlots = 64; explicit FastNewBlockContextStub(int slots) : slots_(slots) { ASSERT(slots_ > 0 && slots_ <= kMaximumSlots); } void Generate(MacroAssembler* masm); private: int slots_; Major MajorKey() { return FastNewBlockContext; } int MinorKey() { return slots_; } }; class FastCloneShallowArrayStub : public HydrogenCodeStub { public: // Maximum length of copied elements array. static const int kMaximumClonedLength = 8; enum Mode { CLONE_ELEMENTS, CLONE_DOUBLE_ELEMENTS, COPY_ON_WRITE_ELEMENTS, CLONE_ANY_ELEMENTS, LAST_CLONE_MODE = CLONE_ANY_ELEMENTS }; static const int kFastCloneModeCount = LAST_CLONE_MODE + 1; FastCloneShallowArrayStub(Mode mode, AllocationSiteMode allocation_site_mode, int length) : mode_(mode), allocation_site_mode_(allocation_site_mode), length_((mode == COPY_ON_WRITE_ELEMENTS) ? 0 : length) { ASSERT_GE(length_, 0); ASSERT_LE(length_, kMaximumClonedLength); } Mode mode() const { return mode_; } int length() const { return length_; } AllocationSiteMode allocation_site_mode() const { return allocation_site_mode_; } ElementsKind ComputeElementsKind() const { switch (mode()) { case CLONE_ELEMENTS: case COPY_ON_WRITE_ELEMENTS: return FAST_ELEMENTS; case CLONE_DOUBLE_ELEMENTS: return FAST_DOUBLE_ELEMENTS; case CLONE_ANY_ELEMENTS: /*fall-through*/; } UNREACHABLE(); return LAST_ELEMENTS_KIND; } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: Mode mode_; AllocationSiteMode allocation_site_mode_; int length_; class AllocationSiteModeBits: public BitField {}; class ModeBits: public BitField {}; class LengthBits: public BitField {}; // Ensure data fits within available bits. STATIC_ASSERT(LAST_ALLOCATION_SITE_MODE == 1); STATIC_ASSERT(kFastCloneModeCount < 16); STATIC_ASSERT(kMaximumClonedLength < 16); Major MajorKey() { return FastCloneShallowArray; } int NotMissMinorKey() { return AllocationSiteModeBits::encode(allocation_site_mode_) | ModeBits::encode(mode_) | LengthBits::encode(length_); } }; class FastCloneShallowObjectStub : public HydrogenCodeStub { public: // Maximum number of properties in copied object. static const int kMaximumClonedProperties = 6; explicit FastCloneShallowObjectStub(int length) : length_(length) { ASSERT_GE(length_, 0); ASSERT_LE(length_, kMaximumClonedProperties); } int length() const { return length_; } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: int length_; Major MajorKey() { return FastCloneShallowObject; } int NotMissMinorKey() { return length_; } DISALLOW_COPY_AND_ASSIGN(FastCloneShallowObjectStub); }; class InstanceofStub: public PlatformCodeStub { public: enum Flags { kNoFlags = 0, kArgsInRegisters = 1 << 0, kCallSiteInlineCheck = 1 << 1, kReturnTrueFalseObject = 1 << 2 }; explicit InstanceofStub(Flags flags) : flags_(flags) { } static Register left(); static Register right(); void Generate(MacroAssembler* masm); private: Major MajorKey() { return Instanceof; } int MinorKey() { return static_cast(flags_); } bool HasArgsInRegisters() const { return (flags_ & kArgsInRegisters) != 0; } bool HasCallSiteInlineCheck() const { return (flags_ & kCallSiteInlineCheck) != 0; } bool ReturnTrueFalseObject() const { return (flags_ & kReturnTrueFalseObject) != 0; } virtual void PrintName(StringStream* stream); Flags flags_; }; class ArrayConstructorStub: public PlatformCodeStub { public: enum ArgumentCountKey { ANY, NONE, ONE, MORE_THAN_ONE }; ArrayConstructorStub(Isolate* isolate, int argument_count); explicit ArrayConstructorStub(Isolate* isolate); void Generate(MacroAssembler* masm); private: virtual CodeStub::Major MajorKey() { return ArrayConstructor; } virtual int MinorKey() { return argument_count_; } ArgumentCountKey argument_count_; }; class InternalArrayConstructorStub: public PlatformCodeStub { public: explicit InternalArrayConstructorStub(Isolate* isolate); void Generate(MacroAssembler* masm); private: virtual CodeStub::Major MajorKey() { return InternalArrayConstructor; } virtual int MinorKey() { return 0; } void GenerateCase(MacroAssembler* masm, ElementsKind kind); }; class MathPowStub: public PlatformCodeStub { public: enum ExponentType { INTEGER, DOUBLE, TAGGED, ON_STACK }; explicit MathPowStub(ExponentType exponent_type) : exponent_type_(exponent_type) { } virtual void Generate(MacroAssembler* masm); private: virtual CodeStub::Major MajorKey() { return MathPow; } virtual int MinorKey() { return exponent_type_; } ExponentType exponent_type_; }; class ICStub: public PlatformCodeStub { public: explicit ICStub(Code::Kind kind) : kind_(kind) { } virtual Code::Kind GetCodeKind() const { return kind_; } virtual InlineCacheState GetICState() { return MONOMORPHIC; } bool Describes(Code* code) { return GetMajorKey(code) == MajorKey() && code->stub_info() == MinorKey(); } protected: class KindBits: public BitField {}; virtual void FinishCode(Handle code) { code->set_stub_info(MinorKey()); } Code::Kind kind() { return kind_; } virtual int MinorKey() { return KindBits::encode(kind_); } private: Code::Kind kind_; }; class FunctionPrototypeStub: public ICStub { public: explicit FunctionPrototypeStub(Code::Kind kind) : ICStub(kind) { } virtual void Generate(MacroAssembler* masm); private: virtual CodeStub::Major MajorKey() { return FunctionPrototype; } }; class StringLengthStub: public ICStub { public: StringLengthStub(Code::Kind kind, bool support_wrapper) : ICStub(kind), support_wrapper_(support_wrapper) { } virtual void Generate(MacroAssembler* masm); private: STATIC_ASSERT(KindBits::kSize == 4); class WrapperModeBits: public BitField {}; virtual CodeStub::Major MajorKey() { return StringLength; } virtual int MinorKey() { return KindBits::encode(kind()) | WrapperModeBits::encode(support_wrapper_); } bool support_wrapper_; }; class StoreICStub: public ICStub { public: StoreICStub(Code::Kind kind, StrictModeFlag strict_mode) : ICStub(kind), strict_mode_(strict_mode) { } protected: virtual Code::ExtraICState GetExtraICState() { return strict_mode_; } private: STATIC_ASSERT(KindBits::kSize == 4); class StrictModeBits: public BitField {}; virtual int MinorKey() { return KindBits::encode(kind()) | StrictModeBits::encode(strict_mode_); } StrictModeFlag strict_mode_; }; class StoreArrayLengthStub: public StoreICStub { public: explicit StoreArrayLengthStub(Code::Kind kind, StrictModeFlag strict_mode) : StoreICStub(kind, strict_mode) { } virtual void Generate(MacroAssembler* masm); private: virtual CodeStub::Major MajorKey() { return StoreArrayLength; } }; class HICStub: public HydrogenCodeStub { public: virtual Code::Kind GetCodeKind() const { return kind(); } virtual InlineCacheState GetICState() { return MONOMORPHIC; } protected: HICStub() { } class KindBits: public BitField {}; virtual Code::Kind kind() const = 0; }; class HandlerStub: public HICStub { public: virtual Code::Kind GetCodeKind() const { return Code::STUB; } virtual int GetStubFlags() { return kind(); } protected: HandlerStub() : HICStub() { } }; class LoadFieldStub: public HandlerStub { public: LoadFieldStub(bool inobject, int index, Representation representation) : HandlerStub() { Initialize(Code::LOAD_IC, inobject, index, representation); } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); Representation representation() { if (unboxed_double()) return Representation::Double(); return Representation::Tagged(); } virtual Code::Kind kind() const { return KindBits::decode(bit_field_); } bool is_inobject() { return InobjectBits::decode(bit_field_); } int offset() { int index = IndexBits::decode(bit_field_); int offset = index * kPointerSize; if (is_inobject()) return offset; return FixedArray::kHeaderSize + offset; } bool unboxed_double() { return UnboxedDoubleBits::decode(bit_field_); } virtual Code::StubType GetStubType() { return Code::FIELD; } protected: LoadFieldStub() : HandlerStub() { } void Initialize(Code::Kind kind, bool inobject, int index, Representation representation) { bool unboxed_double = FLAG_track_double_fields && representation.IsDouble(); bit_field_ = KindBits::encode(kind) | InobjectBits::encode(inobject) | IndexBits::encode(index) | UnboxedDoubleBits::encode(unboxed_double); } private: STATIC_ASSERT(KindBits::kSize == 4); class InobjectBits: public BitField {}; class IndexBits: public BitField {}; class UnboxedDoubleBits: public BitField {}; virtual CodeStub::Major MajorKey() { return LoadField; } virtual int NotMissMinorKey() { return bit_field_; } int bit_field_; }; class KeyedLoadFieldStub: public LoadFieldStub { public: KeyedLoadFieldStub(bool inobject, int index, Representation representation) : LoadFieldStub() { Initialize(Code::KEYED_LOAD_IC, inobject, index, representation); } virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); virtual Handle GenerateCode(); private: virtual CodeStub::Major MajorKey() { return KeyedLoadField; } }; class BinaryOpStub: public PlatformCodeStub { public: BinaryOpStub(Token::Value op, OverwriteMode mode) : op_(op), mode_(mode), platform_specific_bit_(false), left_type_(BinaryOpIC::UNINITIALIZED), right_type_(BinaryOpIC::UNINITIALIZED), result_type_(BinaryOpIC::UNINITIALIZED), encoded_right_arg_(false, encode_arg_value(1)) { Initialize(); ASSERT(OpBits::is_valid(Token::NUM_TOKENS)); } BinaryOpStub( int key, BinaryOpIC::TypeInfo left_type, BinaryOpIC::TypeInfo right_type, BinaryOpIC::TypeInfo result_type, Maybe fixed_right_arg) : op_(OpBits::decode(key)), mode_(ModeBits::decode(key)), platform_specific_bit_(PlatformSpecificBits::decode(key)), left_type_(left_type), right_type_(right_type), result_type_(result_type), encoded_right_arg_(fixed_right_arg.has_value, encode_arg_value(fixed_right_arg.value)) { } static void decode_types_from_minor_key(int minor_key, BinaryOpIC::TypeInfo* left_type, BinaryOpIC::TypeInfo* right_type, BinaryOpIC::TypeInfo* result_type) { *left_type = static_cast(LeftTypeBits::decode(minor_key)); *right_type = static_cast(RightTypeBits::decode(minor_key)); *result_type = static_cast(ResultTypeBits::decode(minor_key)); } static Token::Value decode_op_from_minor_key(int minor_key) { return static_cast(OpBits::decode(minor_key)); } static Maybe decode_fixed_right_arg_from_minor_key(int minor_key) { return Maybe( HasFixedRightArgBits::decode(minor_key), decode_arg_value(FixedRightArgValueBits::decode(minor_key))); } int fixed_right_arg_value() const { return decode_arg_value(encoded_right_arg_.value); } static bool can_encode_arg_value(int32_t value) { return value > 0 && IsPowerOf2(value) && FixedRightArgValueBits::is_valid(WhichPowerOf2(value)); } enum SmiCodeGenerateHeapNumberResults { ALLOW_HEAPNUMBER_RESULTS, NO_HEAPNUMBER_RESULTS }; private: Token::Value op_; OverwriteMode mode_; bool platform_specific_bit_; // Indicates SSE3 on IA32. // Operand type information determined at runtime. BinaryOpIC::TypeInfo left_type_; BinaryOpIC::TypeInfo right_type_; BinaryOpIC::TypeInfo result_type_; Maybe encoded_right_arg_; static int encode_arg_value(int32_t value) { ASSERT(can_encode_arg_value(value)); return WhichPowerOf2(value); } static int32_t decode_arg_value(int value) { return 1 << value; } virtual void PrintName(StringStream* stream); // Minor key encoding in all 25 bits FFFFFHTTTRRRLLLPOOOOOOOMM. // Note: We actually do not need 7 bits for the operation, just 4 bits to // encode ADD, SUB, MUL, DIV, MOD, BIT_OR, BIT_AND, BIT_XOR, SAR, SHL, SHR. class ModeBits: public BitField {}; class OpBits: public BitField {}; class PlatformSpecificBits: public BitField {}; class LeftTypeBits: public BitField {}; class RightTypeBits: public BitField {}; class ResultTypeBits: public BitField {}; class HasFixedRightArgBits: public BitField {}; class FixedRightArgValueBits: public BitField {}; Major MajorKey() { return BinaryOp; } int MinorKey() { return OpBits::encode(op_) | ModeBits::encode(mode_) | PlatformSpecificBits::encode(platform_specific_bit_) | LeftTypeBits::encode(left_type_) | RightTypeBits::encode(right_type_) | ResultTypeBits::encode(result_type_) | HasFixedRightArgBits::encode(encoded_right_arg_.has_value) | FixedRightArgValueBits::encode(encoded_right_arg_.value); } // Platform-independent implementation. void Generate(MacroAssembler* masm); void GenerateCallRuntime(MacroAssembler* masm); // Platform-independent signature, platform-specific implementation. void Initialize(); void GenerateAddStrings(MacroAssembler* masm); void GenerateBothStringStub(MacroAssembler* masm); void GenerateGeneric(MacroAssembler* masm); void GenerateGenericStub(MacroAssembler* masm); void GenerateNumberStub(MacroAssembler* masm); void GenerateInt32Stub(MacroAssembler* masm); void GenerateLoadArguments(MacroAssembler* masm); void GenerateOddballStub(MacroAssembler* masm); void GenerateRegisterArgsPush(MacroAssembler* masm); void GenerateReturn(MacroAssembler* masm); void GenerateSmiStub(MacroAssembler* masm); void GenerateStringStub(MacroAssembler* masm); void GenerateTypeTransition(MacroAssembler* masm); void GenerateTypeTransitionWithSavedArgs(MacroAssembler* masm); void GenerateUninitializedStub(MacroAssembler* masm); // Entirely platform-specific methods are defined as static helper // functions in the /code-stubs-.cc files. virtual Code::Kind GetCodeKind() const { return Code::BINARY_OP_IC; } virtual InlineCacheState GetICState() { return BinaryOpIC::ToState(Max(left_type_, right_type_)); } virtual void FinishCode(Handle code) { code->set_stub_info(MinorKey()); } friend class CodeGenerator; }; class ICCompareStub: public PlatformCodeStub { public: ICCompareStub(Token::Value op, CompareIC::State left, CompareIC::State right, CompareIC::State handler) : op_(op), left_(left), right_(right), state_(handler) { ASSERT(Token::IsCompareOp(op)); } virtual void Generate(MacroAssembler* masm); void set_known_map(Handle map) { known_map_ = map; } static void DecodeMinorKey(int minor_key, CompareIC::State* left_state, CompareIC::State* right_state, CompareIC::State* handler_state, Token::Value* op); static CompareIC::State CompareState(int minor_key) { return static_cast(HandlerStateField::decode(minor_key)); } virtual InlineCacheState GetICState(); private: class OpField: public BitField { }; class LeftStateField: public BitField { }; class RightStateField: public BitField { }; class HandlerStateField: public BitField { }; virtual void FinishCode(Handle code) { code->set_stub_info(MinorKey()); } virtual CodeStub::Major MajorKey() { return CompareIC; } virtual int MinorKey(); virtual Code::Kind GetCodeKind() const { return Code::COMPARE_IC; } void GenerateSmis(MacroAssembler* masm); void GenerateNumbers(MacroAssembler* masm); void GenerateInternalizedStrings(MacroAssembler* masm); void GenerateStrings(MacroAssembler* masm); void GenerateUniqueNames(MacroAssembler* masm); void GenerateObjects(MacroAssembler* masm); void GenerateMiss(MacroAssembler* masm); void GenerateKnownObjects(MacroAssembler* masm); void GenerateGeneric(MacroAssembler* masm); bool strict() const { return op_ == Token::EQ_STRICT; } Condition GetCondition() const { return CompareIC::ComputeCondition(op_); } virtual void AddToSpecialCache(Handle new_object); virtual bool FindCodeInSpecialCache(Code** code_out, Isolate* isolate); virtual bool UseSpecialCache() { return state_ == CompareIC::KNOWN_OBJECT; } Token::Value op_; CompareIC::State left_; CompareIC::State right_; CompareIC::State state_; Handle known_map_; }; class CompareNilICStub : public HydrogenCodeStub { public: enum CompareNilType { UNDEFINED, NULL_TYPE, MONOMORPHIC_MAP, UNDETECTABLE, GENERIC, NUMBER_OF_TYPES }; class State : public EnumSet { public: State() : EnumSet(0) { } explicit State(byte bits) : EnumSet(bits) { } static State Generic() { State set; set.Add(UNDEFINED); set.Add(NULL_TYPE); set.Add(UNDETECTABLE); set.Add(GENERIC); return set; } void Print(StringStream* stream) const; void TraceTransition(State to) const; }; static Handle StateToType( Isolate* isolate, State state, Handle map = Handle()); // At most 6 different types can be distinguished, because the Code object // only has room for a single byte to hold a set and there are two more // boolean flags we need to store. :-P STATIC_ASSERT(NUMBER_OF_TYPES <= 6); CompareNilICStub(NilValue nil, State state = State()) : nil_value_(nil), state_(state) { } CompareNilICStub(Code::ExtraICState ic_state, InitializationState init_state = INITIALIZED) : HydrogenCodeStub(init_state) { nil_value_ = NilValueField::decode(ic_state); state_ = State(ExtractTypesFromExtraICState(ic_state)); } static Handle GetUninitialized(Isolate* isolate, NilValue nil) { return CompareNilICStub(nil, UNINITIALIZED).GetCode(isolate); } virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); static void InitializeForIsolate(Isolate* isolate) { CompareNilICStub compare_stub(kNullValue, UNINITIALIZED); compare_stub.InitializeInterfaceDescriptor( isolate, isolate->code_stub_interface_descriptor(CodeStub::CompareNilIC)); } virtual InlineCacheState GetICState() { if (state_ == State::Generic()) { return MEGAMORPHIC; } else if (state_.Contains(MONOMORPHIC_MAP)) { return MONOMORPHIC; } else { return PREMONOMORPHIC; } } virtual Code::Kind GetCodeKind() const { return Code::COMPARE_NIL_IC; } Handle GenerateCode(); // extra ic state = nil_value | type_n-1 | ... | type_0 virtual Code::ExtraICState GetExtraICState() { return NilValueField::encode(nil_value_) | state_.ToIntegral(); } static byte ExtractTypesFromExtraICState(Code::ExtraICState state) { return state & ((1 << NUMBER_OF_TYPES) - 1); } static NilValue ExtractNilValueFromExtraICState(Code::ExtraICState state) { return NilValueField::decode(state); } void Record(Handle object); bool IsMonomorphic() const { return state_.Contains(MONOMORPHIC_MAP); } NilValue GetNilValue() const { return nil_value_; } State GetState() const { return state_; } void ClearState() { state_.RemoveAll(); } virtual void PrintName(StringStream* stream); private: friend class CompareNilIC; CompareNilICStub(NilValue nil, InitializationState init_state) : HydrogenCodeStub(init_state) { nil_value_ = nil; } class NilValueField : public BitField {}; virtual CodeStub::Major MajorKey() { return CompareNilIC; } virtual int NotMissMinorKey() { return GetExtraICState(); } NilValue nil_value_; State state_; DISALLOW_COPY_AND_ASSIGN(CompareNilICStub); }; class CEntryStub : public PlatformCodeStub { public: explicit CEntryStub(int result_size, SaveFPRegsMode save_doubles = kDontSaveFPRegs) : result_size_(result_size), save_doubles_(save_doubles) { } void Generate(MacroAssembler* masm); // The version of this stub that doesn't save doubles is generated ahead of // time, so it's OK to call it from other stubs that can't cope with GC during // their code generation. On machines that always have gp registers (x64) we // can generate both variants ahead of time. virtual bool IsPregenerated(); static void GenerateAheadOfTime(Isolate* isolate); private: void GenerateCore(MacroAssembler* masm, Label* throw_normal_exception, Label* throw_termination_exception, Label* throw_out_of_memory_exception, bool do_gc, bool always_allocate_scope); // Number of pointers/values returned. const int result_size_; SaveFPRegsMode save_doubles_; Major MajorKey() { return CEntry; } int MinorKey(); bool NeedsImmovableCode(); }; class JSEntryStub : public PlatformCodeStub { public: JSEntryStub() { } void Generate(MacroAssembler* masm) { GenerateBody(masm, false); } protected: void GenerateBody(MacroAssembler* masm, bool is_construct); private: Major MajorKey() { return JSEntry; } int MinorKey() { return 0; } virtual void FinishCode(Handle code); int handler_offset_; }; class JSConstructEntryStub : public JSEntryStub { public: JSConstructEntryStub() { } void Generate(MacroAssembler* masm) { GenerateBody(masm, true); } private: int MinorKey() { return 1; } virtual void PrintName(StringStream* stream) { stream->Add("JSConstructEntryStub"); } }; class ArgumentsAccessStub: public PlatformCodeStub { public: enum Type { READ_ELEMENT, NEW_NON_STRICT_FAST, NEW_NON_STRICT_SLOW, NEW_STRICT }; explicit ArgumentsAccessStub(Type type) : type_(type) { } private: Type type_; Major MajorKey() { return ArgumentsAccess; } int MinorKey() { return type_; } void Generate(MacroAssembler* masm); void GenerateReadElement(MacroAssembler* masm); void GenerateNewStrict(MacroAssembler* masm); void GenerateNewNonStrictFast(MacroAssembler* masm); void GenerateNewNonStrictSlow(MacroAssembler* masm); virtual void PrintName(StringStream* stream); }; class RegExpExecStub: public PlatformCodeStub { public: RegExpExecStub() { } private: Major MajorKey() { return RegExpExec; } int MinorKey() { return 0; } void Generate(MacroAssembler* masm); }; class RegExpConstructResultStub: public PlatformCodeStub { public: RegExpConstructResultStub() { } private: Major MajorKey() { return RegExpConstructResult; } int MinorKey() { return 0; } void Generate(MacroAssembler* masm); }; class CallFunctionStub: public PlatformCodeStub { public: CallFunctionStub(int argc, CallFunctionFlags flags) : argc_(argc), flags_(flags) { } void Generate(MacroAssembler* masm); virtual void FinishCode(Handle code) { code->set_has_function_cache(RecordCallTarget()); } static int ExtractArgcFromMinorKey(int minor_key) { return ArgcBits::decode(minor_key); } private: int argc_; CallFunctionFlags flags_; virtual void PrintName(StringStream* stream); // Minor key encoding in 32 bits with Bitfield . class FlagBits: public BitField {}; class ArgcBits: public BitField {}; Major MajorKey() { return CallFunction; } int MinorKey() { // Encode the parameters in a unique 32 bit value. return FlagBits::encode(flags_) | ArgcBits::encode(argc_); } bool ReceiverMightBeImplicit() { return (flags_ & RECEIVER_MIGHT_BE_IMPLICIT) != 0; } bool RecordCallTarget() { return (flags_ & RECORD_CALL_TARGET) != 0; } }; class CallConstructStub: public PlatformCodeStub { public: explicit CallConstructStub(CallFunctionFlags flags) : flags_(flags) {} void Generate(MacroAssembler* masm); virtual void FinishCode(Handle code) { code->set_has_function_cache(RecordCallTarget()); } private: CallFunctionFlags flags_; virtual void PrintName(StringStream* stream); Major MajorKey() { return CallConstruct; } int MinorKey() { return flags_; } bool RecordCallTarget() { return (flags_ & RECORD_CALL_TARGET) != 0; } }; enum StringIndexFlags { // Accepts smis or heap numbers. STRING_INDEX_IS_NUMBER, // Accepts smis or heap numbers that are valid array indices // (ECMA-262 15.4). Invalid indices are reported as being out of // range. STRING_INDEX_IS_ARRAY_INDEX }; // Generates code implementing String.prototype.charCodeAt. // // Only supports the case when the receiver is a string and the index // is a number (smi or heap number) that is a valid index into the // string. Additional index constraints are specified by the // flags. Otherwise, bails out to the provided labels. // // Register usage: |object| may be changed to another string in a way // that doesn't affect charCodeAt/charAt semantics, |index| is // preserved, |scratch| and |result| are clobbered. class StringCharCodeAtGenerator { public: StringCharCodeAtGenerator(Register object, Register index, Register result, Label* receiver_not_string, Label* index_not_number, Label* index_out_of_range, StringIndexFlags index_flags) : object_(object), index_(index), result_(result), receiver_not_string_(receiver_not_string), index_not_number_(index_not_number), index_out_of_range_(index_out_of_range), index_flags_(index_flags) { ASSERT(!result_.is(object_)); ASSERT(!result_.is(index_)); } // Generates the fast case code. On the fallthrough path |result| // register contains the result. void GenerateFast(MacroAssembler* masm); // Generates the slow case code. Must not be naturally // reachable. Expected to be put after a ret instruction (e.g., in // deferred code). Always jumps back to the fast case. void GenerateSlow(MacroAssembler* masm, const RuntimeCallHelper& call_helper); // Skip handling slow case and directly jump to bailout. void SkipSlow(MacroAssembler* masm, Label* bailout) { masm->bind(&index_not_smi_); masm->bind(&call_runtime_); masm->jmp(bailout); } private: Register object_; Register index_; Register result_; Label* receiver_not_string_; Label* index_not_number_; Label* index_out_of_range_; StringIndexFlags index_flags_; Label call_runtime_; Label index_not_smi_; Label got_smi_index_; Label exit_; DISALLOW_COPY_AND_ASSIGN(StringCharCodeAtGenerator); }; // Generates code for creating a one-char string from a char code. class StringCharFromCodeGenerator { public: StringCharFromCodeGenerator(Register code, Register result) : code_(code), result_(result) { ASSERT(!code_.is(result_)); } // Generates the fast case code. On the fallthrough path |result| // register contains the result. void GenerateFast(MacroAssembler* masm); // Generates the slow case code. Must not be naturally // reachable. Expected to be put after a ret instruction (e.g., in // deferred code). Always jumps back to the fast case. void GenerateSlow(MacroAssembler* masm, const RuntimeCallHelper& call_helper); // Skip handling slow case and directly jump to bailout. void SkipSlow(MacroAssembler* masm, Label* bailout) { masm->bind(&slow_case_); masm->jmp(bailout); } private: Register code_; Register result_; Label slow_case_; Label exit_; DISALLOW_COPY_AND_ASSIGN(StringCharFromCodeGenerator); }; // Generates code implementing String.prototype.charAt. // // Only supports the case when the receiver is a string and the index // is a number (smi or heap number) that is a valid index into the // string. Additional index constraints are specified by the // flags. Otherwise, bails out to the provided labels. // // Register usage: |object| may be changed to another string in a way // that doesn't affect charCodeAt/charAt semantics, |index| is // preserved, |scratch1|, |scratch2|, and |result| are clobbered. class StringCharAtGenerator { public: StringCharAtGenerator(Register object, Register index, Register scratch, Register result, Label* receiver_not_string, Label* index_not_number, Label* index_out_of_range, StringIndexFlags index_flags) : char_code_at_generator_(object, index, scratch, receiver_not_string, index_not_number, index_out_of_range, index_flags), char_from_code_generator_(scratch, result) {} // Generates the fast case code. On the fallthrough path |result| // register contains the result. void GenerateFast(MacroAssembler* masm) { char_code_at_generator_.GenerateFast(masm); char_from_code_generator_.GenerateFast(masm); } // Generates the slow case code. Must not be naturally // reachable. Expected to be put after a ret instruction (e.g., in // deferred code). Always jumps back to the fast case. void GenerateSlow(MacroAssembler* masm, const RuntimeCallHelper& call_helper) { char_code_at_generator_.GenerateSlow(masm, call_helper); char_from_code_generator_.GenerateSlow(masm, call_helper); } // Skip handling slow case and directly jump to bailout. void SkipSlow(MacroAssembler* masm, Label* bailout) { char_code_at_generator_.SkipSlow(masm, bailout); char_from_code_generator_.SkipSlow(masm, bailout); } private: StringCharCodeAtGenerator char_code_at_generator_; StringCharFromCodeGenerator char_from_code_generator_; DISALLOW_COPY_AND_ASSIGN(StringCharAtGenerator); }; class AllowStubCallsScope { public: AllowStubCallsScope(MacroAssembler* masm, bool allow) : masm_(masm), previous_allow_(masm->allow_stub_calls()) { masm_->set_allow_stub_calls(allow); } ~AllowStubCallsScope() { masm_->set_allow_stub_calls(previous_allow_); } private: MacroAssembler* masm_; bool previous_allow_; DISALLOW_COPY_AND_ASSIGN(AllowStubCallsScope); }; class KeyedLoadDictionaryElementStub : public PlatformCodeStub { public: KeyedLoadDictionaryElementStub() {} void Generate(MacroAssembler* masm); private: Major MajorKey() { return KeyedLoadElement; } int MinorKey() { return DICTIONARY_ELEMENTS; } DISALLOW_COPY_AND_ASSIGN(KeyedLoadDictionaryElementStub); }; class KeyedLoadFastElementStub : public HydrogenCodeStub { public: KeyedLoadFastElementStub(bool is_js_array, ElementsKind elements_kind) { bit_field_ = ElementsKindBits::encode(elements_kind) | IsJSArrayBits::encode(is_js_array); } bool is_js_array() const { return IsJSArrayBits::decode(bit_field_); } ElementsKind elements_kind() const { return ElementsKindBits::decode(bit_field_); } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: class ElementsKindBits: public BitField {}; class IsJSArrayBits: public BitField {}; uint32_t bit_field_; Major MajorKey() { return KeyedLoadElement; } int NotMissMinorKey() { return bit_field_; } DISALLOW_COPY_AND_ASSIGN(KeyedLoadFastElementStub); }; class KeyedStoreFastElementStub : public HydrogenCodeStub { public: KeyedStoreFastElementStub(bool is_js_array, ElementsKind elements_kind, KeyedAccessStoreMode mode) { bit_field_ = ElementsKindBits::encode(elements_kind) | IsJSArrayBits::encode(is_js_array) | StoreModeBits::encode(mode); } bool is_js_array() const { return IsJSArrayBits::decode(bit_field_); } ElementsKind elements_kind() const { return ElementsKindBits::decode(bit_field_); } KeyedAccessStoreMode store_mode() const { return StoreModeBits::decode(bit_field_); } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: class ElementsKindBits: public BitField {}; class StoreModeBits: public BitField {}; class IsJSArrayBits: public BitField {}; uint32_t bit_field_; Major MajorKey() { return KeyedStoreElement; } int NotMissMinorKey() { return bit_field_; } DISALLOW_COPY_AND_ASSIGN(KeyedStoreFastElementStub); }; class TransitionElementsKindStub : public HydrogenCodeStub { public: TransitionElementsKindStub(ElementsKind from_kind, ElementsKind to_kind) { bit_field_ = FromKindBits::encode(from_kind) | ToKindBits::encode(to_kind); } ElementsKind from_kind() const { return FromKindBits::decode(bit_field_); } ElementsKind to_kind() const { return ToKindBits::decode(bit_field_); } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: class FromKindBits: public BitField {}; class ToKindBits: public BitField {}; uint32_t bit_field_; Major MajorKey() { return TransitionElementsKind; } int NotMissMinorKey() { return bit_field_; } DISALLOW_COPY_AND_ASSIGN(TransitionElementsKindStub); }; class ArrayConstructorStubBase : public HydrogenCodeStub { public: ArrayConstructorStubBase(ElementsKind kind, bool disable_allocation_sites) { // It only makes sense to override local allocation site behavior // if there is a difference between the global allocation site policy // for an ElementsKind and the desired usage of the stub. ASSERT(!disable_allocation_sites || AllocationSiteInfo::GetMode(kind) == TRACK_ALLOCATION_SITE); bit_field_ = ElementsKindBits::encode(kind) | DisableAllocationSitesBits::encode(disable_allocation_sites); } ElementsKind elements_kind() const { return ElementsKindBits::decode(bit_field_); } bool disable_allocation_sites() const { return DisableAllocationSitesBits::decode(bit_field_); } virtual bool IsPregenerated() { return true; } static void GenerateStubsAheadOfTime(Isolate* isolate); static void InstallDescriptors(Isolate* isolate); // Parameters accessed via CodeStubGraphBuilder::GetParameter() static const int kConstructor = 0; static const int kPropertyCell = 1; private: int NotMissMinorKey() { return bit_field_; } class ElementsKindBits: public BitField {}; class DisableAllocationSitesBits: public BitField {}; uint32_t bit_field_; DISALLOW_COPY_AND_ASSIGN(ArrayConstructorStubBase); }; class ArrayNoArgumentConstructorStub : public ArrayConstructorStubBase { public: ArrayNoArgumentConstructorStub( ElementsKind kind, bool disable_allocation_sites = false) : ArrayConstructorStubBase(kind, disable_allocation_sites) { } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: Major MajorKey() { return ArrayNoArgumentConstructor; } DISALLOW_COPY_AND_ASSIGN(ArrayNoArgumentConstructorStub); }; class ArraySingleArgumentConstructorStub : public ArrayConstructorStubBase { public: ArraySingleArgumentConstructorStub( ElementsKind kind, bool disable_allocation_sites = false) : ArrayConstructorStubBase(kind, disable_allocation_sites) { } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: Major MajorKey() { return ArraySingleArgumentConstructor; } DISALLOW_COPY_AND_ASSIGN(ArraySingleArgumentConstructorStub); }; class ArrayNArgumentsConstructorStub : public ArrayConstructorStubBase { public: ArrayNArgumentsConstructorStub( ElementsKind kind, bool disable_allocation_sites = false) : ArrayConstructorStubBase(kind, disable_allocation_sites) { } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: Major MajorKey() { return ArrayNArgumentsConstructor; } DISALLOW_COPY_AND_ASSIGN(ArrayNArgumentsConstructorStub); }; class InternalArrayConstructorStubBase : public HydrogenCodeStub { public: explicit InternalArrayConstructorStubBase(ElementsKind kind) { kind_ = kind; } virtual bool IsPregenerated() { return true; } static void GenerateStubsAheadOfTime(Isolate* isolate); static void InstallDescriptors(Isolate* isolate); // Parameters accessed via CodeStubGraphBuilder::GetParameter() static const int kConstructor = 0; ElementsKind elements_kind() const { return kind_; } private: int NotMissMinorKey() { return kind_; } ElementsKind kind_; DISALLOW_COPY_AND_ASSIGN(InternalArrayConstructorStubBase); }; class InternalArrayNoArgumentConstructorStub : public InternalArrayConstructorStubBase { public: explicit InternalArrayNoArgumentConstructorStub(ElementsKind kind) : InternalArrayConstructorStubBase(kind) { } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: Major MajorKey() { return InternalArrayNoArgumentConstructor; } DISALLOW_COPY_AND_ASSIGN(InternalArrayNoArgumentConstructorStub); }; class InternalArraySingleArgumentConstructorStub : public InternalArrayConstructorStubBase { public: explicit InternalArraySingleArgumentConstructorStub(ElementsKind kind) : InternalArrayConstructorStubBase(kind) { } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: Major MajorKey() { return InternalArraySingleArgumentConstructor; } DISALLOW_COPY_AND_ASSIGN(InternalArraySingleArgumentConstructorStub); }; class InternalArrayNArgumentsConstructorStub : public InternalArrayConstructorStubBase { public: explicit InternalArrayNArgumentsConstructorStub(ElementsKind kind) : InternalArrayConstructorStubBase(kind) { } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); private: Major MajorKey() { return InternalArrayNArgumentsConstructor; } DISALLOW_COPY_AND_ASSIGN(InternalArrayNArgumentsConstructorStub); }; class KeyedStoreElementStub : public PlatformCodeStub { public: KeyedStoreElementStub(bool is_js_array, ElementsKind elements_kind, KeyedAccessStoreMode store_mode) : is_js_array_(is_js_array), elements_kind_(elements_kind), store_mode_(store_mode), fp_registers_(CanUseFPRegisters()) { } Major MajorKey() { return KeyedStoreElement; } int MinorKey() { return ElementsKindBits::encode(elements_kind_) | IsJSArrayBits::encode(is_js_array_) | StoreModeBits::encode(store_mode_) | FPRegisters::encode(fp_registers_); } void Generate(MacroAssembler* masm); private: class ElementsKindBits: public BitField {}; class StoreModeBits: public BitField {}; class IsJSArrayBits: public BitField {}; class FPRegisters: public BitField {}; bool is_js_array_; ElementsKind elements_kind_; KeyedAccessStoreMode store_mode_; bool fp_registers_; DISALLOW_COPY_AND_ASSIGN(KeyedStoreElementStub); }; class ToBooleanStub: public HydrogenCodeStub { public: enum Type { UNDEFINED, BOOLEAN, NULL_TYPE, SMI, SPEC_OBJECT, STRING, SYMBOL, HEAP_NUMBER, NUMBER_OF_TYPES }; // At most 8 different types can be distinguished, because the Code object // only has room for a single byte to hold a set of these types. :-P STATIC_ASSERT(NUMBER_OF_TYPES <= 8); class Types : public EnumSet { public: Types() : EnumSet(0) {} explicit Types(byte bits) : EnumSet(bits) {} byte ToByte() const { return ToIntegral(); } void Print(StringStream* stream) const; void TraceTransition(Types to) const; bool Record(Handle object); bool NeedsMap() const; bool CanBeUndetectable() const; bool IsGeneric() const { return ToIntegral() == Generic().ToIntegral(); } static Types Generic() { return Types((1 << NUMBER_OF_TYPES) - 1); } }; explicit ToBooleanStub(Types types = Types()) : types_(types) { } explicit ToBooleanStub(Code::ExtraICState state) : types_(static_cast(state)) { } bool Record(Handle object); Types GetTypes() { return types_; } virtual Handle GenerateCode(); virtual void InitializeInterfaceDescriptor( Isolate* isolate, CodeStubInterfaceDescriptor* descriptor); virtual Code::Kind GetCodeKind() const { return Code::TO_BOOLEAN_IC; } virtual void PrintName(StringStream* stream); virtual bool SometimesSetsUpAFrame() { return false; } static void InitializeForIsolate(Isolate* isolate) { ToBooleanStub stub; stub.InitializeInterfaceDescriptor( isolate, isolate->code_stub_interface_descriptor(CodeStub::ToBoolean)); } static Handle GetUninitialized(Isolate* isolate) { return ToBooleanStub(UNINITIALIZED).GetCode(isolate); } virtual Code::ExtraICState GetExtraICState() { return types_.ToIntegral(); } virtual InlineCacheState GetICState() { if (types_.IsEmpty()) { return ::v8::internal::UNINITIALIZED; } else { return MONOMORPHIC; } } private: Major MajorKey() { return ToBoolean; } int NotMissMinorKey() { return GetExtraICState(); } explicit ToBooleanStub(InitializationState init_state) : HydrogenCodeStub(init_state) {} Types types_; }; class ElementsTransitionAndStoreStub : public PlatformCodeStub { public: ElementsTransitionAndStoreStub(ElementsKind from, ElementsKind to, bool is_jsarray, StrictModeFlag strict_mode, KeyedAccessStoreMode store_mode) : from_(from), to_(to), is_jsarray_(is_jsarray), strict_mode_(strict_mode), store_mode_(store_mode) {} private: class FromBits: public BitField {}; class ToBits: public BitField {}; class IsJSArrayBits: public BitField {}; class StrictModeBits: public BitField {}; class StoreModeBits: public BitField {}; Major MajorKey() { return ElementsTransitionAndStore; } int MinorKey() { return FromBits::encode(from_) | ToBits::encode(to_) | IsJSArrayBits::encode(is_jsarray_) | StrictModeBits::encode(strict_mode_) | StoreModeBits::encode(store_mode_); } void Generate(MacroAssembler* masm); ElementsKind from_; ElementsKind to_; bool is_jsarray_; StrictModeFlag strict_mode_; KeyedAccessStoreMode store_mode_; DISALLOW_COPY_AND_ASSIGN(ElementsTransitionAndStoreStub); }; class StoreArrayLiteralElementStub : public PlatformCodeStub { public: StoreArrayLiteralElementStub() : fp_registers_(CanUseFPRegisters()) { } private: class FPRegisters: public BitField {}; Major MajorKey() { return StoreArrayLiteralElement; } int MinorKey() { return FPRegisters::encode(fp_registers_); } void Generate(MacroAssembler* masm); bool fp_registers_; DISALLOW_COPY_AND_ASSIGN(StoreArrayLiteralElementStub); }; class StubFailureTrampolineStub : public PlatformCodeStub { public: explicit StubFailureTrampolineStub(StubFunctionMode function_mode) : fp_registers_(CanUseFPRegisters()), function_mode_(function_mode) {} virtual bool IsPregenerated() { return true; } static void GenerateAheadOfTime(Isolate* isolate); private: class FPRegisters: public BitField {}; class FunctionModeField: public BitField {}; Major MajorKey() { return StubFailureTrampoline; } int MinorKey() { return FPRegisters::encode(fp_registers_) | FunctionModeField::encode(function_mode_); } void Generate(MacroAssembler* masm); bool fp_registers_; StubFunctionMode function_mode_; DISALLOW_COPY_AND_ASSIGN(StubFailureTrampolineStub); }; class ProfileEntryHookStub : public PlatformCodeStub { public: explicit ProfileEntryHookStub() {} // The profile entry hook function is not allowed to cause a GC. virtual bool SometimesSetsUpAFrame() { return false; } // Generates a call to the entry hook if it's enabled. static void MaybeCallEntryHook(MacroAssembler* masm); // Sets or unsets the entry hook function. Returns true on success, // false on an attempt to replace a non-NULL entry hook with another // non-NULL hook. static bool SetFunctionEntryHook(FunctionEntryHook entry_hook); static bool HasEntryHook() { return entry_hook_ != NULL; } private: static void EntryHookTrampoline(intptr_t function, intptr_t stack_pointer); Major MajorKey() { return ProfileEntryHook; } int MinorKey() { return 0; } void Generate(MacroAssembler* masm); // The current function entry hook. static FunctionEntryHook entry_hook_; DISALLOW_COPY_AND_ASSIGN(ProfileEntryHookStub); }; } } // namespace v8::internal #endif // V8_CODE_STUBS_H_