| // Copyright 2012 the V8 project authors. All rights reserved. |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are |
| // met: |
| // |
| // * Redistributions of source code must retain the above copyright |
| // notice, this list of conditions and the following disclaimer. |
| // * Redistributions in binary form must reproduce the above |
| // copyright notice, this list of conditions and the following |
| // disclaimer in the documentation and/or other materials provided |
| // with the distribution. |
| // * Neither the name of Google Inc. nor the names of its |
| // contributors may be used to endorse or promote products derived |
| // from this software without specific prior written permission. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| #ifndef V8_CODE_STUBS_H_ |
| #define V8_CODE_STUBS_H_ |
| |
| #include "allocation.h" |
| #include "assembler.h" |
| #include "globals.h" |
| #include "codegen.h" |
| |
| namespace v8 { |
| namespace internal { |
| |
| // List of code stubs used on all platforms. |
| #define CODE_STUB_LIST_ALL_PLATFORMS(V) \ |
| V(CallFunction) \ |
| V(CallConstruct) \ |
| V(BinaryOp) \ |
| V(StringAdd) \ |
| V(SubString) \ |
| V(StringCompare) \ |
| V(Compare) \ |
| V(CompareIC) \ |
| V(CompareNilIC) \ |
| V(MathPow) \ |
| V(StringLength) \ |
| V(FunctionPrototype) \ |
| V(StoreArrayLength) \ |
| V(RecordWrite) \ |
| V(StoreBufferOverflow) \ |
| V(RegExpExec) \ |
| V(TranscendentalCache) \ |
| V(Instanceof) \ |
| V(ConvertToDouble) \ |
| V(WriteInt32ToHeapNumber) \ |
| V(StackCheck) \ |
| V(Interrupt) \ |
| V(FastNewClosure) \ |
| V(FastNewContext) \ |
| V(FastNewBlockContext) \ |
| V(FastCloneShallowArray) \ |
| V(FastCloneShallowObject) \ |
| V(CreateAllocationSite) \ |
| V(ToBoolean) \ |
| V(ToNumber) \ |
| V(ArgumentsAccess) \ |
| V(RegExpConstructResult) \ |
| V(NumberToString) \ |
| V(DoubleToI) \ |
| V(CEntry) \ |
| V(JSEntry) \ |
| V(KeyedLoadElement) \ |
| V(ArrayNoArgumentConstructor) \ |
| V(ArraySingleArgumentConstructor) \ |
| V(ArrayNArgumentsConstructor) \ |
| V(InternalArrayNoArgumentConstructor) \ |
| V(InternalArraySingleArgumentConstructor) \ |
| V(InternalArrayNArgumentsConstructor) \ |
| V(KeyedStoreElement) \ |
| V(DebuggerStatement) \ |
| V(NameDictionaryLookup) \ |
| V(ElementsTransitionAndStore) \ |
| V(TransitionElementsKind) \ |
| V(StoreArrayLiteralElement) \ |
| V(StubFailureTrampoline) \ |
| V(ArrayConstructor) \ |
| V(InternalArrayConstructor) \ |
| V(ProfileEntryHook) \ |
| V(StoreGlobal) \ |
| /* IC Handler stubs */ \ |
| V(LoadField) \ |
| V(KeyedLoadField) |
| |
| // List of code stubs only used on ARM platforms. |
| #if V8_TARGET_ARCH_ARM |
| #define CODE_STUB_LIST_ARM(V) \ |
| V(GetProperty) \ |
| V(SetProperty) \ |
| V(InvokeBuiltin) \ |
| V(DirectCEntry) |
| #else |
| #define CODE_STUB_LIST_ARM(V) |
| #endif |
| |
| // List of code stubs only used on MIPS platforms. |
| #if V8_TARGET_ARCH_MIPS |
| #define CODE_STUB_LIST_MIPS(V) \ |
| V(RegExpCEntry) \ |
| V(DirectCEntry) |
| #else |
| #define CODE_STUB_LIST_MIPS(V) |
| #endif |
| |
| // Combined list of code stubs. |
| #define CODE_STUB_LIST(V) \ |
| CODE_STUB_LIST_ALL_PLATFORMS(V) \ |
| CODE_STUB_LIST_ARM(V) \ |
| CODE_STUB_LIST_MIPS(V) |
| |
| // Mode to overwrite BinaryExpression values. |
| enum OverwriteMode { NO_OVERWRITE, OVERWRITE_LEFT, OVERWRITE_RIGHT }; |
| |
| // Stub is base classes of all stubs. |
| class CodeStub BASE_EMBEDDED { |
| public: |
| enum Major { |
| #define DEF_ENUM(name) name, |
| CODE_STUB_LIST(DEF_ENUM) |
| #undef DEF_ENUM |
| NoCache, // marker for stubs that do custom caching |
| NUMBER_OF_IDS |
| }; |
| |
| // Retrieve the code for the stub. Generate the code if needed. |
| Handle<Code> GetCode(Isolate* isolate); |
| |
| // Retrieve the code for the stub, make and return a copy of the code. |
| Handle<Code> GetCodeCopyFromTemplate(Isolate* isolate); |
| static Major MajorKeyFromKey(uint32_t key) { |
| return static_cast<Major>(MajorKeyBits::decode(key)); |
| } |
| static int MinorKeyFromKey(uint32_t key) { |
| return MinorKeyBits::decode(key); |
| } |
| |
| // Gets the major key from a code object that is a code stub or binary op IC. |
| static Major GetMajorKey(Code* code_stub) { |
| return static_cast<Major>(code_stub->major_key()); |
| } |
| |
| static const char* MajorName(Major major_key, bool allow_unknown_keys); |
| |
| virtual ~CodeStub() {} |
| |
| bool CompilingCallsToThisStubIsGCSafe(Isolate* isolate) { |
| bool is_pregenerated = IsPregenerated(isolate); |
| Code* code = NULL; |
| CHECK(!is_pregenerated || FindCodeInCache(&code, isolate)); |
| return is_pregenerated; |
| } |
| |
| // See comment above, where Instanceof is defined. |
| virtual bool IsPregenerated(Isolate* isolate) { return false; } |
| |
| static void GenerateStubsAheadOfTime(Isolate* isolate); |
| static void GenerateFPStubs(Isolate* isolate); |
| |
| // Some stubs put untagged junk on the stack that cannot be scanned by the |
| // GC. This means that we must be statically sure that no GC can occur while |
| // they are running. If that is the case they should override this to return |
| // true, which will cause an assertion if we try to call something that can |
| // GC or if we try to put a stack frame on top of the junk, which would not |
| // result in a traversable stack. |
| virtual bool SometimesSetsUpAFrame() { return true; } |
| |
| // Lookup the code in the (possibly custom) cache. |
| bool FindCodeInCache(Code** code_out, Isolate* isolate); |
| |
| // Returns information for computing the number key. |
| virtual Major MajorKey() = 0; |
| virtual int MinorKey() = 0; |
| |
| virtual InlineCacheState GetICState() { |
| return UNINITIALIZED; |
| } |
| virtual Code::ExtraICState GetExtraICState() { |
| return Code::kNoExtraICState; |
| } |
| virtual Code::StubType GetStubType() { |
| return Code::NORMAL; |
| } |
| virtual int GetStubFlags() { |
| return -1; |
| } |
| |
| virtual void PrintName(StringStream* stream); |
| |
| protected: |
| static bool CanUseFPRegisters(); |
| |
| // Generates the assembler code for the stub. |
| virtual Handle<Code> GenerateCode(Isolate* isolate) = 0; |
| |
| |
| // Returns whether the code generated for this stub needs to be allocated as |
| // a fixed (non-moveable) code object. |
| virtual bool NeedsImmovableCode() { return false; } |
| |
| // Returns a name for logging/debugging purposes. |
| SmartArrayPointer<const char> GetName(); |
| virtual void PrintBaseName(StringStream* stream); |
| virtual void PrintState(StringStream* stream) { } |
| |
| private: |
| // Perform bookkeeping required after code generation when stub code is |
| // initially generated. |
| void RecordCodeGeneration(Code* code, Isolate* isolate); |
| |
| // Finish the code object after it has been generated. |
| virtual void FinishCode(Handle<Code> code) { } |
| |
| // Activate newly generated stub. Is called after |
| // registering stub in the stub cache. |
| virtual void Activate(Code* code) { } |
| |
| // BinaryOpStub needs to override this. |
| virtual Code::Kind GetCodeKind() const; |
| |
| // Add the code to a specialized cache, specific to an individual |
| // stub type. Please note, this method must add the code object to a |
| // roots object, otherwise we will remove the code during GC. |
| virtual void AddToSpecialCache(Handle<Code> new_object) { } |
| |
| // Find code in a specialized cache, work is delegated to the specific stub. |
| virtual bool FindCodeInSpecialCache(Code** code_out, Isolate* isolate) { |
| return false; |
| } |
| |
| // If a stub uses a special cache override this. |
| virtual bool UseSpecialCache() { return false; } |
| |
| // Computes the key based on major and minor. |
| uint32_t GetKey() { |
| ASSERT(static_cast<int>(MajorKey()) < NUMBER_OF_IDS); |
| return MinorKeyBits::encode(MinorKey()) | |
| MajorKeyBits::encode(MajorKey()); |
| } |
| |
| class MajorKeyBits: public BitField<uint32_t, 0, kStubMajorKeyBits> {}; |
| class MinorKeyBits: public BitField<uint32_t, |
| kStubMajorKeyBits, kStubMinorKeyBits> {}; // NOLINT |
| |
| friend class BreakPointIterator; |
| }; |
| |
| |
| class PlatformCodeStub : public CodeStub { |
| public: |
| // Retrieve the code for the stub. Generate the code if needed. |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual Code::Kind GetCodeKind() const { return Code::STUB; } |
| |
| protected: |
| // Generates the assembler code for the stub. |
| virtual void Generate(MacroAssembler* masm) = 0; |
| }; |
| |
| |
| enum StubFunctionMode { NOT_JS_FUNCTION_STUB_MODE, JS_FUNCTION_STUB_MODE }; |
| |
| |
| struct CodeStubInterfaceDescriptor { |
| CodeStubInterfaceDescriptor(); |
| int register_param_count_; |
| const Register* stack_parameter_count_; |
| // if hint_stack_parameter_count_ > 0, the code stub can optimize the |
| // return sequence. Default value is -1, which means it is ignored. |
| int hint_stack_parameter_count_; |
| StubFunctionMode function_mode_; |
| Register* register_params_; |
| Address deoptimization_handler_; |
| |
| int environment_length() const { |
| if (stack_parameter_count_ != NULL) { |
| return register_param_count_ + 1; |
| } |
| return register_param_count_; |
| } |
| |
| bool initialized() const { return register_param_count_ >= 0; } |
| |
| void SetMissHandler(ExternalReference handler) { |
| miss_handler_ = handler; |
| has_miss_handler_ = true; |
| } |
| |
| ExternalReference miss_handler() { |
| ASSERT(has_miss_handler_); |
| return miss_handler_; |
| } |
| |
| bool has_miss_handler() { |
| return has_miss_handler_; |
| } |
| |
| private: |
| ExternalReference miss_handler_; |
| bool has_miss_handler_; |
| }; |
| |
| // A helper to make up for the fact that type Register is not fully |
| // defined outside of the platform directories |
| #define DESCRIPTOR_GET_PARAMETER_REGISTER(descriptor, index) \ |
| ((index) == (descriptor)->register_param_count_) \ |
| ? *((descriptor)->stack_parameter_count_) \ |
| : (descriptor)->register_params_[(index)] |
| |
| |
| class HydrogenCodeStub : public CodeStub { |
| public: |
| enum InitializationState { |
| UNINITIALIZED, |
| INITIALIZED |
| }; |
| |
| explicit HydrogenCodeStub(InitializationState state = INITIALIZED) { |
| is_uninitialized_ = (state == UNINITIALIZED); |
| } |
| |
| virtual Code::Kind GetCodeKind() const { return Code::STUB; } |
| |
| CodeStubInterfaceDescriptor* GetInterfaceDescriptor(Isolate* isolate) { |
| return isolate->code_stub_interface_descriptor(MajorKey()); |
| } |
| |
| bool IsUninitialized() { return is_uninitialized_; } |
| |
| template<class SubClass> |
| static Handle<Code> GetUninitialized(Isolate* isolate) { |
| SubClass::GenerateAheadOfTime(isolate); |
| return SubClass().GetCode(isolate); |
| } |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor) = 0; |
| |
| // Retrieve the code for the stub. Generate the code if needed. |
| virtual Handle<Code> GenerateCode(Isolate* isolate) = 0; |
| |
| virtual int NotMissMinorKey() = 0; |
| |
| Handle<Code> GenerateLightweightMissCode(Isolate* isolate); |
| |
| template<class StateType> |
| void TraceTransition(StateType from, StateType to); |
| |
| private: |
| class MinorKeyBits: public BitField<int, 0, kStubMinorKeyBits - 1> {}; |
| class IsMissBits: public BitField<bool, kStubMinorKeyBits - 1, 1> {}; |
| |
| void GenerateLightweightMiss(MacroAssembler* masm); |
| virtual int MinorKey() { |
| return IsMissBits::encode(is_uninitialized_) | |
| MinorKeyBits::encode(NotMissMinorKey()); |
| } |
| |
| bool is_uninitialized_; |
| }; |
| |
| |
| // Helper interface to prepare to/restore after making runtime calls. |
| class RuntimeCallHelper { |
| public: |
| virtual ~RuntimeCallHelper() {} |
| |
| virtual void BeforeCall(MacroAssembler* masm) const = 0; |
| |
| virtual void AfterCall(MacroAssembler* masm) const = 0; |
| |
| protected: |
| RuntimeCallHelper() {} |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(RuntimeCallHelper); |
| }; |
| |
| |
| // TODO(bmeurer): Move to the StringAddStub declaration once we're |
| // done with the translation to a hydrogen code stub. |
| enum StringAddFlags { |
| // Omit both parameter checks. |
| STRING_ADD_CHECK_NONE = 0, |
| // Check left parameter. |
| STRING_ADD_CHECK_LEFT = 1 << 0, |
| // Check right parameter. |
| STRING_ADD_CHECK_RIGHT = 1 << 1, |
| // Check both parameters. |
| STRING_ADD_CHECK_BOTH = STRING_ADD_CHECK_LEFT | STRING_ADD_CHECK_RIGHT, |
| // Stub needs a frame before calling the runtime |
| STRING_ADD_ERECT_FRAME = 1 << 2 |
| }; |
| |
| } } // namespace v8::internal |
| |
| #if V8_TARGET_ARCH_IA32 |
| #include "ia32/code-stubs-ia32.h" |
| #elif V8_TARGET_ARCH_X64 |
| #include "x64/code-stubs-x64.h" |
| #elif V8_TARGET_ARCH_ARM |
| #include "arm/code-stubs-arm.h" |
| #elif V8_TARGET_ARCH_MIPS |
| #include "mips/code-stubs-mips.h" |
| #else |
| #error Unsupported target architecture. |
| #endif |
| |
| namespace v8 { |
| namespace internal { |
| |
| |
| // RuntimeCallHelper implementation used in stubs: enters/leaves a |
| // newly created internal frame before/after the runtime call. |
| class StubRuntimeCallHelper : public RuntimeCallHelper { |
| public: |
| StubRuntimeCallHelper() {} |
| |
| virtual void BeforeCall(MacroAssembler* masm) const; |
| |
| virtual void AfterCall(MacroAssembler* masm) const; |
| }; |
| |
| |
| // Trivial RuntimeCallHelper implementation. |
| class NopRuntimeCallHelper : public RuntimeCallHelper { |
| public: |
| NopRuntimeCallHelper() {} |
| |
| virtual void BeforeCall(MacroAssembler* masm) const {} |
| |
| virtual void AfterCall(MacroAssembler* masm) const {} |
| }; |
| |
| |
| class ToNumberStub: public HydrogenCodeStub { |
| public: |
| ToNumberStub() { } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Major MajorKey() { return ToNumber; } |
| int NotMissMinorKey() { return 0; } |
| }; |
| |
| |
| class FastNewClosureStub : public HydrogenCodeStub { |
| public: |
| explicit FastNewClosureStub(LanguageMode language_mode, bool is_generator) |
| : language_mode_(language_mode), |
| is_generator_(is_generator) { } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| static void InstallDescriptors(Isolate* isolate); |
| |
| LanguageMode language_mode() const { return language_mode_; } |
| bool is_generator() const { return is_generator_; } |
| |
| private: |
| class StrictModeBits: public BitField<bool, 0, 1> {}; |
| class IsGeneratorBits: public BitField<bool, 1, 1> {}; |
| |
| Major MajorKey() { return FastNewClosure; } |
| int NotMissMinorKey() { |
| return StrictModeBits::encode(language_mode_ != CLASSIC_MODE) | |
| IsGeneratorBits::encode(is_generator_); |
| } |
| |
| LanguageMode language_mode_; |
| bool is_generator_; |
| }; |
| |
| |
| class FastNewContextStub : public PlatformCodeStub { |
| public: |
| static const int kMaximumSlots = 64; |
| |
| explicit FastNewContextStub(int slots) : slots_(slots) { |
| ASSERT(slots_ > 0 && slots_ <= kMaximumSlots); |
| } |
| |
| void Generate(MacroAssembler* masm); |
| |
| private: |
| int slots_; |
| |
| Major MajorKey() { return FastNewContext; } |
| int MinorKey() { return slots_; } |
| }; |
| |
| |
| class FastNewBlockContextStub : public PlatformCodeStub { |
| public: |
| static const int kMaximumSlots = 64; |
| |
| explicit FastNewBlockContextStub(int slots) : slots_(slots) { |
| ASSERT(slots_ > 0 && slots_ <= kMaximumSlots); |
| } |
| |
| void Generate(MacroAssembler* masm); |
| |
| private: |
| int slots_; |
| |
| Major MajorKey() { return FastNewBlockContext; } |
| int MinorKey() { return slots_; } |
| }; |
| |
| class StoreGlobalStub : public HydrogenCodeStub { |
| public: |
| StoreGlobalStub(StrictModeFlag strict_mode, bool is_constant) { |
| bit_field_ = StrictModeBits::encode(strict_mode) | |
| IsConstantBits::encode(is_constant); |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| virtual Code::Kind GetCodeKind() const { return Code::STORE_IC; } |
| virtual InlineCacheState GetICState() { return MONOMORPHIC; } |
| virtual Code::ExtraICState GetExtraICState() { return bit_field_; } |
| |
| bool is_constant() { |
| return IsConstantBits::decode(bit_field_); |
| } |
| void set_is_constant(bool value) { |
| bit_field_ = IsConstantBits::update(bit_field_, value); |
| } |
| |
| Representation representation() { |
| return Representation::FromKind(RepresentationBits::decode(bit_field_)); |
| } |
| void set_representation(Representation r) { |
| bit_field_ = RepresentationBits::update(bit_field_, r.kind()); |
| } |
| |
| private: |
| virtual int NotMissMinorKey() { return GetExtraICState(); } |
| Major MajorKey() { return StoreGlobal; } |
| |
| class StrictModeBits: public BitField<StrictModeFlag, 0, 1> {}; |
| class IsConstantBits: public BitField<bool, 1, 1> {}; |
| class RepresentationBits: public BitField<Representation::Kind, 2, 8> {}; |
| |
| int bit_field_; |
| |
| DISALLOW_COPY_AND_ASSIGN(StoreGlobalStub); |
| }; |
| |
| |
| class FastCloneShallowArrayStub : public HydrogenCodeStub { |
| public: |
| // Maximum length of copied elements array. |
| static const int kMaximumClonedLength = 8; |
| enum Mode { |
| CLONE_ELEMENTS, |
| CLONE_DOUBLE_ELEMENTS, |
| COPY_ON_WRITE_ELEMENTS, |
| CLONE_ANY_ELEMENTS, |
| LAST_CLONE_MODE = CLONE_ANY_ELEMENTS |
| }; |
| |
| static const int kFastCloneModeCount = LAST_CLONE_MODE + 1; |
| |
| FastCloneShallowArrayStub(Mode mode, |
| AllocationSiteMode allocation_site_mode, |
| int length) |
| : mode_(mode), |
| allocation_site_mode_(allocation_site_mode), |
| length_((mode == COPY_ON_WRITE_ELEMENTS) ? 0 : length) { |
| ASSERT_GE(length_, 0); |
| ASSERT_LE(length_, kMaximumClonedLength); |
| } |
| |
| Mode mode() const { return mode_; } |
| int length() const { return length_; } |
| AllocationSiteMode allocation_site_mode() const { |
| return allocation_site_mode_; |
| } |
| |
| ElementsKind ComputeElementsKind() const { |
| switch (mode()) { |
| case CLONE_ELEMENTS: |
| case COPY_ON_WRITE_ELEMENTS: |
| return FAST_ELEMENTS; |
| case CLONE_DOUBLE_ELEMENTS: |
| return FAST_DOUBLE_ELEMENTS; |
| case CLONE_ANY_ELEMENTS: |
| /*fall-through*/; |
| } |
| UNREACHABLE(); |
| return LAST_ELEMENTS_KIND; |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Mode mode_; |
| AllocationSiteMode allocation_site_mode_; |
| int length_; |
| |
| class AllocationSiteModeBits: public BitField<AllocationSiteMode, 0, 1> {}; |
| class ModeBits: public BitField<Mode, 1, 4> {}; |
| class LengthBits: public BitField<int, 5, 4> {}; |
| // Ensure data fits within available bits. |
| STATIC_ASSERT(LAST_ALLOCATION_SITE_MODE == 1); |
| STATIC_ASSERT(kFastCloneModeCount < 16); |
| STATIC_ASSERT(kMaximumClonedLength < 16); |
| Major MajorKey() { return FastCloneShallowArray; } |
| int NotMissMinorKey() { |
| return AllocationSiteModeBits::encode(allocation_site_mode_) |
| | ModeBits::encode(mode_) |
| | LengthBits::encode(length_); |
| } |
| }; |
| |
| |
| class FastCloneShallowObjectStub : public HydrogenCodeStub { |
| public: |
| // Maximum number of properties in copied object. |
| static const int kMaximumClonedProperties = 6; |
| |
| explicit FastCloneShallowObjectStub(int length) |
| : length_(length) { |
| ASSERT_GE(length_, 0); |
| ASSERT_LE(length_, kMaximumClonedProperties); |
| } |
| |
| int length() const { return length_; } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| int length_; |
| |
| Major MajorKey() { return FastCloneShallowObject; } |
| int NotMissMinorKey() { return length_; } |
| |
| DISALLOW_COPY_AND_ASSIGN(FastCloneShallowObjectStub); |
| }; |
| |
| |
| class CreateAllocationSiteStub : public HydrogenCodeStub { |
| public: |
| explicit CreateAllocationSiteStub() { } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual bool IsPregenerated(Isolate* isolate) V8_OVERRIDE { return true; } |
| |
| static void GenerateAheadOfTime(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Major MajorKey() { return CreateAllocationSite; } |
| int NotMissMinorKey() { return 0; } |
| |
| DISALLOW_COPY_AND_ASSIGN(CreateAllocationSiteStub); |
| }; |
| |
| |
| class InstanceofStub: public PlatformCodeStub { |
| public: |
| enum Flags { |
| kNoFlags = 0, |
| kArgsInRegisters = 1 << 0, |
| kCallSiteInlineCheck = 1 << 1, |
| kReturnTrueFalseObject = 1 << 2 |
| }; |
| |
| explicit InstanceofStub(Flags flags) : flags_(flags) { } |
| |
| static Register left(); |
| static Register right(); |
| |
| void Generate(MacroAssembler* masm); |
| |
| private: |
| Major MajorKey() { return Instanceof; } |
| int MinorKey() { return static_cast<int>(flags_); } |
| |
| bool HasArgsInRegisters() const { |
| return (flags_ & kArgsInRegisters) != 0; |
| } |
| |
| bool HasCallSiteInlineCheck() const { |
| return (flags_ & kCallSiteInlineCheck) != 0; |
| } |
| |
| bool ReturnTrueFalseObject() const { |
| return (flags_ & kReturnTrueFalseObject) != 0; |
| } |
| |
| virtual void PrintName(StringStream* stream); |
| |
| Flags flags_; |
| }; |
| |
| |
| enum AllocationSiteOverrideMode { |
| DONT_OVERRIDE, |
| DISABLE_ALLOCATION_SITES, |
| LAST_ALLOCATION_SITE_OVERRIDE_MODE = DISABLE_ALLOCATION_SITES |
| }; |
| |
| |
| class ArrayConstructorStub: public PlatformCodeStub { |
| public: |
| enum ArgumentCountKey { ANY, NONE, ONE, MORE_THAN_ONE }; |
| ArrayConstructorStub(Isolate* isolate, int argument_count); |
| explicit ArrayConstructorStub(Isolate* isolate); |
| |
| void Generate(MacroAssembler* masm); |
| |
| private: |
| void GenerateDispatchToArrayStub(MacroAssembler* masm, |
| AllocationSiteOverrideMode mode); |
| |
| virtual CodeStub::Major MajorKey() { return ArrayConstructor; } |
| virtual int MinorKey() { return argument_count_; } |
| |
| ArgumentCountKey argument_count_; |
| }; |
| |
| |
| class InternalArrayConstructorStub: public PlatformCodeStub { |
| public: |
| explicit InternalArrayConstructorStub(Isolate* isolate); |
| |
| void Generate(MacroAssembler* masm); |
| |
| private: |
| virtual CodeStub::Major MajorKey() { return InternalArrayConstructor; } |
| virtual int MinorKey() { return 0; } |
| |
| void GenerateCase(MacroAssembler* masm, ElementsKind kind); |
| }; |
| |
| |
| class MathPowStub: public PlatformCodeStub { |
| public: |
| enum ExponentType { INTEGER, DOUBLE, TAGGED, ON_STACK }; |
| |
| explicit MathPowStub(ExponentType exponent_type) |
| : exponent_type_(exponent_type) { } |
| virtual void Generate(MacroAssembler* masm); |
| |
| private: |
| virtual CodeStub::Major MajorKey() { return MathPow; } |
| virtual int MinorKey() { return exponent_type_; } |
| |
| ExponentType exponent_type_; |
| }; |
| |
| |
| class ICStub: public PlatformCodeStub { |
| public: |
| explicit ICStub(Code::Kind kind) : kind_(kind) { } |
| virtual Code::Kind GetCodeKind() const { return kind_; } |
| virtual InlineCacheState GetICState() { return MONOMORPHIC; } |
| |
| bool Describes(Code* code) { |
| return GetMajorKey(code) == MajorKey() && code->stub_info() == MinorKey(); |
| } |
| |
| protected: |
| class KindBits: public BitField<Code::Kind, 0, 4> {}; |
| virtual void FinishCode(Handle<Code> code) { |
| code->set_stub_info(MinorKey()); |
| } |
| Code::Kind kind() { return kind_; } |
| |
| virtual int MinorKey() { |
| return KindBits::encode(kind_); |
| } |
| |
| private: |
| Code::Kind kind_; |
| }; |
| |
| |
| class FunctionPrototypeStub: public ICStub { |
| public: |
| explicit FunctionPrototypeStub(Code::Kind kind) : ICStub(kind) { } |
| virtual void Generate(MacroAssembler* masm); |
| |
| private: |
| virtual CodeStub::Major MajorKey() { return FunctionPrototype; } |
| }; |
| |
| |
| class StringLengthStub: public ICStub { |
| public: |
| explicit StringLengthStub(Code::Kind kind) : ICStub(kind) { } |
| virtual void Generate(MacroAssembler* masm); |
| |
| private: |
| STATIC_ASSERT(KindBits::kSize == 4); |
| virtual CodeStub::Major MajorKey() { return StringLength; } |
| }; |
| |
| |
| class StoreICStub: public ICStub { |
| public: |
| StoreICStub(Code::Kind kind, StrictModeFlag strict_mode) |
| : ICStub(kind), strict_mode_(strict_mode) { } |
| |
| protected: |
| virtual Code::ExtraICState GetExtraICState() { |
| return strict_mode_; |
| } |
| |
| private: |
| STATIC_ASSERT(KindBits::kSize == 4); |
| class StrictModeBits: public BitField<bool, 4, 1> {}; |
| virtual int MinorKey() { |
| return KindBits::encode(kind()) | StrictModeBits::encode(strict_mode_); |
| } |
| |
| StrictModeFlag strict_mode_; |
| }; |
| |
| |
| class StoreArrayLengthStub: public StoreICStub { |
| public: |
| explicit StoreArrayLengthStub(Code::Kind kind, StrictModeFlag strict_mode) |
| : StoreICStub(kind, strict_mode) { } |
| virtual void Generate(MacroAssembler* masm); |
| |
| private: |
| virtual CodeStub::Major MajorKey() { return StoreArrayLength; } |
| }; |
| |
| |
| class HICStub: public HydrogenCodeStub { |
| public: |
| virtual Code::Kind GetCodeKind() const { return kind(); } |
| virtual InlineCacheState GetICState() { return MONOMORPHIC; } |
| |
| protected: |
| HICStub() { } |
| class KindBits: public BitField<Code::Kind, 0, 4> {}; |
| virtual Code::Kind kind() const = 0; |
| }; |
| |
| |
| class HandlerStub: public HICStub { |
| public: |
| virtual Code::Kind GetCodeKind() const { return Code::STUB; } |
| virtual int GetStubFlags() { return kind(); } |
| |
| protected: |
| HandlerStub() : HICStub() { } |
| }; |
| |
| |
| class LoadFieldStub: public HandlerStub { |
| public: |
| LoadFieldStub(bool inobject, int index, Representation representation) |
| : HandlerStub() { |
| Initialize(Code::LOAD_IC, inobject, index, representation); |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| Representation representation() { |
| if (unboxed_double()) return Representation::Double(); |
| return Representation::Tagged(); |
| } |
| |
| virtual Code::Kind kind() const { |
| return KindBits::decode(bit_field_); |
| } |
| |
| bool is_inobject() { |
| return InobjectBits::decode(bit_field_); |
| } |
| |
| int offset() { |
| int index = IndexBits::decode(bit_field_); |
| int offset = index * kPointerSize; |
| if (is_inobject()) return offset; |
| return FixedArray::kHeaderSize + offset; |
| } |
| |
| bool unboxed_double() { |
| return UnboxedDoubleBits::decode(bit_field_); |
| } |
| |
| virtual Code::StubType GetStubType() { return Code::FIELD; } |
| |
| protected: |
| LoadFieldStub() : HandlerStub() { } |
| |
| void Initialize(Code::Kind kind, |
| bool inobject, |
| int index, |
| Representation representation) { |
| bool unboxed_double = FLAG_track_double_fields && representation.IsDouble(); |
| bit_field_ = KindBits::encode(kind) |
| | InobjectBits::encode(inobject) |
| | IndexBits::encode(index) |
| | UnboxedDoubleBits::encode(unboxed_double); |
| } |
| |
| private: |
| STATIC_ASSERT(KindBits::kSize == 4); |
| class InobjectBits: public BitField<bool, 4, 1> {}; |
| class IndexBits: public BitField<int, 5, 11> {}; |
| class UnboxedDoubleBits: public BitField<bool, 16, 1> {}; |
| virtual CodeStub::Major MajorKey() { return LoadField; } |
| virtual int NotMissMinorKey() { return bit_field_; } |
| |
| int bit_field_; |
| }; |
| |
| |
| class KeyedLoadFieldStub: public LoadFieldStub { |
| public: |
| KeyedLoadFieldStub(bool inobject, int index, Representation representation) |
| : LoadFieldStub() { |
| Initialize(Code::KEYED_LOAD_IC, inobject, index, representation); |
| } |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| private: |
| virtual CodeStub::Major MajorKey() { return KeyedLoadField; } |
| }; |
| |
| |
| class BinaryOpStub: public PlatformCodeStub { |
| public: |
| BinaryOpStub(Token::Value op, OverwriteMode mode) |
| : op_(op), |
| mode_(mode), |
| platform_specific_bit_(false), |
| left_type_(BinaryOpIC::UNINITIALIZED), |
| right_type_(BinaryOpIC::UNINITIALIZED), |
| result_type_(BinaryOpIC::UNINITIALIZED), |
| encoded_right_arg_(false, encode_arg_value(1)) { |
| Initialize(); |
| ASSERT(OpBits::is_valid(Token::NUM_TOKENS)); |
| } |
| |
| BinaryOpStub( |
| int key, |
| BinaryOpIC::TypeInfo left_type, |
| BinaryOpIC::TypeInfo right_type, |
| BinaryOpIC::TypeInfo result_type, |
| Maybe<int32_t> fixed_right_arg) |
| : op_(OpBits::decode(key)), |
| mode_(ModeBits::decode(key)), |
| platform_specific_bit_(PlatformSpecificBits::decode(key)), |
| left_type_(left_type), |
| right_type_(right_type), |
| result_type_(result_type), |
| encoded_right_arg_(fixed_right_arg.has_value, |
| encode_arg_value(fixed_right_arg.value)) { } |
| |
| static void decode_types_from_minor_key(int minor_key, |
| BinaryOpIC::TypeInfo* left_type, |
| BinaryOpIC::TypeInfo* right_type, |
| BinaryOpIC::TypeInfo* result_type) { |
| *left_type = |
| static_cast<BinaryOpIC::TypeInfo>(LeftTypeBits::decode(minor_key)); |
| *right_type = |
| static_cast<BinaryOpIC::TypeInfo>(RightTypeBits::decode(minor_key)); |
| *result_type = |
| static_cast<BinaryOpIC::TypeInfo>(ResultTypeBits::decode(minor_key)); |
| } |
| |
| static Token::Value decode_op_from_minor_key(int minor_key) { |
| return static_cast<Token::Value>(OpBits::decode(minor_key)); |
| } |
| |
| static Maybe<int> decode_fixed_right_arg_from_minor_key(int minor_key) { |
| return Maybe<int>( |
| HasFixedRightArgBits::decode(minor_key), |
| decode_arg_value(FixedRightArgValueBits::decode(minor_key))); |
| } |
| |
| int fixed_right_arg_value() const { |
| return decode_arg_value(encoded_right_arg_.value); |
| } |
| |
| static bool can_encode_arg_value(int32_t value) { |
| return value > 0 && |
| IsPowerOf2(value) && |
| FixedRightArgValueBits::is_valid(WhichPowerOf2(value)); |
| } |
| |
| enum SmiCodeGenerateHeapNumberResults { |
| ALLOW_HEAPNUMBER_RESULTS, |
| NO_HEAPNUMBER_RESULTS |
| }; |
| |
| private: |
| Token::Value op_; |
| OverwriteMode mode_; |
| bool platform_specific_bit_; // Indicates SSE3 on IA32. |
| |
| // Operand type information determined at runtime. |
| BinaryOpIC::TypeInfo left_type_; |
| BinaryOpIC::TypeInfo right_type_; |
| BinaryOpIC::TypeInfo result_type_; |
| |
| Maybe<int> encoded_right_arg_; |
| |
| static int encode_arg_value(int32_t value) { |
| ASSERT(can_encode_arg_value(value)); |
| return WhichPowerOf2(value); |
| } |
| |
| static int32_t decode_arg_value(int value) { |
| return 1 << value; |
| } |
| |
| virtual void PrintName(StringStream* stream); |
| |
| // Minor key encoding in all 25 bits FFFFFHTTTRRRLLLPOOOOOOOMM. |
| // Note: We actually do not need 7 bits for the operation, just 4 bits to |
| // encode ADD, SUB, MUL, DIV, MOD, BIT_OR, BIT_AND, BIT_XOR, SAR, SHL, SHR. |
| class ModeBits: public BitField<OverwriteMode, 0, 2> {}; |
| class OpBits: public BitField<Token::Value, 2, 7> {}; |
| class PlatformSpecificBits: public BitField<bool, 9, 1> {}; |
| class LeftTypeBits: public BitField<BinaryOpIC::TypeInfo, 10, 3> {}; |
| class RightTypeBits: public BitField<BinaryOpIC::TypeInfo, 13, 3> {}; |
| class ResultTypeBits: public BitField<BinaryOpIC::TypeInfo, 16, 3> {}; |
| class HasFixedRightArgBits: public BitField<bool, 19, 1> {}; |
| class FixedRightArgValueBits: public BitField<int, 20, 5> {}; |
| |
| Major MajorKey() { return BinaryOp; } |
| int MinorKey() { |
| return OpBits::encode(op_) |
| | ModeBits::encode(mode_) |
| | PlatformSpecificBits::encode(platform_specific_bit_) |
| | LeftTypeBits::encode(left_type_) |
| | RightTypeBits::encode(right_type_) |
| | ResultTypeBits::encode(result_type_) |
| | HasFixedRightArgBits::encode(encoded_right_arg_.has_value) |
| | FixedRightArgValueBits::encode(encoded_right_arg_.value); |
| } |
| |
| |
| // Platform-independent implementation. |
| void Generate(MacroAssembler* masm); |
| void GenerateCallRuntime(MacroAssembler* masm); |
| |
| // Platform-independent signature, platform-specific implementation. |
| void Initialize(); |
| void GenerateAddStrings(MacroAssembler* masm); |
| void GenerateBothStringStub(MacroAssembler* masm); |
| void GenerateGeneric(MacroAssembler* masm); |
| void GenerateGenericStub(MacroAssembler* masm); |
| void GenerateNumberStub(MacroAssembler* masm); |
| void GenerateInt32Stub(MacroAssembler* masm); |
| void GenerateLoadArguments(MacroAssembler* masm); |
| void GenerateOddballStub(MacroAssembler* masm); |
| void GenerateRegisterArgsPush(MacroAssembler* masm); |
| void GenerateReturn(MacroAssembler* masm); |
| void GenerateSmiStub(MacroAssembler* masm); |
| void GenerateStringStub(MacroAssembler* masm); |
| void GenerateTypeTransition(MacroAssembler* masm); |
| void GenerateTypeTransitionWithSavedArgs(MacroAssembler* masm); |
| void GenerateUninitializedStub(MacroAssembler* masm); |
| |
| // Entirely platform-specific methods are defined as static helper |
| // functions in the <arch>/code-stubs-<arch>.cc files. |
| |
| virtual Code::Kind GetCodeKind() const { return Code::BINARY_OP_IC; } |
| |
| virtual InlineCacheState GetICState() { |
| return BinaryOpIC::ToState(Max(left_type_, right_type_)); |
| } |
| |
| virtual void FinishCode(Handle<Code> code) { |
| code->set_stub_info(MinorKey()); |
| } |
| |
| friend class CodeGenerator; |
| }; |
| |
| |
| class ICCompareStub: public PlatformCodeStub { |
| public: |
| ICCompareStub(Token::Value op, |
| CompareIC::State left, |
| CompareIC::State right, |
| CompareIC::State handler) |
| : op_(op), |
| left_(left), |
| right_(right), |
| state_(handler) { |
| ASSERT(Token::IsCompareOp(op)); |
| } |
| |
| virtual void Generate(MacroAssembler* masm); |
| |
| void set_known_map(Handle<Map> map) { known_map_ = map; } |
| |
| static void DecodeMinorKey(int minor_key, |
| CompareIC::State* left_state, |
| CompareIC::State* right_state, |
| CompareIC::State* handler_state, |
| Token::Value* op); |
| |
| static CompareIC::State CompareState(int minor_key) { |
| return static_cast<CompareIC::State>(HandlerStateField::decode(minor_key)); |
| } |
| |
| virtual InlineCacheState GetICState(); |
| |
| private: |
| class OpField: public BitField<int, 0, 3> { }; |
| class LeftStateField: public BitField<int, 3, 4> { }; |
| class RightStateField: public BitField<int, 7, 4> { }; |
| class HandlerStateField: public BitField<int, 11, 4> { }; |
| |
| virtual void FinishCode(Handle<Code> code) { |
| code->set_stub_info(MinorKey()); |
| } |
| |
| virtual CodeStub::Major MajorKey() { return CompareIC; } |
| virtual int MinorKey(); |
| |
| virtual Code::Kind GetCodeKind() const { return Code::COMPARE_IC; } |
| |
| void GenerateSmis(MacroAssembler* masm); |
| void GenerateNumbers(MacroAssembler* masm); |
| void GenerateInternalizedStrings(MacroAssembler* masm); |
| void GenerateStrings(MacroAssembler* masm); |
| void GenerateUniqueNames(MacroAssembler* masm); |
| void GenerateObjects(MacroAssembler* masm); |
| void GenerateMiss(MacroAssembler* masm); |
| void GenerateKnownObjects(MacroAssembler* masm); |
| void GenerateGeneric(MacroAssembler* masm); |
| |
| bool strict() const { return op_ == Token::EQ_STRICT; } |
| Condition GetCondition() const { return CompareIC::ComputeCondition(op_); } |
| |
| virtual void AddToSpecialCache(Handle<Code> new_object); |
| virtual bool FindCodeInSpecialCache(Code** code_out, Isolate* isolate); |
| virtual bool UseSpecialCache() { return state_ == CompareIC::KNOWN_OBJECT; } |
| |
| Token::Value op_; |
| CompareIC::State left_; |
| CompareIC::State right_; |
| CompareIC::State state_; |
| Handle<Map> known_map_; |
| }; |
| |
| |
| class CompareNilICStub : public HydrogenCodeStub { |
| public: |
| Handle<Type> GetType(Isolate* isolate, Handle<Map> map = Handle<Map>()); |
| Handle<Type> GetInputType(Isolate* isolate, Handle<Map> map); |
| |
| explicit CompareNilICStub(NilValue nil) : nil_value_(nil) { } |
| |
| CompareNilICStub(Code::ExtraICState ic_state, |
| InitializationState init_state = INITIALIZED) |
| : HydrogenCodeStub(init_state), |
| nil_value_(NilValueField::decode(ic_state)), |
| state_(State(TypesField::decode(ic_state))) { |
| } |
| |
| static Handle<Code> GetUninitialized(Isolate* isolate, |
| NilValue nil) { |
| return CompareNilICStub(nil, UNINITIALIZED).GetCode(isolate); |
| } |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| static void InitializeForIsolate(Isolate* isolate) { |
| CompareNilICStub compare_stub(kNullValue, UNINITIALIZED); |
| compare_stub.InitializeInterfaceDescriptor( |
| isolate, |
| isolate->code_stub_interface_descriptor(CodeStub::CompareNilIC)); |
| } |
| |
| virtual InlineCacheState GetICState() { |
| if (state_.Contains(GENERIC)) { |
| return MEGAMORPHIC; |
| } else if (state_.Contains(MONOMORPHIC_MAP)) { |
| return MONOMORPHIC; |
| } else { |
| return PREMONOMORPHIC; |
| } |
| } |
| |
| virtual Code::Kind GetCodeKind() const { return Code::COMPARE_NIL_IC; } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual Code::ExtraICState GetExtraICState() { |
| return NilValueField::encode(nil_value_) | |
| TypesField::encode(state_.ToIntegral()); |
| } |
| |
| void UpdateStatus(Handle<Object> object); |
| |
| bool IsMonomorphic() const { return state_.Contains(MONOMORPHIC_MAP); } |
| NilValue GetNilValue() const { return nil_value_; } |
| void ClearState() { state_.RemoveAll(); } |
| |
| virtual void PrintState(StringStream* stream); |
| virtual void PrintBaseName(StringStream* stream); |
| |
| private: |
| friend class CompareNilIC; |
| |
| enum CompareNilType { |
| UNDEFINED, |
| NULL_TYPE, |
| MONOMORPHIC_MAP, |
| GENERIC, |
| NUMBER_OF_TYPES |
| }; |
| |
| // At most 6 different types can be distinguished, because the Code object |
| // only has room for a single byte to hold a set and there are two more |
| // boolean flags we need to store. :-P |
| STATIC_ASSERT(NUMBER_OF_TYPES <= 6); |
| |
| class State : public EnumSet<CompareNilType, byte> { |
| public: |
| State() : EnumSet<CompareNilType, byte>(0) { } |
| explicit State(byte bits) : EnumSet<CompareNilType, byte>(bits) { } |
| |
| void Print(StringStream* stream) const; |
| }; |
| |
| CompareNilICStub(NilValue nil, InitializationState init_state) |
| : HydrogenCodeStub(init_state), nil_value_(nil) { } |
| |
| class NilValueField : public BitField<NilValue, 0, 1> {}; |
| class TypesField : public BitField<byte, 1, NUMBER_OF_TYPES> {}; |
| |
| virtual CodeStub::Major MajorKey() { return CompareNilIC; } |
| virtual int NotMissMinorKey() { return GetExtraICState(); } |
| |
| NilValue nil_value_; |
| State state_; |
| |
| DISALLOW_COPY_AND_ASSIGN(CompareNilICStub); |
| }; |
| |
| |
| class CEntryStub : public PlatformCodeStub { |
| public: |
| explicit CEntryStub(int result_size, |
| SaveFPRegsMode save_doubles = kDontSaveFPRegs) |
| : result_size_(result_size), save_doubles_(save_doubles) { } |
| |
| void Generate(MacroAssembler* masm); |
| |
| // The version of this stub that doesn't save doubles is generated ahead of |
| // time, so it's OK to call it from other stubs that can't cope with GC during |
| // their code generation. On machines that always have gp registers (x64) we |
| // can generate both variants ahead of time. |
| virtual bool IsPregenerated(Isolate* isolate) V8_OVERRIDE; |
| static void GenerateAheadOfTime(Isolate* isolate); |
| |
| private: |
| void GenerateCore(MacroAssembler* masm, |
| Label* throw_normal_exception, |
| Label* throw_termination_exception, |
| Label* throw_out_of_memory_exception, |
| bool do_gc, |
| bool always_allocate_scope); |
| |
| // Number of pointers/values returned. |
| Isolate* isolate_; |
| const int result_size_; |
| SaveFPRegsMode save_doubles_; |
| |
| Major MajorKey() { return CEntry; } |
| int MinorKey(); |
| |
| bool NeedsImmovableCode(); |
| }; |
| |
| |
| class JSEntryStub : public PlatformCodeStub { |
| public: |
| JSEntryStub() { } |
| |
| void Generate(MacroAssembler* masm) { GenerateBody(masm, false); } |
| |
| protected: |
| void GenerateBody(MacroAssembler* masm, bool is_construct); |
| |
| private: |
| Major MajorKey() { return JSEntry; } |
| int MinorKey() { return 0; } |
| |
| virtual void FinishCode(Handle<Code> code); |
| |
| int handler_offset_; |
| }; |
| |
| |
| class JSConstructEntryStub : public JSEntryStub { |
| public: |
| JSConstructEntryStub() { } |
| |
| void Generate(MacroAssembler* masm) { GenerateBody(masm, true); } |
| |
| private: |
| int MinorKey() { return 1; } |
| |
| virtual void PrintName(StringStream* stream) { |
| stream->Add("JSConstructEntryStub"); |
| } |
| }; |
| |
| |
| class ArgumentsAccessStub: public PlatformCodeStub { |
| public: |
| enum Type { |
| READ_ELEMENT, |
| NEW_NON_STRICT_FAST, |
| NEW_NON_STRICT_SLOW, |
| NEW_STRICT |
| }; |
| |
| explicit ArgumentsAccessStub(Type type) : type_(type) { } |
| |
| private: |
| Type type_; |
| |
| Major MajorKey() { return ArgumentsAccess; } |
| int MinorKey() { return type_; } |
| |
| void Generate(MacroAssembler* masm); |
| void GenerateReadElement(MacroAssembler* masm); |
| void GenerateNewStrict(MacroAssembler* masm); |
| void GenerateNewNonStrictFast(MacroAssembler* masm); |
| void GenerateNewNonStrictSlow(MacroAssembler* masm); |
| |
| virtual void PrintName(StringStream* stream); |
| }; |
| |
| |
| class RegExpExecStub: public PlatformCodeStub { |
| public: |
| RegExpExecStub() { } |
| |
| private: |
| Major MajorKey() { return RegExpExec; } |
| int MinorKey() { return 0; } |
| |
| void Generate(MacroAssembler* masm); |
| }; |
| |
| |
| class RegExpConstructResultStub: public PlatformCodeStub { |
| public: |
| RegExpConstructResultStub() { } |
| |
| private: |
| Major MajorKey() { return RegExpConstructResult; } |
| int MinorKey() { return 0; } |
| |
| void Generate(MacroAssembler* masm); |
| }; |
| |
| |
| class CallFunctionStub: public PlatformCodeStub { |
| public: |
| CallFunctionStub(int argc, CallFunctionFlags flags) |
| : argc_(argc), flags_(flags) { } |
| |
| void Generate(MacroAssembler* masm); |
| |
| virtual void FinishCode(Handle<Code> code) { |
| code->set_has_function_cache(RecordCallTarget()); |
| } |
| |
| static int ExtractArgcFromMinorKey(int minor_key) { |
| return ArgcBits::decode(minor_key); |
| } |
| |
| private: |
| int argc_; |
| CallFunctionFlags flags_; |
| |
| virtual void PrintName(StringStream* stream); |
| |
| // Minor key encoding in 32 bits with Bitfield <Type, shift, size>. |
| class FlagBits: public BitField<CallFunctionFlags, 0, 2> {}; |
| class ArgcBits: public BitField<unsigned, 2, 32 - 2> {}; |
| |
| Major MajorKey() { return CallFunction; } |
| int MinorKey() { |
| // Encode the parameters in a unique 32 bit value. |
| return FlagBits::encode(flags_) | ArgcBits::encode(argc_); |
| } |
| |
| bool ReceiverMightBeImplicit() { |
| return (flags_ & RECEIVER_MIGHT_BE_IMPLICIT) != 0; |
| } |
| |
| bool RecordCallTarget() { |
| return (flags_ & RECORD_CALL_TARGET) != 0; |
| } |
| }; |
| |
| |
| class CallConstructStub: public PlatformCodeStub { |
| public: |
| explicit CallConstructStub(CallFunctionFlags flags) : flags_(flags) {} |
| |
| void Generate(MacroAssembler* masm); |
| |
| virtual void FinishCode(Handle<Code> code) { |
| code->set_has_function_cache(RecordCallTarget()); |
| } |
| |
| private: |
| CallFunctionFlags flags_; |
| |
| virtual void PrintName(StringStream* stream); |
| |
| Major MajorKey() { return CallConstruct; } |
| int MinorKey() { return flags_; } |
| |
| bool RecordCallTarget() { |
| return (flags_ & RECORD_CALL_TARGET) != 0; |
| } |
| }; |
| |
| |
| enum StringIndexFlags { |
| // Accepts smis or heap numbers. |
| STRING_INDEX_IS_NUMBER, |
| |
| // Accepts smis or heap numbers that are valid array indices |
| // (ECMA-262 15.4). Invalid indices are reported as being out of |
| // range. |
| STRING_INDEX_IS_ARRAY_INDEX |
| }; |
| |
| |
| // Generates code implementing String.prototype.charCodeAt. |
| // |
| // Only supports the case when the receiver is a string and the index |
| // is a number (smi or heap number) that is a valid index into the |
| // string. Additional index constraints are specified by the |
| // flags. Otherwise, bails out to the provided labels. |
| // |
| // Register usage: |object| may be changed to another string in a way |
| // that doesn't affect charCodeAt/charAt semantics, |index| is |
| // preserved, |scratch| and |result| are clobbered. |
| class StringCharCodeAtGenerator { |
| public: |
| StringCharCodeAtGenerator(Register object, |
| Register index, |
| Register result, |
| Label* receiver_not_string, |
| Label* index_not_number, |
| Label* index_out_of_range, |
| StringIndexFlags index_flags) |
| : object_(object), |
| index_(index), |
| result_(result), |
| receiver_not_string_(receiver_not_string), |
| index_not_number_(index_not_number), |
| index_out_of_range_(index_out_of_range), |
| index_flags_(index_flags) { |
| ASSERT(!result_.is(object_)); |
| ASSERT(!result_.is(index_)); |
| } |
| |
| // Generates the fast case code. On the fallthrough path |result| |
| // register contains the result. |
| void GenerateFast(MacroAssembler* masm); |
| |
| // Generates the slow case code. Must not be naturally |
| // reachable. Expected to be put after a ret instruction (e.g., in |
| // deferred code). Always jumps back to the fast case. |
| void GenerateSlow(MacroAssembler* masm, |
| const RuntimeCallHelper& call_helper); |
| |
| // Skip handling slow case and directly jump to bailout. |
| void SkipSlow(MacroAssembler* masm, Label* bailout) { |
| masm->bind(&index_not_smi_); |
| masm->bind(&call_runtime_); |
| masm->jmp(bailout); |
| } |
| |
| private: |
| Register object_; |
| Register index_; |
| Register result_; |
| |
| Label* receiver_not_string_; |
| Label* index_not_number_; |
| Label* index_out_of_range_; |
| |
| StringIndexFlags index_flags_; |
| |
| Label call_runtime_; |
| Label index_not_smi_; |
| Label got_smi_index_; |
| Label exit_; |
| |
| DISALLOW_COPY_AND_ASSIGN(StringCharCodeAtGenerator); |
| }; |
| |
| |
| // Generates code for creating a one-char string from a char code. |
| class StringCharFromCodeGenerator { |
| public: |
| StringCharFromCodeGenerator(Register code, |
| Register result) |
| : code_(code), |
| result_(result) { |
| ASSERT(!code_.is(result_)); |
| } |
| |
| // Generates the fast case code. On the fallthrough path |result| |
| // register contains the result. |
| void GenerateFast(MacroAssembler* masm); |
| |
| // Generates the slow case code. Must not be naturally |
| // reachable. Expected to be put after a ret instruction (e.g., in |
| // deferred code). Always jumps back to the fast case. |
| void GenerateSlow(MacroAssembler* masm, |
| const RuntimeCallHelper& call_helper); |
| |
| // Skip handling slow case and directly jump to bailout. |
| void SkipSlow(MacroAssembler* masm, Label* bailout) { |
| masm->bind(&slow_case_); |
| masm->jmp(bailout); |
| } |
| |
| private: |
| Register code_; |
| Register result_; |
| |
| Label slow_case_; |
| Label exit_; |
| |
| DISALLOW_COPY_AND_ASSIGN(StringCharFromCodeGenerator); |
| }; |
| |
| |
| // Generates code implementing String.prototype.charAt. |
| // |
| // Only supports the case when the receiver is a string and the index |
| // is a number (smi or heap number) that is a valid index into the |
| // string. Additional index constraints are specified by the |
| // flags. Otherwise, bails out to the provided labels. |
| // |
| // Register usage: |object| may be changed to another string in a way |
| // that doesn't affect charCodeAt/charAt semantics, |index| is |
| // preserved, |scratch1|, |scratch2|, and |result| are clobbered. |
| class StringCharAtGenerator { |
| public: |
| StringCharAtGenerator(Register object, |
| Register index, |
| Register scratch, |
| Register result, |
| Label* receiver_not_string, |
| Label* index_not_number, |
| Label* index_out_of_range, |
| StringIndexFlags index_flags) |
| : char_code_at_generator_(object, |
| index, |
| scratch, |
| receiver_not_string, |
| index_not_number, |
| index_out_of_range, |
| index_flags), |
| char_from_code_generator_(scratch, result) {} |
| |
| // Generates the fast case code. On the fallthrough path |result| |
| // register contains the result. |
| void GenerateFast(MacroAssembler* masm) { |
| char_code_at_generator_.GenerateFast(masm); |
| char_from_code_generator_.GenerateFast(masm); |
| } |
| |
| // Generates the slow case code. Must not be naturally |
| // reachable. Expected to be put after a ret instruction (e.g., in |
| // deferred code). Always jumps back to the fast case. |
| void GenerateSlow(MacroAssembler* masm, |
| const RuntimeCallHelper& call_helper) { |
| char_code_at_generator_.GenerateSlow(masm, call_helper); |
| char_from_code_generator_.GenerateSlow(masm, call_helper); |
| } |
| |
| // Skip handling slow case and directly jump to bailout. |
| void SkipSlow(MacroAssembler* masm, Label* bailout) { |
| char_code_at_generator_.SkipSlow(masm, bailout); |
| char_from_code_generator_.SkipSlow(masm, bailout); |
| } |
| |
| private: |
| StringCharCodeAtGenerator char_code_at_generator_; |
| StringCharFromCodeGenerator char_from_code_generator_; |
| |
| DISALLOW_COPY_AND_ASSIGN(StringCharAtGenerator); |
| }; |
| |
| |
| class AllowStubCallsScope { |
| public: |
| AllowStubCallsScope(MacroAssembler* masm, bool allow) |
| : masm_(masm), previous_allow_(masm->allow_stub_calls()) { |
| masm_->set_allow_stub_calls(allow); |
| } |
| ~AllowStubCallsScope() { |
| masm_->set_allow_stub_calls(previous_allow_); |
| } |
| |
| private: |
| MacroAssembler* masm_; |
| bool previous_allow_; |
| |
| DISALLOW_COPY_AND_ASSIGN(AllowStubCallsScope); |
| }; |
| |
| |
| class KeyedLoadDictionaryElementStub : public PlatformCodeStub { |
| public: |
| KeyedLoadDictionaryElementStub() {} |
| |
| void Generate(MacroAssembler* masm); |
| |
| private: |
| Major MajorKey() { return KeyedLoadElement; } |
| int MinorKey() { return DICTIONARY_ELEMENTS; } |
| |
| DISALLOW_COPY_AND_ASSIGN(KeyedLoadDictionaryElementStub); |
| }; |
| |
| |
| class DoubleToIStub : public PlatformCodeStub { |
| public: |
| DoubleToIStub(Register source, |
| Register destination, |
| int offset, |
| bool is_truncating, |
| bool skip_fastpath = false) : bit_field_(0) { |
| bit_field_ = SourceRegisterBits::encode(source.code_) | |
| DestinationRegisterBits::encode(destination.code_) | |
| OffsetBits::encode(offset) | |
| IsTruncatingBits::encode(is_truncating) | |
| SkipFastPathBits::encode(skip_fastpath); |
| } |
| |
| Register source() { |
| Register result = { SourceRegisterBits::decode(bit_field_) }; |
| return result; |
| } |
| |
| Register destination() { |
| Register result = { DestinationRegisterBits::decode(bit_field_) }; |
| return result; |
| } |
| |
| bool is_truncating() { |
| return IsTruncatingBits::decode(bit_field_); |
| } |
| |
| bool skip_fastpath() { |
| return SkipFastPathBits::decode(bit_field_); |
| } |
| |
| int offset() { |
| return OffsetBits::decode(bit_field_); |
| } |
| |
| void Generate(MacroAssembler* masm); |
| |
| virtual bool SometimesSetsUpAFrame() { return false; } |
| |
| private: |
| static const int kBitsPerRegisterNumber = 6; |
| STATIC_ASSERT((1L << kBitsPerRegisterNumber) >= Register::kNumRegisters); |
| class SourceRegisterBits: |
| public BitField<int, 0, kBitsPerRegisterNumber> {}; // NOLINT |
| class DestinationRegisterBits: |
| public BitField<int, kBitsPerRegisterNumber, |
| kBitsPerRegisterNumber> {}; // NOLINT |
| class IsTruncatingBits: |
| public BitField<bool, 2 * kBitsPerRegisterNumber, 1> {}; // NOLINT |
| class OffsetBits: |
| public BitField<int, 2 * kBitsPerRegisterNumber + 1, 3> {}; // NOLINT |
| class SkipFastPathBits: |
| public BitField<int, 2 * kBitsPerRegisterNumber + 4, 1> {}; // NOLINT |
| |
| Major MajorKey() { return DoubleToI; } |
| int MinorKey() { return bit_field_; } |
| |
| int bit_field_; |
| |
| DISALLOW_COPY_AND_ASSIGN(DoubleToIStub); |
| }; |
| |
| |
| class KeyedLoadFastElementStub : public HydrogenCodeStub { |
| public: |
| KeyedLoadFastElementStub(bool is_js_array, ElementsKind elements_kind) { |
| bit_field_ = ElementsKindBits::encode(elements_kind) | |
| IsJSArrayBits::encode(is_js_array); |
| } |
| |
| bool is_js_array() const { |
| return IsJSArrayBits::decode(bit_field_); |
| } |
| |
| ElementsKind elements_kind() const { |
| return ElementsKindBits::decode(bit_field_); |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| class ElementsKindBits: public BitField<ElementsKind, 0, 8> {}; |
| class IsJSArrayBits: public BitField<bool, 8, 1> {}; |
| uint32_t bit_field_; |
| |
| Major MajorKey() { return KeyedLoadElement; } |
| int NotMissMinorKey() { return bit_field_; } |
| |
| DISALLOW_COPY_AND_ASSIGN(KeyedLoadFastElementStub); |
| }; |
| |
| |
| class KeyedStoreFastElementStub : public HydrogenCodeStub { |
| public: |
| KeyedStoreFastElementStub(bool is_js_array, |
| ElementsKind elements_kind, |
| KeyedAccessStoreMode mode) { |
| bit_field_ = ElementsKindBits::encode(elements_kind) | |
| IsJSArrayBits::encode(is_js_array) | |
| StoreModeBits::encode(mode); |
| } |
| |
| bool is_js_array() const { |
| return IsJSArrayBits::decode(bit_field_); |
| } |
| |
| ElementsKind elements_kind() const { |
| return ElementsKindBits::decode(bit_field_); |
| } |
| |
| KeyedAccessStoreMode store_mode() const { |
| return StoreModeBits::decode(bit_field_); |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| class ElementsKindBits: public BitField<ElementsKind, 0, 8> {}; |
| class StoreModeBits: public BitField<KeyedAccessStoreMode, 8, 4> {}; |
| class IsJSArrayBits: public BitField<bool, 12, 1> {}; |
| uint32_t bit_field_; |
| |
| Major MajorKey() { return KeyedStoreElement; } |
| int NotMissMinorKey() { return bit_field_; } |
| |
| DISALLOW_COPY_AND_ASSIGN(KeyedStoreFastElementStub); |
| }; |
| |
| |
| class TransitionElementsKindStub : public HydrogenCodeStub { |
| public: |
| TransitionElementsKindStub(ElementsKind from_kind, |
| ElementsKind to_kind) { |
| bit_field_ = FromKindBits::encode(from_kind) | |
| ToKindBits::encode(to_kind); |
| } |
| |
| ElementsKind from_kind() const { |
| return FromKindBits::decode(bit_field_); |
| } |
| |
| ElementsKind to_kind() const { |
| return ToKindBits::decode(bit_field_); |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| class FromKindBits: public BitField<ElementsKind, 8, 8> {}; |
| class ToKindBits: public BitField<ElementsKind, 0, 8> {}; |
| uint32_t bit_field_; |
| |
| Major MajorKey() { return TransitionElementsKind; } |
| int NotMissMinorKey() { return bit_field_; } |
| |
| DISALLOW_COPY_AND_ASSIGN(TransitionElementsKindStub); |
| }; |
| |
| |
| enum ContextCheckMode { |
| CONTEXT_CHECK_REQUIRED, |
| CONTEXT_CHECK_NOT_REQUIRED, |
| LAST_CONTEXT_CHECK_MODE = CONTEXT_CHECK_NOT_REQUIRED |
| }; |
| |
| |
| class ArrayConstructorStubBase : public HydrogenCodeStub { |
| public: |
| ArrayConstructorStubBase(ElementsKind kind, ContextCheckMode context_mode, |
| AllocationSiteOverrideMode override_mode) { |
| // It only makes sense to override local allocation site behavior |
| // if there is a difference between the global allocation site policy |
| // for an ElementsKind and the desired usage of the stub. |
| ASSERT(!(FLAG_track_allocation_sites && |
| override_mode == DISABLE_ALLOCATION_SITES) || |
| AllocationSite::GetMode(kind) == TRACK_ALLOCATION_SITE); |
| bit_field_ = ElementsKindBits::encode(kind) | |
| AllocationSiteOverrideModeBits::encode(override_mode) | |
| ContextCheckModeBits::encode(context_mode); |
| } |
| |
| ElementsKind elements_kind() const { |
| return ElementsKindBits::decode(bit_field_); |
| } |
| |
| AllocationSiteOverrideMode override_mode() const { |
| return AllocationSiteOverrideModeBits::decode(bit_field_); |
| } |
| |
| ContextCheckMode context_mode() const { |
| return ContextCheckModeBits::decode(bit_field_); |
| } |
| |
| virtual bool IsPregenerated(Isolate* isolate) V8_OVERRIDE { |
| // We only pre-generate stubs that verify correct context |
| return context_mode() == CONTEXT_CHECK_REQUIRED; |
| } |
| |
| static void GenerateStubsAheadOfTime(Isolate* isolate); |
| static void InstallDescriptors(Isolate* isolate); |
| |
| // Parameters accessed via CodeStubGraphBuilder::GetParameter() |
| static const int kConstructor = 0; |
| static const int kPropertyCell = 1; |
| |
| private: |
| int NotMissMinorKey() { return bit_field_; } |
| |
| // Ensure data fits within available bits. |
| STATIC_ASSERT(LAST_ALLOCATION_SITE_OVERRIDE_MODE == 1); |
| STATIC_ASSERT(LAST_CONTEXT_CHECK_MODE == 1); |
| |
| class ElementsKindBits: public BitField<ElementsKind, 0, 8> {}; |
| class AllocationSiteOverrideModeBits: public |
| BitField<AllocationSiteOverrideMode, 8, 1> {}; // NOLINT |
| class ContextCheckModeBits: public BitField<ContextCheckMode, 9, 1> {}; |
| uint32_t bit_field_; |
| |
| DISALLOW_COPY_AND_ASSIGN(ArrayConstructorStubBase); |
| }; |
| |
| |
| class ArrayNoArgumentConstructorStub : public ArrayConstructorStubBase { |
| public: |
| ArrayNoArgumentConstructorStub( |
| ElementsKind kind, |
| ContextCheckMode context_mode = CONTEXT_CHECK_REQUIRED, |
| AllocationSiteOverrideMode override_mode = DONT_OVERRIDE) |
| : ArrayConstructorStubBase(kind, context_mode, override_mode) { |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Major MajorKey() { return ArrayNoArgumentConstructor; } |
| |
| DISALLOW_COPY_AND_ASSIGN(ArrayNoArgumentConstructorStub); |
| }; |
| |
| |
| class ArraySingleArgumentConstructorStub : public ArrayConstructorStubBase { |
| public: |
| ArraySingleArgumentConstructorStub( |
| ElementsKind kind, |
| ContextCheckMode context_mode = CONTEXT_CHECK_REQUIRED, |
| AllocationSiteOverrideMode override_mode = DONT_OVERRIDE) |
| : ArrayConstructorStubBase(kind, context_mode, override_mode) { |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Major MajorKey() { return ArraySingleArgumentConstructor; } |
| |
| DISALLOW_COPY_AND_ASSIGN(ArraySingleArgumentConstructorStub); |
| }; |
| |
| |
| class ArrayNArgumentsConstructorStub : public ArrayConstructorStubBase { |
| public: |
| ArrayNArgumentsConstructorStub( |
| ElementsKind kind, |
| ContextCheckMode context_mode = CONTEXT_CHECK_REQUIRED, |
| AllocationSiteOverrideMode override_mode = DONT_OVERRIDE) |
| : ArrayConstructorStubBase(kind, context_mode, override_mode) { |
| } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Major MajorKey() { return ArrayNArgumentsConstructor; } |
| |
| DISALLOW_COPY_AND_ASSIGN(ArrayNArgumentsConstructorStub); |
| }; |
| |
| |
| class InternalArrayConstructorStubBase : public HydrogenCodeStub { |
| public: |
| explicit InternalArrayConstructorStubBase(ElementsKind kind) { |
| kind_ = kind; |
| } |
| |
| virtual bool IsPregenerated(Isolate* isolate) V8_OVERRIDE { return true; } |
| static void GenerateStubsAheadOfTime(Isolate* isolate); |
| static void InstallDescriptors(Isolate* isolate); |
| |
| // Parameters accessed via CodeStubGraphBuilder::GetParameter() |
| static const int kConstructor = 0; |
| |
| ElementsKind elements_kind() const { return kind_; } |
| |
| private: |
| int NotMissMinorKey() { return kind_; } |
| |
| ElementsKind kind_; |
| |
| DISALLOW_COPY_AND_ASSIGN(InternalArrayConstructorStubBase); |
| }; |
| |
| |
| class InternalArrayNoArgumentConstructorStub : public |
| InternalArrayConstructorStubBase { |
| public: |
| explicit InternalArrayNoArgumentConstructorStub(ElementsKind kind) |
| : InternalArrayConstructorStubBase(kind) { } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Major MajorKey() { return InternalArrayNoArgumentConstructor; } |
| |
| DISALLOW_COPY_AND_ASSIGN(InternalArrayNoArgumentConstructorStub); |
| }; |
| |
| |
| class InternalArraySingleArgumentConstructorStub : public |
| InternalArrayConstructorStubBase { |
| public: |
| explicit InternalArraySingleArgumentConstructorStub(ElementsKind kind) |
| : InternalArrayConstructorStubBase(kind) { } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Major MajorKey() { return InternalArraySingleArgumentConstructor; } |
| |
| DISALLOW_COPY_AND_ASSIGN(InternalArraySingleArgumentConstructorStub); |
| }; |
| |
| |
| class InternalArrayNArgumentsConstructorStub : public |
| InternalArrayConstructorStubBase { |
| public: |
| explicit InternalArrayNArgumentsConstructorStub(ElementsKind kind) |
| : InternalArrayConstructorStubBase(kind) { } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| Major MajorKey() { return InternalArrayNArgumentsConstructor; } |
| |
| DISALLOW_COPY_AND_ASSIGN(InternalArrayNArgumentsConstructorStub); |
| }; |
| |
| |
| class KeyedStoreElementStub : public PlatformCodeStub { |
| public: |
| KeyedStoreElementStub(bool is_js_array, |
| ElementsKind elements_kind, |
| KeyedAccessStoreMode store_mode) |
| : is_js_array_(is_js_array), |
| elements_kind_(elements_kind), |
| store_mode_(store_mode), |
| fp_registers_(CanUseFPRegisters()) { } |
| |
| Major MajorKey() { return KeyedStoreElement; } |
| int MinorKey() { |
| return ElementsKindBits::encode(elements_kind_) | |
| IsJSArrayBits::encode(is_js_array_) | |
| StoreModeBits::encode(store_mode_) | |
| FPRegisters::encode(fp_registers_); |
| } |
| |
| void Generate(MacroAssembler* masm); |
| |
| private: |
| class ElementsKindBits: public BitField<ElementsKind, 0, 8> {}; |
| class StoreModeBits: public BitField<KeyedAccessStoreMode, 8, 4> {}; |
| class IsJSArrayBits: public BitField<bool, 12, 1> {}; |
| class FPRegisters: public BitField<bool, 13, 1> {}; |
| |
| bool is_js_array_; |
| ElementsKind elements_kind_; |
| KeyedAccessStoreMode store_mode_; |
| bool fp_registers_; |
| |
| DISALLOW_COPY_AND_ASSIGN(KeyedStoreElementStub); |
| }; |
| |
| |
| class ToBooleanStub: public HydrogenCodeStub { |
| public: |
| enum Type { |
| UNDEFINED, |
| BOOLEAN, |
| NULL_TYPE, |
| SMI, |
| SPEC_OBJECT, |
| STRING, |
| SYMBOL, |
| HEAP_NUMBER, |
| NUMBER_OF_TYPES |
| }; |
| |
| // At most 8 different types can be distinguished, because the Code object |
| // only has room for a single byte to hold a set of these types. :-P |
| STATIC_ASSERT(NUMBER_OF_TYPES <= 8); |
| |
| class Types : public EnumSet<Type, byte> { |
| public: |
| Types() : EnumSet<Type, byte>(0) {} |
| explicit Types(byte bits) : EnumSet<Type, byte>(bits) {} |
| |
| byte ToByte() const { return ToIntegral(); } |
| void Print(StringStream* stream) const; |
| bool UpdateStatus(Handle<Object> object); |
| bool NeedsMap() const; |
| bool CanBeUndetectable() const; |
| bool IsGeneric() const { return ToIntegral() == Generic().ToIntegral(); } |
| |
| static Types Generic() { return Types((1 << NUMBER_OF_TYPES) - 1); } |
| }; |
| |
| explicit ToBooleanStub(Types types = Types()) |
| : types_(types) { } |
| explicit ToBooleanStub(Code::ExtraICState state) |
| : types_(static_cast<byte>(state)) { } |
| |
| bool UpdateStatus(Handle<Object> object); |
| Types GetTypes() { return types_; } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| virtual void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| virtual Code::Kind GetCodeKind() const { return Code::TO_BOOLEAN_IC; } |
| virtual void PrintState(StringStream* stream); |
| |
| virtual bool SometimesSetsUpAFrame() { return false; } |
| |
| static void InitializeForIsolate(Isolate* isolate) { |
| ToBooleanStub stub; |
| stub.InitializeInterfaceDescriptor( |
| isolate, |
| isolate->code_stub_interface_descriptor(CodeStub::ToBoolean)); |
| } |
| |
| static Handle<Code> GetUninitialized(Isolate* isolate) { |
| return ToBooleanStub(UNINITIALIZED).GetCode(isolate); |
| } |
| |
| virtual Code::ExtraICState GetExtraICState() { |
| return types_.ToIntegral(); |
| } |
| |
| virtual InlineCacheState GetICState() { |
| if (types_.IsEmpty()) { |
| return ::v8::internal::UNINITIALIZED; |
| } else { |
| return MONOMORPHIC; |
| } |
| } |
| |
| private: |
| Major MajorKey() { return ToBoolean; } |
| int NotMissMinorKey() { return GetExtraICState(); } |
| |
| explicit ToBooleanStub(InitializationState init_state) : |
| HydrogenCodeStub(init_state) {} |
| |
| Types types_; |
| }; |
| |
| |
| class ElementsTransitionAndStoreStub : public HydrogenCodeStub { |
| public: |
| ElementsTransitionAndStoreStub(ElementsKind from_kind, |
| ElementsKind to_kind, |
| bool is_jsarray, |
| KeyedAccessStoreMode store_mode) |
| : from_kind_(from_kind), |
| to_kind_(to_kind), |
| is_jsarray_(is_jsarray), |
| store_mode_(store_mode) {} |
| |
| ElementsKind from_kind() const { return from_kind_; } |
| ElementsKind to_kind() const { return to_kind_; } |
| bool is_jsarray() const { return is_jsarray_; } |
| KeyedAccessStoreMode store_mode() const { return store_mode_; } |
| |
| virtual Handle<Code> GenerateCode(Isolate* isolate); |
| |
| void InitializeInterfaceDescriptor( |
| Isolate* isolate, |
| CodeStubInterfaceDescriptor* descriptor); |
| |
| private: |
| class FromBits: public BitField<ElementsKind, 0, 8> {}; |
| class ToBits: public BitField<ElementsKind, 8, 8> {}; |
| class IsJSArrayBits: public BitField<bool, 16, 1> {}; |
| class StoreModeBits: public BitField<KeyedAccessStoreMode, 17, 4> {}; |
| |
| Major MajorKey() { return ElementsTransitionAndStore; } |
| int NotMissMinorKey() { |
| return FromBits::encode(from_kind_) | |
| ToBits::encode(to_kind_) | |
| IsJSArrayBits::encode(is_jsarray_) | |
| StoreModeBits::encode(store_mode_); |
| } |
| |
| ElementsKind from_kind_; |
| ElementsKind to_kind_; |
| bool is_jsarray_; |
| KeyedAccessStoreMode store_mode_; |
| |
| DISALLOW_COPY_AND_ASSIGN(ElementsTransitionAndStoreStub); |
| }; |
| |
| |
| class StoreArrayLiteralElementStub : public PlatformCodeStub { |
| public: |
| StoreArrayLiteralElementStub() |
| : fp_registers_(CanUseFPRegisters()) { } |
| |
| private: |
| class FPRegisters: public BitField<bool, 0, 1> {}; |
| |
| Major MajorKey() { return StoreArrayLiteralElement; } |
| int MinorKey() { return FPRegisters::encode(fp_registers_); } |
| |
| void Generate(MacroAssembler* masm); |
| |
| bool fp_registers_; |
| |
| DISALLOW_COPY_AND_ASSIGN(StoreArrayLiteralElementStub); |
| }; |
| |
| |
| class StubFailureTrampolineStub : public PlatformCodeStub { |
| public: |
| explicit StubFailureTrampolineStub(StubFunctionMode function_mode) |
| : fp_registers_(CanUseFPRegisters()), function_mode_(function_mode) {} |
| |
| virtual bool IsPregenerated(Isolate* isolate) V8_OVERRIDE { return true; } |
| |
| static void GenerateAheadOfTime(Isolate* isolate); |
| |
| private: |
| class FPRegisters: public BitField<bool, 0, 1> {}; |
| class FunctionModeField: public BitField<StubFunctionMode, 1, 1> {}; |
| |
| Major MajorKey() { return StubFailureTrampoline; } |
| int MinorKey() { |
| return FPRegisters::encode(fp_registers_) | |
| FunctionModeField::encode(function_mode_); |
| } |
| |
| void Generate(MacroAssembler* masm); |
| |
| bool fp_registers_; |
| StubFunctionMode function_mode_; |
| |
| DISALLOW_COPY_AND_ASSIGN(StubFailureTrampolineStub); |
| }; |
| |
| |
| class ProfileEntryHookStub : public PlatformCodeStub { |
| public: |
| explicit ProfileEntryHookStub() {} |
| |
| // The profile entry hook function is not allowed to cause a GC. |
| virtual bool SometimesSetsUpAFrame() { return false; } |
| |
| // Generates a call to the entry hook if it's enabled. |
| static void MaybeCallEntryHook(MacroAssembler* masm); |
| |
| private: |
| static void EntryHookTrampoline(intptr_t function, |
| intptr_t stack_pointer, |
| Isolate* isolate); |
| |
| Major MajorKey() { return ProfileEntryHook; } |
| int MinorKey() { return 0; } |
| |
| void Generate(MacroAssembler* masm); |
| |
| DISALLOW_COPY_AND_ASSIGN(ProfileEntryHookStub); |
| }; |
| |
| } } // namespace v8::internal |
| |
| #endif // V8_CODE_STUBS_H_ |