blob: 8800e4b08f31cd7370ba9e41ed0af60dfbf9a14a [file] [log] [blame]
/*
* Copyright (C) 2012 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef ART_COMPILER_DEX_COMPILER_ENUMS_H_
#define ART_COMPILER_DEX_COMPILER_ENUMS_H_
#include "dex_instruction.h"
namespace art {
enum RegisterClass {
kInvalidRegClass,
kCoreReg,
kFPReg,
kRefReg,
kAnyReg,
};
std::ostream& operator<<(std::ostream& os, const RegisterClass& rhs);
enum BitsUsed {
kSize32Bits,
kSize64Bits,
kSize128Bits,
kSize256Bits,
kSize512Bits,
kSize1024Bits,
};
std::ostream& operator<<(std::ostream& os, const BitsUsed& rhs);
enum SpecialTargetRegister {
kSelf, // Thread pointer.
kSuspend, // Used to reduce suspend checks for some targets.
kLr,
kPc,
kSp,
kArg0,
kArg1,
kArg2,
kArg3,
kArg4,
kArg5,
kArg6,
kArg7,
kFArg0,
kFArg1,
kFArg2,
kFArg3,
kFArg4,
kFArg5,
kFArg6,
kFArg7,
kFArg8,
kFArg9,
kFArg10,
kFArg11,
kFArg12,
kFArg13,
kFArg14,
kFArg15,
kRet0,
kRet1,
kInvokeTgt,
kHiddenArg,
kHiddenFpArg,
kCount
};
std::ostream& operator<<(std::ostream& os, const SpecialTargetRegister& code);
enum RegLocationType {
kLocDalvikFrame = 0, // Normal Dalvik register
kLocPhysReg,
kLocCompilerTemp,
kLocInvalid
};
std::ostream& operator<<(std::ostream& os, const RegLocationType& rhs);
enum BBType {
kNullBlock,
kEntryBlock,
kDalvikByteCode,
kExitBlock,
kExceptionHandling,
kDead,
};
std::ostream& operator<<(std::ostream& os, const BBType& code);
// Shared pseudo opcodes - must be < 0.
enum LIRPseudoOpcode {
kPseudoPrologueBegin = -18,
kPseudoPrologueEnd = -17,
kPseudoEpilogueBegin = -16,
kPseudoEpilogueEnd = -15,
kPseudoExportedPC = -14,
kPseudoSafepointPC = -13,
kPseudoIntrinsicRetry = -12,
kPseudoSuspendTarget = -11,
kPseudoThrowTarget = -10,
kPseudoCaseLabel = -9,
kPseudoBarrier = -8,
kPseudoEntryBlock = -7,
kPseudoExitBlock = -6,
kPseudoTargetLabel = -5,
kPseudoDalvikByteCodeBoundary = -4,
kPseudoPseudoAlign4 = -3,
kPseudoEHBlockLabel = -2,
kPseudoNormalBlockLabel = -1,
};
std::ostream& operator<<(std::ostream& os, const LIRPseudoOpcode& rhs);
enum ExtendedMIROpcode {
kMirOpFirst = kNumPackedOpcodes,
kMirOpPhi = kMirOpFirst,
// @brief Copy from one VR to another.
// @details
// vA: destination VR
// vB: source VR
kMirOpCopy,
// @brief Used to do float comparison with less-than bias.
// @details Unlike cmpl-float, this does not store result of comparison in VR.
// vA: left-hand side VR for comparison.
// vB: right-hand side VR for comparison.
kMirOpFusedCmplFloat,
// @brief Used to do float comparison with greater-than bias.
// @details Unlike cmpg-float, this does not store result of comparison in VR.
// vA: left-hand side VR for comparison.
// vB: right-hand side VR for comparison.
kMirOpFusedCmpgFloat,
// @brief Used to do double comparison with less-than bias.
// @details Unlike cmpl-double, this does not store result of comparison in VR.
// vA: left-hand side wide VR for comparison.
// vB: right-hand side wide VR for comparison.
kMirOpFusedCmplDouble,
// @brief Used to do double comparison with greater-than bias.
// @details Unlike cmpl-double, this does not store result of comparison in VR.
// vA: left-hand side wide VR for comparison.
// vB: right-hand side wide VR for comparison.
kMirOpFusedCmpgDouble,
// @brief Used to do comparison of 64-bit long integers.
// @details Unlike cmp-long, this does not store result of comparison in VR.
// vA: left-hand side wide VR for comparison.
// vB: right-hand side wide VR for comparison.
kMirOpFusedCmpLong,
// @brief This represents no-op.
kMirOpNop,
// @brief Do a null check on the object register.
// @details The backends may implement this implicitly or explicitly. This MIR is guaranteed
// to have the correct offset as an exception thrower.
// vA: object register
kMirOpNullCheck,
kMirOpRangeCheck,
kMirOpDivZeroCheck,
kMirOpCheck,
kMirOpSelect,
// Vector opcodes:
// TypeSize is an encoded field giving the element type and the vector size.
// It is encoded as OpSize << 16 | (number of bits in vector)
//
// Destination and source are integers that will be interpreted by the
// backend that supports Vector operations. Backends are permitted to support only
// certain vector register sizes.
//
// At this point, only two operand instructions are supported. Three operand instructions
// could be supported by using a bit in TypeSize and arg[0] where needed.
// @brief MIR to move constant data to a vector register
// vA: destination
// vB: number of bits in register
// args[0]~args[3]: up to 128 bits of data for initialization
kMirOpConstVector,
// @brief MIR to move a vectorized register to another
// vA: destination
// vB: source
// vC: TypeSize
kMirOpMoveVector,
// @brief Packed multiply of units in two vector registers: vB = vB .* vC using vA to know the type of the vector.
// vA: destination and source
// vB: source
// vC: TypeSize
kMirOpPackedMultiply,
// @brief Packed addition of units in two vector registers: vB = vB .+ vC using vA to know the type of the vector.
// vA: destination and source
// vB: source
// vC: TypeSize
kMirOpPackedAddition,
// @brief Packed subtraction of units in two vector registers: vB = vB .- vC using vA to know the type of the vector.
// vA: destination and source
// vB: source
// vC: TypeSize
kMirOpPackedSubtract,
// @brief Packed shift left of units in two vector registers: vB = vB .<< vC using vA to know the type of the vector.
// vA: destination and source
// vB: amount to shift
// vC: TypeSize
kMirOpPackedShiftLeft,
// @brief Packed signed shift right of units in two vector registers: vB = vB .>> vC using vA to know the type of the vector.
// vA: destination and source
// vB: amount to shift
// vC: TypeSize
kMirOpPackedSignedShiftRight,
// @brief Packed unsigned shift right of units in two vector registers: vB = vB .>>> vC using vA to know the type of the vector.
// vA: destination and source
// vB: amount to shift
// vC: TypeSize
kMirOpPackedUnsignedShiftRight,
// @brief Packed bitwise and of units in two vector registers: vB = vB .& vC using vA to know the type of the vector.
// vA: destination and source
// vB: source
// vC: TypeSize
kMirOpPackedAnd,
// @brief Packed bitwise or of units in two vector registers: vB = vB .| vC using vA to know the type of the vector.
// vA: destination and source
// vB: source
// vC: TypeSize
kMirOpPackedOr,
// @brief Packed bitwise xor of units in two vector registers: vB = vB .^ vC using vA to know the type of the vector.
// vA: destination and source
// vB: source
// vC: TypeSize
kMirOpPackedXor,
// @brief Reduce a 128-bit packed element into a single VR by taking lower bits
// @details Instruction does a horizontal addition of the packed elements and then adds it to VR
// vA: destination and source VR (not vector register)
// vB: source (vector register)
// vC: TypeSize
kMirOpPackedAddReduce,
// @brief Extract a packed element into a single VR.
// vA: destination VR (not vector register)
// vB: source (vector register)
// vC: TypeSize
// arg[0]: The index to use for extraction from vector register (which packed element)
kMirOpPackedReduce,
// @brief Create a vector value, with all TypeSize values equal to vC
// vA: destination vector register
// vB: source VR (not vector register)
// vC: TypeSize
kMirOpPackedSet,
// @brief Reserve a range of vector registers.
// vA: Start vector register to reserve.
// vB: Inclusive end vector register to reserve.
// @note: The backend may choose to map vector numbers used in vector opcodes.
// Reserved registers are removed from the list of backend temporary pool.
kMirOpReserveVectorRegisters,
// @brief Free a range of reserved vector registers
// vA: Start vector register to unreserve.
// vB: Inclusive end vector register to unreserve.
// @note: All currently reserved vector registers are returned to the temporary pool.
kMirOpReturnVectorRegisters,
// @brief Create a memory barrier.
// vA: a constant defined by enum MemBarrierKind.
kMirOpMemBarrier,
// @brief Used to fill a vector register with array values.
// @details Just as with normal arrays, access on null object register must ensure NullPointerException
// and invalid index must ensure ArrayIndexOutOfBoundsException. Exception behavior must be the same
// as the aget it replaced and must happen at same index. Therefore, it is generally recommended that
// before using this MIR, it is proven that exception is guaranteed to not be thrown and marked with
// MIR_IGNORE_NULL_CHECK and MIR_IGNORE_RANGE_CHECK.
// vA: destination vector register
// vB: array register
// vC: index register
// arg[0]: TypeSize (most other vector opcodes have this in vC)
kMirOpPackedArrayGet,
// @brief Used to store a vector register into array.
// @details Just as with normal arrays, access on null object register must ensure NullPointerException
// and invalid index must ensure ArrayIndexOutOfBoundsException. Exception behavior must be the same
// as the aget it replaced and must happen at same index. Therefore, it is generally recommended that
// before using this MIR, it is proven that exception is guaranteed to not be thrown and marked with
// MIR_IGNORE_NULL_CHECK and MIR_IGNORE_RANGE_CHECK.
// vA: source vector register
// vB: array register
// vC: index register
// arg[0]: TypeSize (most other vector opcodes have this in vC)
kMirOpPackedArrayPut,
// @brief Multiply-add integer.
// vA: destination
// vB: multiplicand
// vC: multiplier
// arg[0]: addend
kMirOpMaddInt,
// @brief Multiply-subtract integer.
// vA: destination
// vB: multiplicand
// vC: multiplier
// arg[0]: minuend
kMirOpMsubInt,
// @brief Multiply-add long.
// vA: destination
// vB: multiplicand
// vC: multiplier
// arg[0]: addend
kMirOpMaddLong,
// @brief Multiply-subtract long.
// vA: destination
// vB: multiplicand
// vC: multiplier
// arg[0]: minuend
kMirOpMsubLong,
kMirOpLast,
};
enum MIROptimizationFlagPositions {
kMIRIgnoreNullCheck = 0,
kMIRIgnoreRangeCheck,
kMIRIgnoreCheckCast,
kMIRStoreNonNullValue, // Storing non-null value, always mark GC card.
kMIRClassIsInitialized,
kMIRClassIsInDexCache,
kMirIgnoreDivZeroCheck,
kMIRInlined, // Invoke is inlined (ie dead).
kMIRInlinedPred, // Invoke is inlined via prediction.
kMIRCallee, // Instruction is inlined from callee.
kMIRIgnoreSuspendCheck,
kMIRDup,
kMIRMark, // Temporary node mark can be used by
// opt passes for their private needs.
kMIRStoreNonTemporal,
kMIRLastMIRFlag,
};
// For successor_block_list.
enum BlockListType {
kNotUsed = 0,
kCatch,
kPackedSwitch,
kSparseSwitch,
};
std::ostream& operator<<(std::ostream& os, const BlockListType& rhs);
enum AssemblerStatus {
kSuccess,
kRetryAll,
};
std::ostream& operator<<(std::ostream& os, const AssemblerStatus& rhs);
enum OpSize {
kWord, // Natural word size of target (32/64).
k32,
k64,
kReference, // Object reference; compressed on 64-bit targets.
kSingle,
kDouble,
kUnsignedHalf,
kSignedHalf,
kUnsignedByte,
kSignedByte,
};
std::ostream& operator<<(std::ostream& os, const OpSize& kind);
enum OpKind {
kOpMov,
kOpCmov,
kOpMvn,
kOpCmp,
kOpLsl,
kOpLsr,
kOpAsr,
kOpRor,
kOpNot,
kOpAnd,
kOpOr,
kOpXor,
kOpNeg,
kOpAdd,
kOpAdc,
kOpSub,
kOpSbc,
kOpRsub,
kOpMul,
kOpDiv,
kOpRem,
kOpBic,
kOpCmn,
kOpTst,
kOpRev,
kOpRevsh,
kOpBkpt,
kOpBlx,
kOpPush,
kOpPop,
kOp2Char,
kOp2Short,
kOp2Byte,
kOpCondBr,
kOpUncondBr,
kOpBx,
kOpInvalid,
};
std::ostream& operator<<(std::ostream& os, const OpKind& rhs);
enum MoveType {
kMov8GP, // Move 8-bit general purpose register.
kMov16GP, // Move 16-bit general purpose register.
kMov32GP, // Move 32-bit general purpose register.
kMov64GP, // Move 64-bit general purpose register.
kMov32FP, // Move 32-bit FP register.
kMov64FP, // Move 64-bit FP register.
kMovLo64FP, // Move low 32-bits of 64-bit FP register.
kMovHi64FP, // Move high 32-bits of 64-bit FP register.
kMovU128FP, // Move 128-bit FP register to/from possibly unaligned region.
kMov128FP = kMovU128FP,
kMovA128FP, // Move 128-bit FP register to/from region surely aligned to 16-bytes.
kMovLo128FP, // Move low 64-bits of 128-bit FP register.
kMovHi128FP, // Move high 64-bits of 128-bit FP register.
};
std::ostream& operator<<(std::ostream& os, const MoveType& kind);
enum ConditionCode {
kCondEq, // equal
kCondNe, // not equal
kCondCs, // carry set
kCondCc, // carry clear
kCondUlt, // unsigned less than
kCondUge, // unsigned greater than or same
kCondMi, // minus
kCondPl, // plus, positive or zero
kCondVs, // overflow
kCondVc, // no overflow
kCondHi, // unsigned greater than
kCondLs, // unsigned lower or same
kCondGe, // signed greater than or equal
kCondLt, // signed less than
kCondGt, // signed greater than
kCondLe, // signed less than or equal
kCondAl, // always
kCondNv, // never
};
std::ostream& operator<<(std::ostream& os, const ConditionCode& kind);
// Target specific condition encodings
enum ArmConditionCode {
kArmCondEq = 0x0, // 0000
kArmCondNe = 0x1, // 0001
kArmCondCs = 0x2, // 0010
kArmCondCc = 0x3, // 0011
kArmCondMi = 0x4, // 0100
kArmCondPl = 0x5, // 0101
kArmCondVs = 0x6, // 0110
kArmCondVc = 0x7, // 0111
kArmCondHi = 0x8, // 1000
kArmCondLs = 0x9, // 1001
kArmCondGe = 0xa, // 1010
kArmCondLt = 0xb, // 1011
kArmCondGt = 0xc, // 1100
kArmCondLe = 0xd, // 1101
kArmCondAl = 0xe, // 1110
kArmCondNv = 0xf, // 1111
};
std::ostream& operator<<(std::ostream& os, const ArmConditionCode& kind);
enum X86ConditionCode {
kX86CondO = 0x0, // overflow
kX86CondNo = 0x1, // not overflow
kX86CondB = 0x2, // below
kX86CondNae = kX86CondB, // not-above-equal
kX86CondC = kX86CondB, // carry
kX86CondNb = 0x3, // not-below
kX86CondAe = kX86CondNb, // above-equal
kX86CondNc = kX86CondNb, // not-carry
kX86CondZ = 0x4, // zero
kX86CondEq = kX86CondZ, // equal
kX86CondNz = 0x5, // not-zero
kX86CondNe = kX86CondNz, // not-equal
kX86CondBe = 0x6, // below-equal
kX86CondNa = kX86CondBe, // not-above
kX86CondNbe = 0x7, // not-below-equal
kX86CondA = kX86CondNbe, // above
kX86CondS = 0x8, // sign
kX86CondNs = 0x9, // not-sign
kX86CondP = 0xa, // 8-bit parity even
kX86CondPE = kX86CondP,
kX86CondNp = 0xb, // 8-bit parity odd
kX86CondPo = kX86CondNp,
kX86CondL = 0xc, // less-than
kX86CondNge = kX86CondL, // not-greater-equal
kX86CondNl = 0xd, // not-less-than
kX86CondGe = kX86CondNl, // not-greater-equal
kX86CondLe = 0xe, // less-than-equal
kX86CondNg = kX86CondLe, // not-greater
kX86CondNle = 0xf, // not-less-than
kX86CondG = kX86CondNle, // greater
};
std::ostream& operator<<(std::ostream& os, const X86ConditionCode& kind);
enum DividePattern {
DivideNone,
Divide3,
Divide5,
Divide7,
};
std::ostream& operator<<(std::ostream& os, const DividePattern& pattern);
/**
* @brief Memory barrier types (see "The JSR-133 Cookbook for Compiler Writers").
* @details We define the combined barrier types that are actually required
* by the Java Memory Model, rather than using exactly the terminology from
* the JSR-133 cookbook. These should, in many cases, be replaced by acquire/release
* primitives. Note that the JSR-133 cookbook generally does not deal with
* store atomicity issues, and the recipes there are not always entirely sufficient.
* The current recipe is as follows:
* -# Use AnyStore ~= (LoadStore | StoreStore) ~= release barrier before volatile store.
* -# Use AnyAny barrier after volatile store. (StoreLoad is as expensive.)
* -# Use LoadAny barrier ~= (LoadLoad | LoadStore) ~= acquire barrier after each volatile load.
* -# Use StoreStore barrier after all stores but before return from any constructor whose
* class has final fields.
* -# Use NTStoreStore to order non-temporal stores with respect to all later
* store-to-memory instructions. Only generated together with non-temporal stores.
*/
enum MemBarrierKind {
kAnyStore,
kLoadAny,
kStoreStore,
kAnyAny,
kNTStoreStore,
kLastBarrierKind = kNTStoreStore
};
std::ostream& operator<<(std::ostream& os, const MemBarrierKind& kind);
enum OpFeatureFlags {
kIsBranch = 0,
kNoOperand,
kIsUnaryOp,
kIsBinaryOp,
kIsTertiaryOp,
kIsQuadOp,
kIsQuinOp,
kIsSextupleOp,
kIsIT,
kIsMoveOp,
kMemLoad,
kMemStore,
kMemVolatile,
kMemScaledx0,
kMemScaledx2,
kMemScaledx4,
kPCRelFixup, // x86 FIXME: add NEEDS_FIXUP to instruction attributes.
kRegDef0,
kRegDef1,
kRegDef2,
kRegDefA,
kRegDefD,
kRegDefFPCSList0,
kRegDefFPCSList2,
kRegDefList0,
kRegDefList1,
kRegDefList2,
kRegDefLR,
kRegDefSP,
kRegUse0,
kRegUse1,
kRegUse2,
kRegUse3,
kRegUse4,
kRegUseA,
kRegUseC,
kRegUseD,
kRegUseB,
kRegUseFPCSList0,
kRegUseFPCSList2,
kRegUseList0,
kRegUseList1,
kRegUseLR,
kRegUsePC,
kRegUseSP,
kSetsCCodes,
kUsesCCodes,
kUseFpStack,
kUseHi,
kUseLo,
kDefHi,
kDefLo
};
std::ostream& operator<<(std::ostream& os, const OpFeatureFlags& rhs);
enum SelectInstructionKind {
kSelectNone,
kSelectConst,
kSelectMove,
kSelectGoto
};
std::ostream& operator<<(std::ostream& os, const SelectInstructionKind& kind);
// LIR fixup kinds for Arm and X86.
enum FixupKind {
kFixupNone,
kFixupLabel, // For labels we just adjust the offset.
kFixupLoad, // Mostly for immediates.
kFixupVLoad, // FP load which *may* be pc-relative.
kFixupCBxZ, // Cbz, Cbnz.
kFixupTBxZ, // Tbz, Tbnz.
kFixupCondBranch, // Conditional branch
kFixupT1Branch, // Thumb1 Unconditional branch
kFixupT2Branch, // Thumb2 Unconditional branch
kFixupBlx1, // Blx1 (start of Blx1/Blx2 pair).
kFixupBl1, // Bl1 (start of Bl1/Bl2 pair).
kFixupAdr, // Adr.
kFixupMovImmLST, // kThumb2MovImm16LST.
kFixupMovImmHST, // kThumb2MovImm16HST.
kFixupAlign4, // Align to 4-byte boundary.
kFixupA53Erratum835769, // Cortex A53 Erratum 835769.
kFixupSwitchTable, // X86_64 packed switch table.
};
std::ostream& operator<<(std::ostream& os, const FixupKind& kind);
enum VolatileKind {
kNotVolatile, // Load/Store is not volatile
kVolatile // Load/Store is volatile
};
std::ostream& operator<<(std::ostream& os, const VolatileKind& kind);
enum WideKind {
kNotWide, // Non-wide view
kWide, // Wide view
kRef // Ref width
};
std::ostream& operator<<(std::ostream& os, const WideKind& kind);
} // namespace art
#endif // ART_COMPILER_DEX_COMPILER_ENUMS_H_