/* * Copyright (C) 2009 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "Dalvik.h" #include "compiler/CompilerInternals.h" #ifndef _DALVIK_VM_COMPILER_CODEGEN_ARM_ARMLIR_H #define _DALVIK_VM_COMPILER_CODEGEN_ARM_ARMLIR_H /* * r0, r1, r2, r3 are always scratch * r4 (rPC) is scratch for Jit, but most be restored when resuming interp * r5 (rFP) is reserved [holds Dalvik frame pointer] * r6 (rGLUE) is reserved [holds current &interpState] * r7 (rINST) is scratch for Jit * r8 (rIBASE) is scratch for Jit, but must be restored when resuming interp * r9 is reserved * r10 is always scratch * r11 (fp) used by gcc unless -fomit-frame-pointer set [available for jit?] * r12 is always scratch * r13 (sp) is reserved * r14 (lr) is scratch for Jit * r15 (pc) is reserved * * Preserved across C calls: r4, r5, r6, r7, r8, r10, r11 * Trashed across C calls: r0, r1, r2, r3, r12, r14 * * Floating pointer registers * s0-s31 * d0-d15, where d0={s0,s1}, d1={s2,s3}, ... , d15={s30,s31} * * s16-s31 (d8-d15) preserved across C calls * s0-s15 (d0-d7) trashed across C calls * * For Thumb code use: * r0, r1, r2, r3 to hold operands/results * r4, r7 for temps * * For Thumb2 code use: * r0, r1, r2, r3, r8, r9, r10, r11, r12, r14 for operands/results * r4, r7 for temps * s16-s31/d8-d15 for operands/results * s0-s15/d0-d7 for temps * * When transitioning from code cache to interp: * restore rIBASE * restore rPC * restore r11? */ /* Offset to distingish FP regs */ #define FP_REG_OFFSET 32 /* Offset to distinguish DP FP regs */ #define FP_DOUBLE 64 /* Reg types */ #define REGTYPE(x) (x & (FP_REG_OFFSET | FP_DOUBLE)) #define FPREG(x) ((x & FP_REG_OFFSET) == FP_REG_OFFSET) #define LOWREG(x) ((x & 0x7) == x) #define DOUBLEREG(x) ((x & FP_DOUBLE) == FP_DOUBLE) #define SINGLEREG(x) (FPREG(x) && !DOUBLEREG(x)) /* * Note: the low register of a floating point pair is sufficient to * create the name of a double, but require both names to be passed to * allow for asserts to verify that the pair is consecutive if significant * rework is done in this area. Also, it is a good reminder in the calling * code that reg locations always describe doubles as a pair of singles. */ #define S2D(x,y) ((x) | FP_DOUBLE) /* Mask to strip off fp flags */ #define FP_REG_MASK (FP_REG_OFFSET-1) /* non-existent Dalvik register */ #define vNone (-1) /* non-existant physical register */ #define rNone (-1) /* RegisterLocation templates return values (r0, or r0/r1) */ #define LOC_C_RETURN {kLocPhysReg, 0, 0, r0, 0, -1} #define LOC_C_RETURN_WIDE {kLocPhysReg, 1, 0, r0, r1, -1} /* RegisterLocation templates for interpState->retVal; */ #define LOC_DALVIK_RETURN_VAL {kLocRetval, 0, 0, 0, 0, -1} #define LOC_DALVIK_RETURN_VAL_WIDE {kLocRetval, 1, 0, 0, 0, -1} /* * Data structure tracking the mapping between a Dalvik register (pair) and a * native register (pair). The idea is to reuse the previously loaded value * if possible, otherwise to keep the value in a native register as long as * possible. */ typedef struct RegisterInfo { int reg; // Reg number bool inUse; // Has it been allocated? bool pair; // Part of a register pair? int partner; // If pair, other reg of pair bool live; // Is there an associated SSA name? bool dirty; // If live, is it dirty? int sReg; // Name of live value struct LIR *defStart; // Starting inst in last def sequence struct LIR *defEnd; // Ending inst in last def sequence } RegisterInfo; typedef struct RegisterPool { BitVector *nullCheckedRegs; // Track which registers have been null-checked int numCoreTemps; RegisterInfo *coreTemps; int nextCoreTemp; int numFPTemps; RegisterInfo *FPTemps; int nextFPTemp; int numCoreRegs; RegisterInfo *coreRegs; int numFPRegs; RegisterInfo *FPRegs; } RegisterPool; typedef enum ResourceEncodingPos { kGPReg0 = 0, kRegSP = 13, kRegLR = 14, kRegPC = 15, kFPReg0 = 16, kRegEnd = 48, kCCode = kRegEnd, kFPStatus, kDalvikReg, kLiteral, kFrameRef, kHeapRef, kLitPoolRef } ResourceEncodingPos; #define ENCODE_REG_LIST(N) ((u8) N) #define ENCODE_REG_SP (1ULL << kRegSP) #define ENCODE_REG_LR (1ULL << kRegLR) #define ENCODE_REG_PC (1ULL << kRegPC) #define ENCODE_CCODE (1ULL << kCCode) #define ENCODE_FP_STATUS (1ULL << kFPStatus) /* Must alias */ #define ENCODE_DALVIK_REG (1ULL << kDalvikReg) #define ENCODE_LITERAL (1ULL << kLiteral) /* May alias */ #define ENCODE_FRAME_REF (1ULL << kFrameRef) #define ENCODE_HEAP_REF (1ULL << kHeapRef) #define ENCODE_LITPOOL_REF (1ULL << kLitPoolRef) #define ENCODE_ALL (~0ULL) #define ENCODE_MEM_DEF (ENCODE_FRAME_REF | ENCODE_HEAP_REF) #define ENCODE_MEM_USE (ENCODE_FRAME_REF | ENCODE_HEAP_REF \ | ENCODE_LITPOOL_REF) #define DECODE_ALIAS_INFO_REG(X) (X & 0xffff) #define DECODE_ALIAS_INFO_WIDE(X) ((X & 0x80000000) ? 1 : 0) typedef enum OpSize { kWord, kLong, kSingle, kDouble, kUnsignedHalf, kSignedHalf, kUnsignedByte, kSignedByte, } OpSize; typedef enum OpKind { kOpMov, kOpMvn, kOpCmp, kOpLsl, kOpLsr, kOpAsr, kOpRor, kOpNot, kOpAnd, kOpOr, kOpXor, kOpNeg, kOpAdd, kOpAdc, kOpSub, kOpSbc, kOpRsub, kOpMul, kOpDiv, kOpRem, kOpBic, kOpCmn, kOpTst, kOpBkpt, kOpBlx, kOpPush, kOpPop, kOp2Char, kOp2Short, kOp2Byte, kOpCondBr, kOpUncondBr, } OpKind; typedef enum NativeRegisterPool { r0 = 0, r1 = 1, r2 = 2, r3 = 3, r4PC = 4, rFP = 5, rGLUE = 6, r7 = 7, r8 = 8, r9 = 9, r10 = 10, r11 = 11, r12 = 12, r13 = 13, rlr = 14, rpc = 15, fr0 = 0 + FP_REG_OFFSET, fr1 = 1 + FP_REG_OFFSET, fr2 = 2 + FP_REG_OFFSET, fr3 = 3 + FP_REG_OFFSET, fr4 = 4 + FP_REG_OFFSET, fr5 = 5 + FP_REG_OFFSET, fr6 = 6 + FP_REG_OFFSET, fr7 = 7 + FP_REG_OFFSET, fr8 = 8 + FP_REG_OFFSET, fr9 = 9 + FP_REG_OFFSET, fr10 = 10 + FP_REG_OFFSET, fr11 = 11 + FP_REG_OFFSET, fr12 = 12 + FP_REG_OFFSET, fr13 = 13 + FP_REG_OFFSET, fr14 = 14 + FP_REG_OFFSET, fr15 = 15 + FP_REG_OFFSET, fr16 = 16 + FP_REG_OFFSET, fr17 = 17 + FP_REG_OFFSET, fr18 = 18 + FP_REG_OFFSET, fr19 = 19 + FP_REG_OFFSET, fr20 = 20 + FP_REG_OFFSET, fr21 = 21 + FP_REG_OFFSET, fr22 = 22 + FP_REG_OFFSET, fr23 = 23 + FP_REG_OFFSET, fr24 = 24 + FP_REG_OFFSET, fr25 = 25 + FP_REG_OFFSET, fr26 = 26 + FP_REG_OFFSET, fr27 = 27 + FP_REG_OFFSET, fr28 = 28 + FP_REG_OFFSET, fr29 = 29 + FP_REG_OFFSET, fr30 = 30 + FP_REG_OFFSET, fr31 = 31 + FP_REG_OFFSET, dr0 = fr0 + FP_DOUBLE, dr1 = fr2 + FP_DOUBLE, dr2 = fr4 + FP_DOUBLE, dr3 = fr6 + FP_DOUBLE, dr4 = fr8 + FP_DOUBLE, dr5 = fr10 + FP_DOUBLE, dr6 = fr12 + FP_DOUBLE, dr7 = fr14 + FP_DOUBLE, dr8 = fr16 + FP_DOUBLE, dr9 = fr18 + FP_DOUBLE, dr10 = fr20 + FP_DOUBLE, dr11 = fr22 + FP_DOUBLE, dr12 = fr24 + FP_DOUBLE, dr13 = fr26 + FP_DOUBLE, dr14 = fr28 + FP_DOUBLE, dr15 = fr30 + FP_DOUBLE, } NativeRegisterPool; /* Shift encodings */ typedef enum ArmShiftEncodings { kArmLsl = 0x0, kArmLsr = 0x1, kArmAsr = 0x2, kArmRor = 0x3 } ArmShiftEncodings; /* Thumb condition encodings */ typedef enum ArmConditionCode { kArmCondEq = 0x0, /* 0000 */ kArmCondNe = 0x1, /* 0001 */ kArmCondCs = 0x2, /* 0010 */ kArmCondCc = 0x3, /* 0011 */ kArmCondMi = 0x4, /* 0100 */ kArmCondPl = 0x5, /* 0101 */ kArmCondVs = 0x6, /* 0110 */ kArmCondVc = 0x7, /* 0111 */ kArmCondHi = 0x8, /* 1000 */ kArmCondLs = 0x9, /* 1001 */ kArmCondGe = 0xa, /* 1010 */ kArmCondLt = 0xb, /* 1011 */ kArmCondGt = 0xc, /* 1100 */ kArmCondLe = 0xd, /* 1101 */ kArmCondAl = 0xe, /* 1110 */ kArmCondNv = 0xf, /* 1111 */ } ArmConditionCode; #define isPseudoOpCode(opCode) ((int)(opCode) < 0) /* * The following enum defines the list of supported Thumb instructions by the * assembler. Their corresponding snippet positions will be defined in * Assemble.c. */ typedef enum ArmOpCode { kArmChainingCellBottom = -18, kArmPseudoBarrier = -17, kArmPseudoExtended = -16, kArmPseudoSSARep = -15, kArmPseudoEntryBlock = -14, kArmPseudoExitBlock = -13, kArmPseudoTargetLabel = -12, kArmPseudoChainingCellBackwardBranch = -11, kArmPseudoChainingCellHot = -10, kArmPseudoChainingCellInvokePredicted = -9, kArmPseudoChainingCellInvokeSingleton = -8, kArmPseudoChainingCellNormal = -7, kArmPseudoDalvikByteCodeBoundary = -6, kArmPseudoPseudoAlign4 = -5, kArmPseudoPCReconstructionCell = -4, kArmPseudoPCReconstructionBlockLabel = -3, kArmPseudoEHBlockLabel = -2, kArmPseudoNormalBlockLabel = -1, /************************************************************************/ kArm16BitData, /* DATA [0] rd[15..0] */ kThumbAdcRR, /* adc [0100000101] rm[5..3] rd[2..0] */ kThumbAddRRI3, /* add(1) [0001110] imm_3[8..6] rn[5..3] rd[2..0]*/ kThumbAddRI8, /* add(2) [00110] rd[10..8] imm_8[7..0] */ kThumbAddRRR, /* add(3) [0001100] rm[8..6] rn[5..3] rd[2..0] */ kThumbAddRRLH, /* add(4) [01000100] H12[01] rm[5..3] rd[2..0] */ kThumbAddRRHL, /* add(4) [01001000] H12[10] rm[5..3] rd[2..0] */ kThumbAddRRHH, /* add(4) [01001100] H12[11] rm[5..3] rd[2..0] */ kThumbAddPcRel, /* add(5) [10100] rd[10..8] imm_8[7..0] */ kThumbAddSpRel, /* add(6) [10101] rd[10..8] imm_8[7..0] */ kThumbAddSpI7, /* add(7) [101100000] imm_7[6..0] */ kThumbAndRR, /* and [0100000000] rm[5..3] rd[2..0] */ kThumbAsrRRI5, /* asr(1) [00010] imm_5[10..6] rm[5..3] rd[2..0] */ kThumbAsrRR, /* asr(2) [0100000100] rs[5..3] rd[2..0] */ kThumbBCond, /* b(1) [1101] cond[11..8] offset_8[7..0] */ kThumbBUncond, /* b(2) [11100] offset_11[10..0] */ kThumbBicRR, /* bic [0100001110] rm[5..3] rd[2..0] */ kThumbBkpt, /* bkpt [10111110] imm_8[7..0] */ kThumbBlx1, /* blx(1) [111] H[10] offset_11[10..0] */ kThumbBlx2, /* blx(1) [111] H[01] offset_11[10..0] */ kThumbBl1, /* blx(1) [111] H[10] offset_11[10..0] */ kThumbBl2, /* blx(1) [111] H[11] offset_11[10..0] */ kThumbBlxR, /* blx(2) [010001111] rm[6..3] [000] */ kThumbBx, /* bx [010001110] H2[6..6] rm[5..3] SBZ[000] */ kThumbCmnRR, /* cmn [0100001011] rm[5..3] rd[2..0] */ kThumbCmpRI8, /* cmp(1) [00101] rn[10..8] imm_8[7..0] */ kThumbCmpRR, /* cmp(2) [0100001010] rm[5..3] rd[2..0] */ kThumbCmpLH, /* cmp(3) [01000101] H12[01] rm[5..3] rd[2..0] */ kThumbCmpHL, /* cmp(3) [01000110] H12[10] rm[5..3] rd[2..0] */ kThumbCmpHH, /* cmp(3) [01000111] H12[11] rm[5..3] rd[2..0] */ kThumbEorRR, /* eor [0100000001] rm[5..3] rd[2..0] */ kThumbLdmia, /* ldmia [11001] rn[10..8] reglist [7..0] */ kThumbLdrRRI5, /* ldr(1) [01101] imm_5[10..6] rn[5..3] rd[2..0] */ kThumbLdrRRR, /* ldr(2) [0101100] rm[8..6] rn[5..3] rd[2..0] */ kThumbLdrPcRel, /* ldr(3) [01001] rd[10..8] imm_8[7..0] */ kThumbLdrSpRel, /* ldr(4) [10011] rd[10..8] imm_8[7..0] */ kThumbLdrbRRI5, /* ldrb(1) [01111] imm_5[10..6] rn[5..3] rd[2..0] */ kThumbLdrbRRR, /* ldrb(2) [0101110] rm[8..6] rn[5..3] rd[2..0] */ kThumbLdrhRRI5, /* ldrh(1) [10001] imm_5[10..6] rn[5..3] rd[2..0] */ kThumbLdrhRRR, /* ldrh(2) [0101101] rm[8..6] rn[5..3] rd[2..0] */ kThumbLdrsbRRR, /* ldrsb [0101011] rm[8..6] rn[5..3] rd[2..0] */ kThumbLdrshRRR, /* ldrsh [0101111] rm[8..6] rn[5..3] rd[2..0] */ kThumbLslRRI5, /* lsl(1) [00000] imm_5[10..6] rm[5..3] rd[2..0] */ kThumbLslRR, /* lsl(2) [0100000010] rs[5..3] rd[2..0] */ kThumbLsrRRI5, /* lsr(1) [00001] imm_5[10..6] rm[5..3] rd[2..0] */ kThumbLsrRR, /* lsr(2) [0100000011] rs[5..3] rd[2..0] */ kThumbMovImm, /* mov(1) [00100] rd[10..8] imm_8[7..0] */ kThumbMovRR, /* mov(2) [0001110000] rn[5..3] rd[2..0] */ kThumbMovRR_H2H, /* mov(3) [01000111] H12[11] rm[5..3] rd[2..0] */ kThumbMovRR_H2L, /* mov(3) [01000110] H12[01] rm[5..3] rd[2..0] */ kThumbMovRR_L2H, /* mov(3) [01000101] H12[10] rm[5..3] rd[2..0] */ kThumbMul, /* mul [0100001101] rm[5..3] rd[2..0] */ kThumbMvn, /* mvn [0100001111] rm[5..3] rd[2..0] */ kThumbNeg, /* neg [0100001001] rm[5..3] rd[2..0] */ kThumbOrr, /* orr [0100001100] rm[5..3] rd[2..0] */ kThumbPop, /* pop [1011110] r[8..8] rl[7..0] */ kThumbPush, /* push [1011010] r[8..8] rl[7..0] */ kThumbRorRR, /* ror [0100000111] rs[5..3] rd[2..0] */ kThumbSbc, /* sbc [0100000110] rm[5..3] rd[2..0] */ kThumbStmia, /* stmia [11000] rn[10..8] reglist [7.. 0] */ kThumbStrRRI5, /* str(1) [01100] imm_5[10..6] rn[5..3] rd[2..0] */ kThumbStrRRR, /* str(2) [0101000] rm[8..6] rn[5..3] rd[2..0] */ kThumbStrSpRel, /* str(3) [10010] rd[10..8] imm_8[7..0] */ kThumbStrbRRI5, /* strb(1) [01110] imm_5[10..6] rn[5..3] rd[2..0] */ kThumbStrbRRR, /* strb(2) [0101010] rm[8..6] rn[5..3] rd[2..0] */ kThumbStrhRRI5, /* strh(1) [10000] imm_5[10..6] rn[5..3] rd[2..0] */ kThumbStrhRRR, /* strh(2) [0101001] rm[8..6] rn[5..3] rd[2..0] */ kThumbSubRRI3, /* sub(1) [0001111] imm_3[8..6] rn[5..3] rd[2..0]*/ kThumbSubRI8, /* sub(2) [00111] rd[10..8] imm_8[7..0] */ kThumbSubRRR, /* sub(3) [0001101] rm[8..6] rn[5..3] rd[2..0] */ kThumbSubSpI7, /* sub(4) [101100001] imm_7[6..0] */ kThumbSwi, /* swi [11011111] imm_8[7..0] */ kThumbTst, /* tst [0100001000] rm[5..3] rn[2..0] */ kThumb2Vldrs, /* vldr low sx [111011011001] rn[19..16] rd[15-12] [1010] imm_8[7..0] */ kThumb2Vldrd, /* vldr low dx [111011011001] rn[19..16] rd[15-12] [1011] imm_8[7..0] */ kThumb2Vmuls, /* vmul vd, vn, vm [111011100010] rn[19..16] rd[15-12] [10100000] rm[3..0] */ kThumb2Vmuld, /* vmul vd, vn, vm [111011100010] rn[19..16] rd[15-12] [10110000] rm[3..0] */ kThumb2Vstrs, /* vstr low sx [111011011000] rn[19..16] rd[15-12] [1010] imm_8[7..0] */ kThumb2Vstrd, /* vstr low dx [111011011000] rn[19..16] rd[15-12] [1011] imm_8[7..0] */ kThumb2Vsubs, /* vsub vd, vn, vm [111011100011] rn[19..16] rd[15-12] [10100040] rm[3..0] */ kThumb2Vsubd, /* vsub vd, vn, vm [111011100011] rn[19..16] rd[15-12] [10110040] rm[3..0] */ kThumb2Vadds, /* vadd vd, vn, vm [111011100011] rn[19..16] rd[15-12] [10100000] rm[3..0] */ kThumb2Vaddd, /* vadd vd, vn, vm [111011100011] rn[19..16] rd[15-12] [10110000] rm[3..0] */ kThumb2Vdivs, /* vdiv vd, vn, vm [111011101000] rn[19..16] rd[15-12] [10100000] rm[3..0] */ kThumb2Vdivd, /* vdiv vd, vn, vm [111011101000] rn[19..16] rd[15-12] [10110000] rm[3..0] */ kThumb2VcvtIF, /* vcvt.F32 vd, vm [1110111010111000] vd[15..12] [10101100] vm[3..0] */ kThumb2VcvtID, /* vcvt.F64 vd, vm [1110111010111000] vd[15..12] [10111100] vm[3..0] */ kThumb2VcvtFI, /* vcvt.S32.F32 vd, vm [1110111010111101] vd[15..12] [10101100] vm[3..0] */ kThumb2VcvtDI, /* vcvt.S32.F32 vd, vm [1110111010111101] vd[15..12] [10111100] vm[3..0] */ kThumb2VcvtFd, /* vcvt.F64.F32 vd, vm [1110111010110111] vd[15..12] [10101100] vm[3..0] */ kThumb2VcvtDF, /* vcvt.F32.F64 vd, vm [1110111010110111] vd[15..12] [10111100] vm[3..0] */ kThumb2Vsqrts, /* vsqrt.f32 vd, vm [1110111010110001] vd[15..12] [10101100] vm[3..0] */ kThumb2Vsqrtd, /* vsqrt.f64 vd, vm [1110111010110001] vd[15..12] [10111100] vm[3..0] */ kThumb2MovImmShift, /* mov(T2) rd, #<const> [11110] i [00001001111] imm3 rd[11..8] imm8 */ kThumb2MovImm16, /* mov(T3) rd, #<const> [11110] i [0010100] imm4 [0] imm3 rd[11..8] imm8 */ kThumb2StrRRI12, /* str(Imm,T3) rd,[rn,#imm12] [111110001100] rn[19..16] rt[15..12] imm12[11..0] */ kThumb2LdrRRI12, /* str(Imm,T3) rd,[rn,#imm12] [111110001100] rn[19..16] rt[15..12] imm12[11..0] */ kThumb2StrRRI8Predec, /* str(Imm,T4) rd,[rn,#-imm8] [111110000100] rn[19..16] rt[15..12] [1100] imm[7..0]*/ kThumb2LdrRRI8Predec, /* ldr(Imm,T4) rd,[rn,#-imm8] [111110000101] rn[19..16] rt[15..12] [1100] imm[7..0]*/ kThumb2Cbnz, /* cbnz rd,<label> [101110] i [1] imm5[7..3] rn[2..0] */ kThumb2Cbz, /* cbn rd,<label> [101100] i [1] imm5[7..3] rn[2..0] */ kThumb2AddRRI12, /* add rd, rn, #imm12 [11110] i [100000] rn[19..16] [0] imm3[14..12] rd[11..8] imm8[7..0] */ kThumb2MovRR, /* mov rd, rm [11101010010011110000] rd[11..8] [0000] rm[3..0] */ kThumb2Vmovs, /* vmov.f32 vd, vm [111011101] D [110000] vd[15..12] 101001] M [0] vm[3..0] */ kThumb2Vmovd, /* vmov.f64 vd, vm [111011101] D [110000] vd[15..12] 101101] M [0] vm[3..0] */ kThumb2Ldmia, /* ldmia [111010001001[ rn[19..16] mask[15..0] */ kThumb2Stmia, /* stmia [111010001000[ rn[19..16] mask[15..0] */ kThumb2AddRRR, /* add [111010110000] rn[19..16] [0000] rd[11..8] [0000] rm[3..0] */ kThumb2SubRRR, /* sub [111010111010] rn[19..16] [0000] rd[11..8] [0000] rm[3..0] */ kThumb2SbcRRR, /* sbc [111010110110] rn[19..16] [0000] rd[11..8] [0000] rm[3..0] */ kThumb2CmpRR, /* cmp [111010111011] rn[19..16] [0000] [1111] [0000] rm[3..0] */ kThumb2SubRRI12, /* sub rd, rn, #imm12 [11110] i [01010] rn[19..16] [0] imm3[14..12] rd[11..8] imm8[7..0] */ kThumb2MvnImmShift, /* mov(T2) rd, #<const> [11110] i [00011011110] imm3 rd[11..8] imm8 */ kThumb2Sel, /* sel rd, rn, rm [111110101010] rn[19-16] rd[11-8] rm[3-0] */ kThumb2Ubfx, /* ubfx rd,rn,#lsb,#width [111100111100] rn[19..16] [0] imm3[14-12] rd[11-8] w[4-0] */ kThumb2Sbfx, /* ubfx rd,rn,#lsb,#width [111100110100] rn[19..16] [0] imm3[14-12] rd[11-8] w[4-0] */ kThumb2LdrRRR, /* ldr rt,[rn,rm,LSL #imm] [111110000101] rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ kThumb2LdrhRRR, /* ldrh rt,[rn,rm,LSL #imm] [111110000101] rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ kThumb2LdrshRRR, /* ldrsh rt,[rn,rm,LSL #imm] [111110000101] rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ kThumb2LdrbRRR, /* ldrb rt,[rn,rm,LSL #imm] [111110000101] rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ kThumb2LdrsbRRR, /* ldrsb rt,[rn,rm,LSL #imm] [111110000101] rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ kThumb2StrRRR, /* str rt,[rn,rm,LSL #imm] [111110000100] rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ kThumb2StrhRRR, /* str rt,[rn,rm,LSL #imm] [111110000010] rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ kThumb2StrbRRR, /* str rt,[rn,rm,LSL #imm] [111110000000] rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ kThumb2LdrhRRI12, /* ldrh rt,[rn,#imm12] [111110001011] rt[15..12] rn[19..16] imm12[11..0] */ kThumb2LdrshRRI12, /* ldrsh rt,[rn,#imm12] [111110011011] rt[15..12] rn[19..16] imm12[11..0] */ kThumb2LdrbRRI12, /* ldrb rt,[rn,#imm12] [111110001001] rt[15..12] rn[19..16] imm12[11..0] */ kThumb2LdrsbRRI12, /* ldrsb rt,[rn,#imm12] [111110011001] rt[15..12] rn[19..16] imm12[11..0] */ kThumb2StrhRRI12, /* strh rt,[rn,#imm12] [111110001010] rt[15..12] rn[19..16] imm12[11..0] */ kThumb2StrbRRI12, /* strb rt,[rn,#imm12] [111110001000] rt[15..12] rn[19..16] imm12[11..0] */ kThumb2Pop, /* pop [1110100010111101] list[15-0]*/ kThumb2Push, /* push [1110100010101101] list[15-0]*/ kThumb2CmpRI8, /* cmp rn, #<const> [11110] i [011011] rn[19-16] [0] imm3 [1111] imm8[7..0] */ kThumb2AdcRRR, /* adc [111010110101] rn[19..16] [0000] rd[11..8] [0000] rm[3..0] */ kThumb2AndRRR, /* and [111010100000] rn[19..16] [0000] rd[11..8] [0000] rm[3..0] */ kThumb2BicRRR, /* bic [111010100010] rn[19..16] [0000] rd[11..8] [0000] rm[3..0] */ kThumb2CmnRR, /* cmn [111010110001] rn[19..16] [0000] [1111] [0000] rm[3..0] */ kThumb2EorRRR, /* eor [111010101000] rn[19..16] [0000] rd[11..8] [0000] rm[3..0] */ kThumb2MulRRR, /* mul [111110110000] rn[19..16] [1111] rd[11..8] [0000] rm[3..0] */ kThumb2MnvRR, /* mvn [11101010011011110] rd[11-8] [0000] rm[3..0] */ kThumb2RsubRRI8, /* rsub [111100011100] rn[19..16] [0000] rd[11..8] imm8[7..0] */ kThumb2NegRR, /* actually rsub rd, rn, #0 */ kThumb2OrrRRR, /* orr [111010100100] rn[19..16] [0000] rd[11..8] [0000] rm[3..0] */ kThumb2TstRR, /* tst [111010100001] rn[19..16] [0000] [1111] [0000] rm[3..0] */ kThumb2LslRRR, /* lsl [111110100000] rn[19..16] [1111] rd[11..8] [0000] rm[3..0] */ kThumb2LsrRRR, /* lsr [111110100010] rn[19..16] [1111] rd[11..8] [0000] rm[3..0] */ kThumb2AsrRRR, /* asr [111110100100] rn[19..16] [1111] rd[11..8] [0000] rm[3..0] */ kThumb2RorRRR, /* ror [111110100110] rn[19..16] [1111] rd[11..8] [0000] rm[3..0] */ kThumb2LslRRI5, /* lsl [11101010010011110] imm[14.12] rd[11..8] [00] rm[3..0] */ kThumb2LsrRRI5, /* lsr [11101010010011110] imm[14.12] rd[11..8] [01] rm[3..0] */ kThumb2AsrRRI5, /* asr [11101010010011110] imm[14.12] rd[11..8] [10] rm[3..0] */ kThumb2RorRRI5, /* ror [11101010010011110] imm[14.12] rd[11..8] [11] rm[3..0] */ kThumb2BicRRI8, /* bic [111100000010] rn[19..16] [0] imm3 rd[11..8] imm8 */ kThumb2AndRRI8, /* bic [111100000000] rn[19..16] [0] imm3 rd[11..8] imm8 */ kThumb2OrrRRI8, /* orr [111100000100] rn[19..16] [0] imm3 rd[11..8] imm8 */ kThumb2EorRRI8, /* eor [111100001000] rn[19..16] [0] imm3 rd[11..8] imm8 */ kThumb2AddRRI8, /* add [111100001000] rn[19..16] [0] imm3 rd[11..8] imm8 */ kThumb2AdcRRI8, /* adc [111100010101] rn[19..16] [0] imm3 rd[11..8] imm8 */ kThumb2SubRRI8, /* sub [111100011011] rn[19..16] [0] imm3 rd[11..8] imm8 */ kThumb2SbcRRI8, /* sbc [111100010111] rn[19..16] [0] imm3 rd[11..8] imm8 */ kThumb2It, /* it [10111111] firstcond[7-4] mask[3-0] */ kThumb2Fmstat, /* fmstat [11101110111100011111101000010000] */ kThumb2Vcmpd, /* vcmp [111011101] D [11011] rd[15-12] [1011] E [1] M [0] rm[3-0] */ kThumb2Vcmps, /* vcmp [111011101] D [11010] rd[15-12] [1011] E [1] M [0] rm[3-0] */ kThumb2LdrPcRel12, /* ldr rd,[pc,#imm12] [1111100011011111] rt[15-12] imm12[11-0] */ kThumb2BCond, /* b<c> [1110] S cond[25-22] imm6[21-16] [10] J1 [0] J2 imm11[10..0] */ kThumb2Vmovd_RR, /* vmov [111011101] D [110000] vd[15-12 [101101] M [0] vm[3-0] */ kThumb2Vmovs_RR, /* vmov [111011101] D [110000] vd[15-12 [101001] M [0] vm[3-0] */ kThumb2Fmrs, /* vmov [111011100000] vn[19-16] rt[15-12] [1010] N [0010000] */ kThumb2Fmsr, /* vmov [111011100001] vn[19-16] rt[15-12] [1010] N [0010000] */ kThumb2Fmrrd, /* vmov [111011000100] rt2[19-16] rt[15-12] [101100] M [1] vm[3-0] */ kThumb2Fmdrr, /* vmov [111011000101] rt2[19-16] rt[15-12] [101100] M [1] vm[3-0] */ kThumb2Vabsd, /* vabs.f64 [111011101] D [110000] rd[15-12] [1011110] M [0] vm[3-0] */ kThumb2Vabss, /* vabs.f32 [111011101] D [110000] rd[15-12] [1010110] M [0] vm[3-0] */ kThumb2Vnegd, /* vneg.f64 [111011101] D [110000] rd[15-12] [1011110] M [0] vm[3-0] */ kThumb2Vnegs, /* vneg.f32 [111011101] D [110000] rd[15-12] [1010110] M [0] vm[3-0] */ kThumb2Vmovs_IMM8, /* vmov.f32 [111011101] D [11] imm4h[19-16] vd[15-12] [10100000] imm4l[3-0] */ kThumb2Vmovd_IMM8, /* vmov.f64 [111011101] D [11] imm4h[19-16] vd[15-12] [10110000] imm4l[3-0] */ kThumb2Mla, /* mla [111110110000] rn[19-16] ra[15-12] rd[7-4] [0000] rm[3-0] */ kThumb2Umull, /* umull [111110111010] rn[19-16], rdlo[15-12] rdhi[11-8] [0000] rm[3-0] */ kThumb2Ldrex, /* ldrex [111010000101] rn[19-16] rt[11-8] [1111] imm8[7-0] */ kThumb2Strex, /* strex [111010000100] rn[19-16] rt[11-8] rd[11-8] imm8[7-0] */ kThumb2Clrex, /* clrex [111100111011111110000111100101111] */ kThumb2Bfi, /* bfi [111100110110] rn[19-16] [0] imm3[14-12] rd[11-8] imm2[7-6] [0] msb[4-0] */ kThumb2Bfc, /* bfc [11110011011011110] [0] imm3[14-12] rd[11-8] imm2[7-6] [0] msb[4-0] */ kThumb2Dmb, /* dmb [1111001110111111100011110101] option[3-0] */ kArmLast, } ArmOpCode; /* DMB option encodings */ typedef enum ArmOpDmbOptions { kSY = 0xf, kST = 0xe, kISH = 0xb, kISHST = 0xa, kNSH = 0x7, kNSHST = 0x6 } ArmOpDmbOptions; /* Bit flags describing the behavior of each native opcode */ typedef enum ArmOpFeatureFlags { kIsBranch = 0, kRegDef0, kRegDef1, kRegDefSP, kRegDefLR, kRegDefList0, kRegDefList1, kRegUse0, kRegUse1, kRegUse2, kRegUse3, kRegUseSP, kRegUsePC, kRegUseList0, kRegUseList1, kNoOperand, kIsUnaryOp, kIsBinaryOp, kIsTertiaryOp, kIsQuadOp, kIsIT, kSetsCCodes, kUsesCCodes, kMemLoad, kMemStore, } ArmOpFeatureFlags; #define IS_LOAD (1 << kMemLoad) #define IS_STORE (1 << kMemStore) #define IS_BRANCH (1 << kIsBranch) #define REG_DEF0 (1 << kRegDef0) #define REG_DEF1 (1 << kRegDef1) #define REG_DEF_SP (1 << kRegDefSP) #define REG_DEF_LR (1 << kRegDefLR) #define REG_DEF_LIST0 (1 << kRegDefList0) #define REG_DEF_LIST1 (1 << kRegDefList1) #define REG_USE0 (1 << kRegUse0) #define REG_USE1 (1 << kRegUse1) #define REG_USE2 (1 << kRegUse2) #define REG_USE3 (1 << kRegUse3) #define REG_USE_SP (1 << kRegUseSP) #define REG_USE_PC (1 << kRegUsePC) #define REG_USE_LIST0 (1 << kRegUseList0) #define REG_USE_LIST1 (1 << kRegUseList1) #define NO_OPERAND (1 << kNoOperand) #define IS_UNARY_OP (1 << kIsUnaryOp) #define IS_BINARY_OP (1 << kIsBinaryOp) #define IS_TERTIARY_OP (1 << kIsTertiaryOp) #define IS_QUAD_OP (1 << kIsQuadOp) #define IS_IT (1 << kIsIT) #define SETS_CCODES (1 << kSetsCCodes) #define USES_CCODES (1 << kUsesCCodes) /* Common combo register usage patterns */ #define REG_USE01 (REG_USE0 | REG_USE1) #define REG_USE012 (REG_USE01 | REG_USE2) #define REG_USE12 (REG_USE1 | REG_USE2) #define REG_DEF0_USE0 (REG_DEF0 | REG_USE0) #define REG_DEF0_USE1 (REG_DEF0 | REG_USE1) #define REG_DEF0_USE01 (REG_DEF0 | REG_USE01) #define REG_DEF0_USE12 (REG_DEF0 | REG_USE12) #define REG_DEF01_USE2 (REG_DEF0 | REG_DEF1 | REG_USE2) /* Instruction assembly fieldLoc kind */ typedef enum ArmEncodingKind { kFmtUnused, kFmtBitBlt, /* Bit string using end/start */ kFmtDfp, /* Double FP reg */ kFmtSfp, /* Single FP reg */ kFmtModImm, /* Shifted 8-bit immed using [26,14..12,7..0] */ kFmtImm16, /* Zero-extended immed using [26,19..16,14..12,7..0] */ kFmtImm6, /* Encoded branch target using [9,7..3]0 */ kFmtImm12, /* Zero-extended immediate using [26,14..12,7..0] */ kFmtShift, /* Shift descriptor, [14..12,7..4] */ kFmtLsb, /* least significant bit using [14..12][7..6] */ kFmtBWidth, /* bit-field width, encoded as width-1 */ kFmtShift5, /* Shift count, [14..12,7..6] */ kFmtBrOffset, /* Signed extended [26,11,13,21-16,10-0]:0 */ kFmtFPImm, /* Encoded floating point immediate */ } ArmEncodingKind; /* Struct used to define the snippet positions for each Thumb opcode */ typedef struct ArmEncodingMap { u4 skeleton; struct { ArmEncodingKind kind; int end; /* end for kFmtBitBlt, 1-bit slice end for FP regs */ int start; /* start for kFmtBitBlt, 4-bit slice end for FP regs */ } fieldLoc[4]; ArmOpCode opCode; int flags; char *name; char* fmt; int size; } ArmEncodingMap; /* Keys for target-specific scheduling and other optimization hints */ typedef enum ArmTargetOptHints { kMaxHoistDistance, } ArmTargetOptHints; extern ArmEncodingMap EncodingMap[kArmLast]; /* * Each instance of this struct holds a pseudo or real LIR instruction: * - pseudo ones (eg labels and marks) and will be discarded by the assembler. * - real ones will be assembled into Thumb instructions. * * Machine resources are encoded into a 64-bit vector, where the encodings are * as following: * - [ 0..15]: general purpose registers including PC, SP, and LR * - [16..47]: floating-point registers where d0 is expanded to s[01] and s0 * starts at bit 16 * - [48]: IT block * - [49]: integer condition code * - [50]: floatint-point status word */ typedef struct ArmLIR { LIR generic; ArmOpCode opCode; int operands[4]; // [0..3] = [dest, src1, src2, extra] bool isNop; // LIR is optimized away bool branchInsertSV;// mark for insertion of branch before this instruction, // used to identify mem ops for self verification mode int age; // default is 0, set lazily by the optimizer int size; // 16-bit unit size (1 for thumb, 1 or 2 for thumb2) int aliasInfo; // For Dalvik register access & litpool disambiguation u8 useMask; // Resource mask for use u8 defMask; // Resource mask for def } ArmLIR; /* Init values when a predicted chain is initially assembled */ /* E7FE is branch to self */ #define PREDICTED_CHAIN_BX_PAIR_INIT 0xe7fe /* Utility macros to traverse the LIR/ArmLIR list */ #define NEXT_LIR(lir) ((ArmLIR *) lir->generic.next) #define PREV_LIR(lir) ((ArmLIR *) lir->generic.prev) #define NEXT_LIR_LVALUE(lir) (lir)->generic.next #define PREV_LIR_LVALUE(lir) (lir)->generic.prev #define CHAIN_CELL_OFFSET_TAG 0xcdab #define CHAIN_CELL_NORMAL_SIZE 12 #define CHAIN_CELL_PREDICTED_SIZE 16 #endif /* _DALVIK_VM_COMPILER_CODEGEN_ARM_ARMLIR_H */