//===- subzero/src/IceTargetLoweringX8664Traits.h - x86-64 traits -*- C++ -*-=// // // The Subzero Code Generator // // This file is distributed under the University of Illinois Open Source // License. See LICENSE.TXT for details. // //===----------------------------------------------------------------------===// /// /// \file /// \brief Declares the X8664 Target Lowering Traits. /// //===----------------------------------------------------------------------===// #ifndef SUBZERO_SRC_ICETARGETLOWERINGX8664TRAITS_H #define SUBZERO_SRC_ICETARGETLOWERINGX8664TRAITS_H #include "IceAssembler.h" #include "IceConditionCodesX8664.h" #include "IceDefs.h" #include "IceInst.h" #include "IceInstX8664.def" #include "IceOperand.h" #include "IceRegistersX8664.h" #include "IceTargetLowering.h" #include "IceTargetLoweringX8664.def" #include "IceTargetLoweringX86RegClass.h" #include <array> #include <initializer_list> namespace Ice { namespace X8664 { using namespace ::Ice::X86; template <class TraitsType> class AssemblerX86Base; template <class TraitsType> struct Insts; template <class TraitsType> class TargetX86Base; class TargetX8664; struct TargetX8664Traits { //---------------------------------------------------------------------------- // ______ ______ __ __ // /\ __ \/\ ___\/\ "-./ \ // \ \ __ \ \___ \ \ \-./\ \ // \ \_\ \_\/\_____\ \_\ \ \_\ // \/_/\/_/\/_____/\/_/ \/_/ // //---------------------------------------------------------------------------- static constexpr ::Ice::Assembler::AssemblerKind AsmKind = ::Ice::Assembler::Asm_X8664; static constexpr bool Is64Bit = true; static constexpr bool HasPopa = false; static constexpr bool HasPusha = false; static constexpr bool UsesX87 = false; static constexpr ::Ice::RegX8664::GPRRegister Last8BitGPR = ::Ice::RegX8664::GPRRegister::Encoded_Reg_r15d; enum ScaleFactor { TIMES_1 = 0, TIMES_2 = 1, TIMES_4 = 2, TIMES_8 = 3 }; using GPRRegister = ::Ice::RegX8664::GPRRegister; using ByteRegister = ::Ice::RegX8664::ByteRegister; using XmmRegister = ::Ice::RegX8664::XmmRegister; using Cond = ::Ice::CondX8664; using RegisterSet = ::Ice::RegX8664; static constexpr RegisterSet::AllRegisters StackPtr = RegX8664::Reg_rsp; static constexpr RegisterSet::AllRegisters FramePtr = RegX8664::Reg_rbp; static constexpr GPRRegister Encoded_Reg_Accumulator = RegX8664::Encoded_Reg_eax; static constexpr GPRRegister Encoded_Reg_Counter = RegX8664::Encoded_Reg_ecx; static constexpr FixupKind FK_PcRel = llvm::ELF::R_X86_64_PC32; static constexpr FixupKind FK_Abs = llvm::ELF::R_X86_64_32S; static constexpr FixupKind FK_Gotoff = llvm::ELF::R_X86_64_GOTOFF64; static constexpr FixupKind FK_GotPC = llvm::ELF::R_X86_64_GOTPC32; class Operand { public: enum RexBits { RexNone = 0x00, RexBase = 0x40, RexW = RexBase | (1 << 3), RexR = RexBase | (1 << 2), RexX = RexBase | (1 << 1), RexB = RexBase | (1 << 0), }; protected: // Needed by subclass Address. Operand() = default; public: Operand(const Operand &) = default; Operand(Operand &&) = default; Operand &operator=(const Operand &) = default; Operand &operator=(Operand &&) = default; uint8_t mod() const { return (encoding_at(0) >> 6) & 3; } uint8_t rexX() const { return (rex_ & RexX) != RexX ? RexNone : RexX; } uint8_t rexB() const { return (rex_ & RexB) != RexB ? RexNone : RexB; } GPRRegister rm() const { return static_cast<GPRRegister>((rexB() != 0 ? 0x08 : 0) | (encoding_at(0) & 7)); } ScaleFactor scale() const { return static_cast<ScaleFactor>((encoding_at(1) >> 6) & 3); } GPRRegister index() const { return static_cast<GPRRegister>((rexX() != 0 ? 0x08 : 0) | ((encoding_at(1) >> 3) & 7)); } GPRRegister base() const { return static_cast<GPRRegister>((rexB() != 0 ? 0x08 : 0) | (encoding_at(1) & 7)); } int8_t disp8() const { assert(length_ >= 2); return static_cast<int8_t>(encoding_[length_ - 1]); } AssemblerFixup *fixup() const { return fixup_; } protected: void SetModRM(int mod, GPRRegister rm) { assert((mod & ~3) == 0); encoding_[0] = (mod << 6) | (rm & 0x07); rex_ = (rm & 0x08) ? RexB : RexNone; length_ = 1; } void SetSIB(ScaleFactor scale, GPRRegister index, GPRRegister base) { assert(length_ == 1); assert((scale & ~3) == 0); encoding_[1] = (scale << 6) | ((index & 0x07) << 3) | (base & 0x07); rex_ = ((base & 0x08) ? RexB : RexNone) | ((index & 0x08) ? RexX : RexNone); length_ = 2; } void SetDisp8(int8_t disp) { assert(length_ == 1 || length_ == 2); encoding_[length_++] = static_cast<uint8_t>(disp); } void SetDisp32(int32_t disp) { assert(length_ == 1 || length_ == 2); intptr_t disp_size = sizeof(disp); memmove(&encoding_[length_], &disp, disp_size); length_ += disp_size; } void SetFixup(AssemblerFixup *fixup) { fixup_ = fixup; } private: AssemblerFixup *fixup_ = nullptr; uint8_t rex_ = 0; uint8_t encoding_[6]; uint8_t length_ = 0; explicit Operand(GPRRegister reg) : fixup_(nullptr) { SetModRM(3, reg); } /// Get the operand encoding byte at the given index. uint8_t encoding_at(intptr_t index) const { assert(index >= 0 && index < length_); return encoding_[index]; } /// Returns whether or not this operand is really the given register in /// disguise. Used from the assembler to generate better encodings. bool IsRegister(GPRRegister reg) const { return ((encoding_[0] & 0xF8) == 0xC0) // Addressing mode is register only. && (rm() == reg); // Register codes match. } friend class AssemblerX86Base<TargetX8664Traits>; }; class Address : public Operand { Address() = default; public: Address(const Address &) = default; Address(Address &&) = default; Address &operator=(const Address &) = default; Address &operator=(Address &&) = default; Address(GPRRegister Base, int32_t Disp, AssemblerFixup *Fixup) { if (Fixup == nullptr && Disp == 0 && (Base & 7) != RegX8664::Encoded_Reg_rbp) { SetModRM(0, Base); if ((Base & 7) == RegX8664::Encoded_Reg_rsp) SetSIB(TIMES_1, RegX8664::Encoded_Reg_rsp, Base); } else if (Fixup == nullptr && Utils::IsInt(8, Disp)) { SetModRM(1, Base); if ((Base & 7) == RegX8664::Encoded_Reg_rsp) SetSIB(TIMES_1, RegX8664::Encoded_Reg_rsp, Base); SetDisp8(Disp); } else { SetModRM(2, Base); if ((Base & 7) == RegX8664::Encoded_Reg_rsp) SetSIB(TIMES_1, RegX8664::Encoded_Reg_rsp, Base); SetDisp32(Disp); if (Fixup) SetFixup(Fixup); } } Address(GPRRegister Index, ScaleFactor Scale, int32_t Disp, AssemblerFixup *Fixup) { assert(Index != RegX8664::Encoded_Reg_rsp); // Illegal addressing mode. SetModRM(0, RegX8664::Encoded_Reg_rsp); SetSIB(Scale, Index, RegX8664::Encoded_Reg_rbp); SetDisp32(Disp); if (Fixup) SetFixup(Fixup); } Address(GPRRegister Base, GPRRegister Index, ScaleFactor Scale, int32_t Disp, AssemblerFixup *Fixup) { assert(Index != RegX8664::Encoded_Reg_rsp); // Illegal addressing mode. if (Fixup == nullptr && Disp == 0 && (Base & 7) != RegX8664::Encoded_Reg_rbp) { SetModRM(0, RegX8664::Encoded_Reg_rsp); SetSIB(Scale, Index, Base); } else if (Fixup == nullptr && Utils::IsInt(8, Disp)) { SetModRM(1, RegX8664::Encoded_Reg_rsp); SetSIB(Scale, Index, Base); SetDisp8(Disp); } else { SetModRM(2, RegX8664::Encoded_Reg_rsp); SetSIB(Scale, Index, Base); SetDisp32(Disp); if (Fixup) SetFixup(Fixup); } } /// Generate a RIP-relative address expression on x86-64. static Address RipRelative(RelocOffsetT Offset, AssemblerFixup *Fixup) { assert(Fixup != nullptr); assert(Fixup->kind() == FK_PcRel); Address NewAddress; NewAddress.SetModRM(0x0, RegX8664::Encoded_Reg_rbp); // Use the Offset in the displacement for now. If we decide to process // fixups later, we'll need to patch up the emitted displacement. NewAddress.SetDisp32(Offset); if (Fixup) NewAddress.SetFixup(Fixup); return NewAddress; } /// Generate an absolute address. static Address Absolute(RelocOffsetT Addr) { Address NewAddress; NewAddress.SetModRM(0x0, RegX8664::Encoded_Reg_rsp); static constexpr ScaleFactor NoScale = TIMES_1; NewAddress.SetSIB(NoScale, RegX8664::Encoded_Reg_rsp, RegX8664::Encoded_Reg_rbp); NewAddress.SetDisp32(Addr); return NewAddress; } static Address ofConstPool(Assembler *Asm, const Constant *Imm) { // TODO(jpp): ??? AssemblerFixup *Fixup = Asm->createFixup(FK_Abs, Imm); const RelocOffsetT Offset = 4; return Address::RipRelative(Offset, Fixup); } }; //---------------------------------------------------------------------------- // __ ______ __ __ ______ ______ __ __ __ ______ // /\ \ /\ __ \/\ \ _ \ \/\ ___\/\ == \/\ \/\ "-.\ \/\ ___\ // \ \ \___\ \ \/\ \ \ \/ ".\ \ \ __\\ \ __<\ \ \ \ \-. \ \ \__ \ // \ \_____\ \_____\ \__/".~\_\ \_____\ \_\ \_\ \_\ \_\\"\_\ \_____\ // \/_____/\/_____/\/_/ \/_/\/_____/\/_/ /_/\/_/\/_/ \/_/\/_____/ // //---------------------------------------------------------------------------- enum InstructionSet { Begin, // SSE2 is the PNaCl baseline instruction set. SSE2 = Begin, SSE4_1, End }; static const char *TargetName; static constexpr Type WordType = IceType_i64; static const char *getRegName(RegNumT RegNum) { static const char *const RegNames[RegisterSet::Reg_NUM] = { #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ name, REGX8664_TABLE #undef X }; RegNum.assertIsValid(); return RegNames[RegNum]; } static GPRRegister getEncodedGPR(RegNumT RegNum) { static const GPRRegister GPRRegs[RegisterSet::Reg_NUM] = { #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ GPRRegister(isGPR ? encode : GPRRegister::Encoded_Not_GPR), REGX8664_TABLE #undef X }; RegNum.assertIsValid(); assert(GPRRegs[RegNum] != GPRRegister::Encoded_Not_GPR); return GPRRegs[RegNum]; } static ByteRegister getEncodedByteReg(RegNumT RegNum) { static const ByteRegister ByteRegs[RegisterSet::Reg_NUM] = { #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ ByteRegister(is8 ? encode : ByteRegister::Encoded_Not_ByteReg), REGX8664_TABLE #undef X }; RegNum.assertIsValid(); assert(ByteRegs[RegNum] != ByteRegister::Encoded_Not_ByteReg); return ByteRegs[RegNum]; } static XmmRegister getEncodedXmm(RegNumT RegNum) { static const XmmRegister XmmRegs[RegisterSet::Reg_NUM] = { #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ XmmRegister(isXmm ? encode : XmmRegister::Encoded_Not_Xmm), REGX8664_TABLE #undef X }; RegNum.assertIsValid(); assert(XmmRegs[RegNum] != XmmRegister::Encoded_Not_Xmm); return XmmRegs[RegNum]; } static uint32_t getEncoding(RegNumT RegNum) { static const uint32_t Encoding[RegisterSet::Reg_NUM] = { #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ encode, REGX8664_TABLE #undef X }; RegNum.assertIsValid(); return Encoding[RegNum]; } static inline RegNumT getBaseReg(RegNumT RegNum) { static const RegNumT BaseRegs[RegisterSet::Reg_NUM] = { #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ RegisterSet::base, REGX8664_TABLE #undef X }; RegNum.assertIsValid(); return BaseRegs[RegNum]; } private: static RegNumT getFirstGprForType(Type Ty) { switch (Ty) { default: llvm_unreachable("Invalid type for GPR."); case IceType_i1: case IceType_i8: return RegisterSet::Reg_al; case IceType_i16: return RegisterSet::Reg_ax; case IceType_i32: return RegisterSet::Reg_eax; case IceType_i64: return RegisterSet::Reg_rax; } } public: static RegNumT getGprForType(Type Ty, RegNumT RegNum) { assert(RegNum.hasValue()); if (!isScalarIntegerType(Ty)) { return RegNum; } assert(Ty == IceType_i1 || Ty == IceType_i8 || Ty == IceType_i16 || Ty == IceType_i32 || Ty == IceType_i64); if (RegNum == RegisterSet::Reg_ah) { assert(Ty == IceType_i8); return RegNum; } assert(RegNum != RegisterSet::Reg_bh); assert(RegNum != RegisterSet::Reg_ch); assert(RegNum != RegisterSet::Reg_dh); const RegNumT FirstGprForType = getFirstGprForType(Ty); switch (RegNum) { default: llvm::report_fatal_error("Unknown register."); #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ case RegisterSet::val: { \ if (!isGPR) \ return RegisterSet::val; \ assert((is64) || (is32) || (is16) || (is8) || \ getBaseReg(RegisterSet::val) == RegisterSet::Reg_rsp); \ constexpr RegisterSet::AllRegisters FirstGprWithRegNumSize = \ ((is64) || RegisterSet::val == RegisterSet::Reg_rsp) \ ? RegisterSet::Reg_rax \ : (((is32) || RegisterSet::val == RegisterSet::Reg_esp) \ ? RegisterSet::Reg_eax \ : (((is16) || RegisterSet::val == RegisterSet::Reg_sp) \ ? RegisterSet::Reg_ax \ : RegisterSet::Reg_al)); \ const auto NewRegNum = \ RegNumT::fixme(RegNum - FirstGprWithRegNumSize + FirstGprForType); \ assert(getBaseReg(RegNum) == getBaseReg(NewRegNum) && \ "Error involving " #val); \ return NewRegNum; \ } REGX8664_TABLE #undef X } } private: /// SizeOf is used to obtain the size of an initializer list as a constexpr /// expression. This is only needed until our C++ library is updated to /// C++ 14 -- which defines constexpr members to std::initializer_list. class SizeOf { SizeOf(const SizeOf &) = delete; SizeOf &operator=(const SizeOf &) = delete; public: constexpr SizeOf() : Size(0) {} template <typename... T> explicit constexpr SizeOf(T...) : Size(length<T...>::value) {} constexpr SizeT size() const { return Size; } private: template <typename T, typename... U> struct length { static constexpr std::size_t value = 1 + length<U...>::value; }; template <typename T> struct length<T> { static constexpr std::size_t value = 1; }; const std::size_t Size; }; public: static void initRegisterSet( const ::Ice::ClFlags &Flags, std::array<SmallBitVector, RCX86_NUM> *TypeToRegisterSet, std::array<SmallBitVector, RegisterSet::Reg_NUM> *RegisterAliases) { SmallBitVector IntegerRegistersI64(RegisterSet::Reg_NUM); SmallBitVector IntegerRegistersI32(RegisterSet::Reg_NUM); SmallBitVector IntegerRegistersI16(RegisterSet::Reg_NUM); SmallBitVector IntegerRegistersI8(RegisterSet::Reg_NUM); SmallBitVector FloatRegisters(RegisterSet::Reg_NUM); SmallBitVector VectorRegisters(RegisterSet::Reg_NUM); SmallBitVector Trunc64To8Registers(RegisterSet::Reg_NUM); SmallBitVector Trunc32To8Registers(RegisterSet::Reg_NUM); SmallBitVector Trunc16To8Registers(RegisterSet::Reg_NUM); SmallBitVector Trunc8RcvrRegisters(RegisterSet::Reg_NUM); SmallBitVector AhRcvrRegisters(RegisterSet::Reg_NUM); SmallBitVector InvalidRegisters(RegisterSet::Reg_NUM); static constexpr struct { uint16_t Val; unsigned IsReservedWhenSandboxing : 1; unsigned Is64 : 1; unsigned Is32 : 1; unsigned Is16 : 1; unsigned Is8 : 1; unsigned IsXmm : 1; unsigned Is64To8 : 1; unsigned Is32To8 : 1; unsigned Is16To8 : 1; unsigned IsTrunc8Rcvr : 1; unsigned IsAhRcvr : 1; #define NUM_ALIASES_BITS 2 SizeT NumAliases : (NUM_ALIASES_BITS + 1); uint16_t Aliases[1 << NUM_ALIASES_BITS]; #undef NUM_ALIASES_BITS } X8664RegTable[RegisterSet::Reg_NUM] = { #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ { \ RegisterSet::val, sboxres, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, (SizeOf aliases).size(), aliases, \ } \ , REGX8664_TABLE #undef X }; const bool NeedSandboxing = Flags.getUseSandboxing(); for (SizeT ii = 0; ii < llvm::array_lengthof(X8664RegTable); ++ii) { const auto &Entry = X8664RegTable[ii]; // Even though the register is disabled for register allocation, it might // still be used by the Target Lowering (e.g., base pointer), so the // register alias table still needs to be defined. (*RegisterAliases)[Entry.Val].resize(RegisterSet::Reg_NUM); for (int J = 0; J < Entry.NumAliases; ++J) { SizeT Alias = Entry.Aliases[J]; assert(!(*RegisterAliases)[Entry.Val][Alias] && "Duplicate alias"); (*RegisterAliases)[Entry.Val].set(Alias); } (*RegisterAliases)[Entry.Val].set(Entry.Val); const bool DisabledRegister = NeedSandboxing && Entry.IsReservedWhenSandboxing; if (DisabledRegister) { continue; } (IntegerRegistersI64)[Entry.Val] = Entry.Is64; (IntegerRegistersI32)[Entry.Val] = Entry.Is32; (IntegerRegistersI16)[Entry.Val] = Entry.Is16; (IntegerRegistersI8)[Entry.Val] = Entry.Is8; (FloatRegisters)[Entry.Val] = Entry.IsXmm; (VectorRegisters)[Entry.Val] = Entry.IsXmm; (Trunc64To8Registers)[Entry.Val] = Entry.Is64To8; (Trunc32To8Registers)[Entry.Val] = Entry.Is32To8; (Trunc16To8Registers)[Entry.Val] = Entry.Is16To8; (Trunc8RcvrRegisters)[Entry.Val] = Entry.IsTrunc8Rcvr; (AhRcvrRegisters)[Entry.Val] = Entry.IsAhRcvr; } (*TypeToRegisterSet)[RC_void] = InvalidRegisters; (*TypeToRegisterSet)[RC_i1] = IntegerRegistersI8; (*TypeToRegisterSet)[RC_i8] = IntegerRegistersI8; (*TypeToRegisterSet)[RC_i16] = IntegerRegistersI16; (*TypeToRegisterSet)[RC_i32] = IntegerRegistersI32; (*TypeToRegisterSet)[RC_i64] = IntegerRegistersI64; (*TypeToRegisterSet)[RC_f32] = FloatRegisters; (*TypeToRegisterSet)[RC_f64] = FloatRegisters; (*TypeToRegisterSet)[RC_v4i1] = VectorRegisters; (*TypeToRegisterSet)[RC_v8i1] = VectorRegisters; (*TypeToRegisterSet)[RC_v16i1] = VectorRegisters; (*TypeToRegisterSet)[RC_v16i8] = VectorRegisters; (*TypeToRegisterSet)[RC_v8i16] = VectorRegisters; (*TypeToRegisterSet)[RC_v4i32] = VectorRegisters; (*TypeToRegisterSet)[RC_v4f32] = VectorRegisters; (*TypeToRegisterSet)[RCX86_Is64To8] = Trunc64To8Registers; (*TypeToRegisterSet)[RCX86_Is32To8] = Trunc32To8Registers; (*TypeToRegisterSet)[RCX86_Is16To8] = Trunc16To8Registers; (*TypeToRegisterSet)[RCX86_IsTrunc8Rcvr] = Trunc8RcvrRegisters; (*TypeToRegisterSet)[RCX86_IsAhRcvr] = AhRcvrRegisters; } static SmallBitVector getRegisterSet(const ::Ice::ClFlags &Flags, TargetLowering::RegSetMask Include, TargetLowering::RegSetMask Exclude) { SmallBitVector Registers(RegisterSet::Reg_NUM); const bool NeedSandboxing = Flags.getUseSandboxing(); #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ if (!NeedSandboxing || !(sboxres)) { \ if (scratch && (Include & ::Ice::TargetLowering::RegSet_CallerSave)) \ Registers[RegisterSet::val] = true; \ if (preserved && (Include & ::Ice::TargetLowering::RegSet_CalleeSave)) \ Registers[RegisterSet::val] = true; \ if (stackptr && (Include & ::Ice::TargetLowering::RegSet_StackPointer)) \ Registers[RegisterSet::val] = true; \ if (frameptr && (Include & ::Ice::TargetLowering::RegSet_FramePointer)) \ Registers[RegisterSet::val] = true; \ if (scratch && (Exclude & ::Ice::TargetLowering::RegSet_CallerSave)) \ Registers[RegisterSet::val] = false; \ if (preserved && (Exclude & ::Ice::TargetLowering::RegSet_CalleeSave)) \ Registers[RegisterSet::val] = false; \ if (stackptr && (Exclude & ::Ice::TargetLowering::RegSet_StackPointer)) \ Registers[RegisterSet::val] = false; \ if (frameptr && (Exclude & ::Ice::TargetLowering::RegSet_FramePointer)) \ Registers[RegisterSet::val] = false; \ } REGX8664_TABLE #undef X return Registers; } static void makeRandomRegisterPermutation( Cfg *Func, llvm::SmallVectorImpl<RegNumT> &Permutation, const SmallBitVector &ExcludeRegisters, uint64_t Salt) { // TODO(stichnot): Declaring Permutation this way loses type/size // information. Fix this in conjunction with the caller-side TODO. assert(Permutation.size() >= RegisterSet::Reg_NUM); // Expected upper bound on the number of registers in a single equivalence // class. For x86-64, this would comprise the 16 XMM registers. This is // for performance, not correctness. static const unsigned MaxEquivalenceClassSize = 8; using RegisterList = llvm::SmallVector<RegNumT, MaxEquivalenceClassSize>; using EquivalenceClassMap = std::map<uint32_t, RegisterList>; EquivalenceClassMap EquivalenceClasses; SizeT NumShuffled = 0, NumPreserved = 0; // Build up the equivalence classes of registers by looking at the register // properties as well as whether the registers should be explicitly excluded // from shuffling. #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ if (ExcludeRegisters[RegisterSet::val]) { \ /* val stays the same in the resulting permutation. */ \ Permutation[RegisterSet::val] = RegisterSet::val; \ ++NumPreserved; \ } else { \ uint32_t AttrKey = 0; \ uint32_t Index = 0; \ /* Combine relevant attributes into an equivalence class key. */ \ Index |= (scratch << (AttrKey++)); \ Index |= (preserved << (AttrKey++)); \ Index |= (is8 << (AttrKey++)); \ Index |= (is16 << (AttrKey++)); \ Index |= (is32 << (AttrKey++)); \ Index |= (is64 << (AttrKey++)); \ Index |= (isXmm << (AttrKey++)); \ Index |= (is16To8 << (AttrKey++)); \ Index |= (is32To8 << (AttrKey++)); \ Index |= (is64To8 << (AttrKey++)); \ Index |= (isTrunc8Rcvr << (AttrKey++)); \ /* val is assigned to an equivalence class based on its properties. */ \ EquivalenceClasses[Index].push_back(RegisterSet::val); \ } REGX8664_TABLE #undef X // Create a random number generator for regalloc randomization. RandomNumberGenerator RNG(getFlags().getRandomSeed(), RPE_RegAllocRandomization, Salt); RandomNumberGeneratorWrapper RNGW(RNG); // Shuffle the resulting equivalence classes. for (auto I : EquivalenceClasses) { const RegisterList &List = I.second; RegisterList Shuffled(List); RandomShuffle(Shuffled.begin(), Shuffled.end(), RNGW); for (size_t SI = 0, SE = Shuffled.size(); SI < SE; ++SI) { Permutation[List[SI]] = Shuffled[SI]; ++NumShuffled; } } assert(NumShuffled + NumPreserved == RegisterSet::Reg_NUM); if (Func->isVerbose(IceV_Random)) { OstreamLocker L(Func->getContext()); Ostream &Str = Func->getContext()->getStrDump(); Str << "Register equivalence classes:\n"; for (auto I : EquivalenceClasses) { Str << "{"; const RegisterList &List = I.second; bool First = true; for (RegNumT Register : List) { if (!First) Str << " "; First = false; Str << getRegName(Register); } Str << "}\n"; } } } static RegNumT getRaxOrDie() { return RegisterSet::Reg_rax; } static RegNumT getRdxOrDie() { return RegisterSet::Reg_rdx; } #if defined(SUBZERO_USE_MICROSOFT_ABI) // Microsoft x86-64 calling convention: // // * The first four arguments of vector/fp type, regardless of their // position relative to the other arguments in the argument list, are placed // in registers %xmm0 - %xmm3. // // * The first four arguments of integer types, regardless of their position // relative to the other arguments in the argument list, are placed in // registers %rcx, %rdx, %r8, and %r9. /// The maximum number of arguments to pass in XMM registers static constexpr uint32_t X86_MAX_XMM_ARGS = 4; /// The maximum number of arguments to pass in GPR registers static constexpr uint32_t X86_MAX_GPR_ARGS = 4; static RegNumT getRegisterForGprArgNum(Type Ty, uint32_t ArgNum) { if (ArgNum >= X86_MAX_GPR_ARGS) { return RegNumT(); } static const RegisterSet::AllRegisters GprForArgNum[] = { RegisterSet::Reg_rcx, RegisterSet::Reg_rdx, RegisterSet::Reg_r8, RegisterSet::Reg_r9, }; static_assert(llvm::array_lengthof(GprForArgNum) == X86_MAX_GPR_ARGS, "Mismatch between MAX_GPR_ARGS and GprForArgNum."); assert(Ty == IceType_i64 || Ty == IceType_i32); return getGprForType(Ty, GprForArgNum[ArgNum]); } #else // System V x86-64 calling convention: // // * The first eight arguments of vector/fp type, regardless of their // position relative to the other arguments in the argument list, are placed // in registers %xmm0 - %xmm7. // // * The first six arguments of integer types, regardless of their position // relative to the other arguments in the argument list, are placed in // registers %rdi, %rsi, %rdx, %rcx, %r8, and %r9. // // This intends to match the section "Function Calling Sequence" of the // document "System V Application Binary Interface." /// The maximum number of arguments to pass in XMM registers static constexpr uint32_t X86_MAX_XMM_ARGS = 8; /// The maximum number of arguments to pass in GPR registers static constexpr uint32_t X86_MAX_GPR_ARGS = 6; /// Get the register for a given argument slot in the GPRs. static RegNumT getRegisterForGprArgNum(Type Ty, uint32_t ArgNum) { if (ArgNum >= X86_MAX_GPR_ARGS) { return RegNumT(); } static const RegisterSet::AllRegisters GprForArgNum[] = { RegisterSet::Reg_rdi, RegisterSet::Reg_rsi, RegisterSet::Reg_rdx, RegisterSet::Reg_rcx, RegisterSet::Reg_r8, RegisterSet::Reg_r9, }; static_assert(llvm::array_lengthof(GprForArgNum) == X86_MAX_GPR_ARGS, "Mismatch between MAX_GPR_ARGS and GprForArgNum."); assert(Ty == IceType_i64 || Ty == IceType_i32); return getGprForType(Ty, GprForArgNum[ArgNum]); } #endif /// Whether scalar floating point arguments are passed in XMM registers static constexpr bool X86_PASS_SCALAR_FP_IN_XMM = true; /// Get the register for a given argument slot in the XMM registers. static RegNumT getRegisterForXmmArgNum(uint32_t ArgNum) { // TODO(sehr): Change to use the CCArg technique used in ARM32. static_assert(RegisterSet::Reg_xmm0 + 1 == RegisterSet::Reg_xmm1, "Inconsistency between XMM register numbers and ordinals"); if (ArgNum >= X86_MAX_XMM_ARGS) { return RegNumT(); } return RegNumT::fixme(RegisterSet::Reg_xmm0 + ArgNum); } /// The number of bits in a byte static constexpr uint32_t X86_CHAR_BIT = 8; /// Stack alignment. This is defined in IceTargetLoweringX8664.cpp because it /// is used as an argument to std::max(), and the default std::less<T> has an /// operator(T const&, T const&) which requires this member to have an /// address. static const uint32_t X86_STACK_ALIGNMENT_BYTES; /// Size of the return address on the stack static constexpr uint32_t X86_RET_IP_SIZE_BYTES = 8; /// The number of different NOP instructions static constexpr uint32_t X86_NUM_NOP_VARIANTS = 5; /// \name Limits for unrolling memory intrinsics. /// @{ static constexpr uint32_t MEMCPY_UNROLL_LIMIT = 8; static constexpr uint32_t MEMMOVE_UNROLL_LIMIT = 8; static constexpr uint32_t MEMSET_UNROLL_LIMIT = 8; /// @} /// Value is in bytes. Return Value adjusted to the next highest multiple of /// the stack alignment. static uint32_t applyStackAlignment(uint32_t Value) { return Utils::applyAlignment(Value, X86_STACK_ALIGNMENT_BYTES); } /// Return the type which the elements of the vector have in the X86 /// representation of the vector. static Type getInVectorElementType(Type Ty) { assert(isVectorType(Ty)); assert(Ty < TableTypeX8664AttributesSize); return TableTypeX8664Attributes[Ty].InVectorElementType; } // Note: The following data structures are defined in // IceTargetLoweringX8664.cpp. /// The following table summarizes the logic for lowering the fcmp /// instruction. There is one table entry for each of the 16 conditions. /// /// The first four columns describe the case when the operands are floating /// point scalar values. A comment in lowerFcmp() describes the lowering /// template. In the most general case, there is a compare followed by two /// conditional branches, because some fcmp conditions don't map to a single /// x86 conditional branch. However, in many cases it is possible to swap the /// operands in the comparison and have a single conditional branch. Since /// it's quite tedious to validate the table by hand, good execution tests are /// helpful. /// /// The last two columns describe the case when the operands are vectors of /// floating point values. For most fcmp conditions, there is a clear mapping /// to a single x86 cmpps instruction variant. Some fcmp conditions require /// special code to handle and these are marked in the table with a /// Cmpps_Invalid predicate. /// {@ static const struct TableFcmpType { uint32_t Default; bool SwapScalarOperands; Cond::BrCond C1, C2; bool SwapVectorOperands; Cond::CmppsCond Predicate; } TableFcmp[]; static const size_t TableFcmpSize; /// @} /// The following table summarizes the logic for lowering the icmp instruction /// for i32 and narrower types. Each icmp condition has a clear mapping to an /// x86 conditional branch instruction. /// {@ static const struct TableIcmp32Type { Cond::BrCond Mapping; } TableIcmp32[]; static const size_t TableIcmp32Size; /// @} /// The following table summarizes the logic for lowering the icmp instruction /// for the i64 type. For Eq and Ne, two separate 32-bit comparisons and /// conditional branches are needed. For the other conditions, three separate /// conditional branches are needed. /// {@ static const struct TableIcmp64Type { Cond::BrCond C1, C2, C3; } TableIcmp64[]; static const size_t TableIcmp64Size; /// @} static Cond::BrCond getIcmp32Mapping(InstIcmp::ICond Cond) { assert(Cond < TableIcmp32Size); return TableIcmp32[Cond].Mapping; } static const struct TableTypeX8664AttributesType { Type InVectorElementType; } TableTypeX8664Attributes[]; static const size_t TableTypeX8664AttributesSize; //---------------------------------------------------------------------------- // __ __ __ ______ ______ // /\ \/\ "-.\ \/\ ___\/\__ _\ // \ \ \ \ \-. \ \___ \/_/\ \/ // \ \_\ \_\\"\_\/\_____\ \ \_\ // \/_/\/_/ \/_/\/_____/ \/_/ // //---------------------------------------------------------------------------- using Traits = TargetX8664Traits; using Insts = ::Ice::X8664::Insts<Traits>; using TargetLowering = ::Ice::X8664::TargetX86Base<Traits>; using ConcreteTarget = ::Ice::X8664::TargetX8664; using Assembler = ::Ice::X8664::AssemblerX86Base<Traits>; /// X86Operand extends the Operand hierarchy. Its subclasses are X86OperandMem /// and VariableSplit. class X86Operand : public ::Ice::Operand { X86Operand() = delete; X86Operand(const X86Operand &) = delete; X86Operand &operator=(const X86Operand &) = delete; public: enum OperandKindX8664 { k__Start = ::Ice::Operand::kTarget, kMem, kSplit }; using ::Ice::Operand::dump; void dump(const Cfg *, Ostream &Str) const override; protected: X86Operand(OperandKindX8664 Kind, Type Ty) : Operand(static_cast<::Ice::Operand::OperandKind>(Kind), Ty) {} }; /// X86OperandMem represents the m64 addressing mode, with optional base and /// index registers, a constant offset, and a fixed shift value for the index /// register. class X86OperandMem : public X86Operand { X86OperandMem() = delete; X86OperandMem(const X86OperandMem &) = delete; X86OperandMem &operator=(const X86OperandMem &) = delete; public: enum SegmentRegisters { DefaultSegment = -1, SegReg_NUM }; static X86OperandMem * create(Cfg *Func, Type Ty, Variable *Base, Constant *Offset, Variable *Index = nullptr, uint16_t Shift = 0, SegmentRegisters SegmentRegister = DefaultSegment, bool IsRebased = false) { assert(SegmentRegister == DefaultSegment); (void)SegmentRegister; return new (Func->allocate<X86OperandMem>()) X86OperandMem(Func, Ty, Base, Offset, Index, Shift, IsRebased); } static X86OperandMem *create(Cfg *Func, Type Ty, Variable *Base, Constant *Offset, bool IsRebased) { constexpr Variable *NoIndex = nullptr; constexpr uint16_t NoShift = 0; return new (Func->allocate<X86OperandMem>()) X86OperandMem(Func, Ty, Base, Offset, NoIndex, NoShift, IsRebased); } Variable *getBase() const { return Base; } Constant *getOffset() const { return Offset; } Variable *getIndex() const { return Index; } uint16_t getShift() const { return Shift; } SegmentRegisters getSegmentRegister() const { return DefaultSegment; } void emitSegmentOverride(Assembler *) const {} bool getIsRebased() const { return IsRebased; } Address toAsmAddress(Assembler *Asm, const Ice::TargetLowering *Target, bool IsLeaAddr = false) const; void emit(const Cfg *Func) const override; using X86Operand::dump; void dump(const Cfg *Func, Ostream &Str) const override; static bool classof(const Operand *Operand) { return Operand->getKind() == static_cast<OperandKind>(kMem); } void setRandomized(bool R) { Randomized = R; } bool getRandomized() const { return Randomized; } private: X86OperandMem(Cfg *Func, Type Ty, Variable *Base, Constant *Offset, Variable *Index, uint16_t Shift, bool IsRebased); Variable *const Base; Constant *const Offset; Variable *const Index; const uint16_t Shift; const bool IsRebased; /// A flag to show if this memory operand is a randomized one. Randomized /// memory operands are generated in /// TargetX86Base::randomizeOrPoolImmediate() bool Randomized = false; }; /// VariableSplit is a way to treat an f64 memory location as a pair of i32 /// locations (Low and High). This is needed for some cases of the Bitcast /// instruction. Since it's not possible for integer registers to access the /// XMM registers and vice versa, the lowering forces the f64 to be spilled to /// the stack and then accesses through the VariableSplit. // TODO(jpp): remove references to VariableSplit from IceInstX86Base as 64bit // targets can natively handle these. class VariableSplit : public X86Operand { VariableSplit() = delete; VariableSplit(const VariableSplit &) = delete; VariableSplit &operator=(const VariableSplit &) = delete; public: enum Portion { Low, High }; static VariableSplit *create(Cfg *Func, Variable *Var, Portion Part) { return new (Func->allocate<VariableSplit>()) VariableSplit(Func, Var, Part); } int32_t getOffset() const { return Part == High ? 4 : 0; } Address toAsmAddress(const Cfg *Func) const; void emit(const Cfg *Func) const override; using X86Operand::dump; void dump(const Cfg *Func, Ostream &Str) const override; static bool classof(const Operand *Operand) { return Operand->getKind() == static_cast<OperandKind>(kSplit); } private: VariableSplit(Cfg *Func, Variable *Var, Portion Part) : X86Operand(kSplit, IceType_i32), Var(Var), Part(Part) { assert(Var->getType() == IceType_f64); Vars = Func->allocateArrayOf<Variable *>(1); Vars[0] = Var; NumVars = 1; } Variable *Var; Portion Part; }; // Note: The following data structures are defined in IceInstX8664.cpp. static const struct InstBrAttributesType { Cond::BrCond Opposite; const char *DisplayString; const char *EmitString; } InstBrAttributes[]; static const struct InstCmppsAttributesType { const char *EmitString; } InstCmppsAttributes[]; static const struct TypeAttributesType { const char *CvtString; // i (integer), s (single FP), d (double FP) const char *SdSsString; // ss, sd, or <blank> const char *PdPsString; // ps, pd, or <blank> const char *SpSdString; // ss, sd, ps, pd, or <blank> const char *IntegralString; // b, w, d, or <blank> const char *UnpackString; // bw, wd, dq, or <blank> const char *PackString; // wb, dw, or <blank> const char *WidthString; // b, w, l, q, or <blank> const char *FldString; // s, l, or <blank> } TypeAttributes[]; }; using Traits = ::Ice::X8664::TargetX8664Traits; } // end of namespace X8664 } // end of namespace Ice #endif // SUBZERO_SRC_ICETARGETLOWERINGX8664TRAITS_H