• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 //===- subzero/src/IceTargetLoweringX8664.cpp - x86-64 lowering -----------===//
2 //
3 //                        The Subzero Code Generator
4 //
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
7 //
8 //===----------------------------------------------------------------------===//
9 ///
10 /// \file
11 /// \brief Implements the TargetLoweringX8664 class, which consists almost
12 /// entirely of the lowering sequence for each high-level instruction.
13 ///
14 //===----------------------------------------------------------------------===//
15 #include "IceTargetLoweringX8664.h"
16 
17 #include "IceDefs.h"
18 #include "IceTargetLoweringX8664Traits.h"
19 
20 namespace X8664 {
createTargetLowering(::Ice::Cfg * Func)21 std::unique_ptr<::Ice::TargetLowering> createTargetLowering(::Ice::Cfg *Func) {
22   return ::Ice::X8664::TargetX8664::create(Func);
23 }
24 
25 std::unique_ptr<::Ice::TargetDataLowering>
createTargetDataLowering(::Ice::GlobalContext * Ctx)26 createTargetDataLowering(::Ice::GlobalContext *Ctx) {
27   return ::Ice::X8664::TargetDataX86<::Ice::X8664::TargetX8664Traits>::create(
28       Ctx);
29 }
30 
31 std::unique_ptr<::Ice::TargetHeaderLowering>
createTargetHeaderLowering(::Ice::GlobalContext * Ctx)32 createTargetHeaderLowering(::Ice::GlobalContext *Ctx) {
33   return ::Ice::X8664::TargetHeaderX86::create(Ctx);
34 }
35 
staticInit(::Ice::GlobalContext * Ctx)36 void staticInit(::Ice::GlobalContext *Ctx) {
37   ::Ice::X8664::TargetX8664::staticInit(Ctx);
38 }
39 
shouldBePooled(const class::Ice::Constant * C)40 bool shouldBePooled(const class ::Ice::Constant *C) {
41   return ::Ice::X8664::TargetX8664::shouldBePooled(C);
42 }
43 
getPointerType()44 ::Ice::Type getPointerType() {
45   return ::Ice::X8664::TargetX8664::getPointerType();
46 }
47 
48 } // end of namespace X8664
49 
50 namespace Ice {
51 namespace X8664 {
52 
53 //------------------------------------------------------------------------------
54 //      ______   ______     ______     __     ______   ______
55 //     /\__  _\ /\  == \   /\  __ \   /\ \   /\__  _\ /\  ___\
56 //     \/_/\ \/ \ \  __<   \ \  __ \  \ \ \  \/_/\ \/ \ \___  \
57 //        \ \_\  \ \_\ \_\  \ \_\ \_\  \ \_\    \ \_\  \/\_____\
58 //         \/_/   \/_/ /_/   \/_/\/_/   \/_/     \/_/   \/_____/
59 //
60 //------------------------------------------------------------------------------
61 const TargetX8664Traits::TableFcmpType TargetX8664Traits::TableFcmp[] = {
62 #define X(val, dflt, swapS, C1, C2, swapV, pred)                               \
63   {                                                                            \
64     dflt, swapS, X8664::Traits::Cond::C1, X8664::Traits::Cond::C2, swapV,      \
65         X8664::Traits::Cond::pred                                              \
66   }                                                                            \
67   ,
68     FCMPX8664_TABLE
69 #undef X
70 };
71 
72 const size_t TargetX8664Traits::TableFcmpSize = llvm::array_lengthof(TableFcmp);
73 
74 const TargetX8664Traits::TableIcmp32Type TargetX8664Traits::TableIcmp32[] = {
75 #define X(val, C_32, C1_64, C2_64, C3_64)                                      \
76   { X8664::Traits::Cond::C_32 }                                                \
77   ,
78     ICMPX8664_TABLE
79 #undef X
80 };
81 
82 const size_t TargetX8664Traits::TableIcmp32Size =
83     llvm::array_lengthof(TableIcmp32);
84 
85 const TargetX8664Traits::TableIcmp64Type TargetX8664Traits::TableIcmp64[] = {
86 #define X(val, C_32, C1_64, C2_64, C3_64)                                      \
87   {                                                                            \
88     X8664::Traits::Cond::C1_64, X8664::Traits::Cond::C2_64,                    \
89         X8664::Traits::Cond::C3_64                                             \
90   }                                                                            \
91   ,
92     ICMPX8664_TABLE
93 #undef X
94 };
95 
96 const size_t TargetX8664Traits::TableIcmp64Size =
97     llvm::array_lengthof(TableIcmp64);
98 
99 const TargetX8664Traits::TableTypeX8664AttributesType
100     TargetX8664Traits::TableTypeX8664Attributes[] = {
101 #define X(tag, elty, cvt, sdss, pdps, spsd, int_, unpack, pack, width, fld)    \
102   { IceType_##elty }                                                           \
103   ,
104         ICETYPEX8664_TABLE
105 #undef X
106 };
107 
108 const size_t TargetX8664Traits::TableTypeX8664AttributesSize =
109     llvm::array_lengthof(TableTypeX8664Attributes);
110 
111 const uint32_t TargetX8664Traits::X86_STACK_ALIGNMENT_BYTES = 16;
112 const char *TargetX8664Traits::TargetName = "X8664";
113 
114 template <>
115 std::array<SmallBitVector, RCX86_NUM>
116     TargetX86Base<X8664::Traits>::TypeToRegisterSet = {{}};
117 
118 template <>
119 std::array<SmallBitVector, RCX86_NUM>
120     TargetX86Base<X8664::Traits>::TypeToRegisterSetUnfiltered = {{}};
121 
122 template <>
123 std::array<SmallBitVector,
124            TargetX86Base<X8664::Traits>::Traits::RegisterSet::Reg_NUM>
125     TargetX86Base<X8664::Traits>::RegisterAliases = {{}};
126 
127 template <>
128 FixupKind TargetX86Base<X8664::Traits>::PcRelFixup =
129     TargetX86Base<X8664::Traits>::Traits::FK_PcRel;
130 
131 template <>
132 FixupKind TargetX86Base<X8664::Traits>::AbsFixup =
133     TargetX86Base<X8664::Traits>::Traits::FK_Abs;
134 
135 //------------------------------------------------------------------------------
136 //     __      ______  __     __  ______  ______  __  __   __  ______
137 //    /\ \    /\  __ \/\ \  _ \ \/\  ___\/\  == \/\ \/\ "-.\ \/\  ___\
138 //    \ \ \___\ \ \/\ \ \ \/ ".\ \ \  __\\ \  __<\ \ \ \ \-.  \ \ \__ \
139 //     \ \_____\ \_____\ \__/".~\_\ \_____\ \_\ \_\ \_\ \_\\"\_\ \_____\
140 //      \/_____/\/_____/\/_/   \/_/\/_____/\/_/ /_/\/_/\/_/ \/_/\/_____/
141 //
142 //------------------------------------------------------------------------------
_add_sp(Operand * Adjustment)143 void TargetX8664::_add_sp(Operand *Adjustment) {
144   Variable *rsp =
145       getPhysicalRegister(Traits::RegisterSet::Reg_rsp, IceType_i64);
146   if (!NeedSandboxing) {
147     _add(rsp, Adjustment);
148     return;
149   }
150 
151   Variable *esp =
152       getPhysicalRegister(Traits::RegisterSet::Reg_esp, IceType_i32);
153   Variable *r15 =
154       getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
155 
156   // When incrementing rsp, NaCl sandboxing requires the following sequence
157   //
158   // .bundle_start
159   // add Adjustment, %esp
160   // add %r15, %rsp
161   // .bundle_end
162   //
163   // In Subzero, even though rsp and esp alias each other, defining one does not
164   // define the other. Therefore, we must emit
165   //
166   // .bundle_start
167   // %esp = fake-def %rsp
168   // add Adjustment, %esp
169   // %rsp = fake-def %esp
170   // add %r15, %rsp
171   // .bundle_end
172   //
173   // The fake-defs ensure that the
174   //
175   // add Adjustment, %esp
176   //
177   // instruction is not DCE'd.
178   AutoBundle _(this);
179   _redefined(Context.insert<InstFakeDef>(esp, rsp));
180   _add(esp, Adjustment);
181   _redefined(Context.insert<InstFakeDef>(rsp, esp));
182   _add(rsp, r15);
183 }
184 
_mov_sp(Operand * NewValue)185 void TargetX8664::_mov_sp(Operand *NewValue) {
186   assert(NewValue->getType() == IceType_i32);
187 
188   Variable *esp = getPhysicalRegister(Traits::RegisterSet::Reg_esp);
189   Variable *rsp =
190       getPhysicalRegister(Traits::RegisterSet::Reg_rsp, IceType_i64);
191 
192   AutoBundle _(this);
193 
194   _redefined(Context.insert<InstFakeDef>(esp, rsp));
195   _redefined(_mov(esp, NewValue));
196   _redefined(Context.insert<InstFakeDef>(rsp, esp));
197 
198   if (!NeedSandboxing) {
199     return;
200   }
201 
202   Variable *r15 =
203       getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
204   _add(rsp, r15);
205 }
206 
_push_rbp()207 void TargetX8664::_push_rbp() {
208   assert(NeedSandboxing);
209 
210   Constant *_0 = Ctx->getConstantZero(IceType_i32);
211   Variable *ebp =
212       getPhysicalRegister(Traits::RegisterSet::Reg_ebp, IceType_i32);
213   Variable *rsp =
214       getPhysicalRegister(Traits::RegisterSet::Reg_rsp, IceType_i64);
215   auto *TopOfStack = llvm::cast<X86OperandMem>(
216       legalize(X86OperandMem::create(Func, IceType_i32, rsp, _0),
217                Legal_Reg | Legal_Mem));
218 
219   // Emits a sequence:
220   //
221   //   .bundle_start
222   //   push 0
223   //   mov %ebp, %(rsp)
224   //   .bundle_end
225   //
226   // to avoid leaking the upper 32-bits (i.e., the sandbox address.)
227   AutoBundle _(this);
228   _push(_0);
229   Context.insert<typename Traits::Insts::Store>(ebp, TopOfStack);
230 }
231 
_link_bp()232 void TargetX8664::_link_bp() {
233   Variable *esp =
234       getPhysicalRegister(Traits::RegisterSet::Reg_esp, IceType_i32);
235   Variable *rsp =
236       getPhysicalRegister(Traits::RegisterSet::Reg_rsp, Traits::WordType);
237   Variable *ebp =
238       getPhysicalRegister(Traits::RegisterSet::Reg_ebp, IceType_i32);
239   Variable *rbp =
240       getPhysicalRegister(Traits::RegisterSet::Reg_rbp, Traits::WordType);
241   Variable *r15 =
242       getPhysicalRegister(Traits::RegisterSet::Reg_r15, Traits::WordType);
243 
244   if (!NeedSandboxing) {
245     _push(rbp);
246     _mov(rbp, rsp);
247   } else {
248     _push_rbp();
249 
250     AutoBundle _(this);
251     _redefined(Context.insert<InstFakeDef>(ebp, rbp));
252     _redefined(Context.insert<InstFakeDef>(esp, rsp));
253     _mov(ebp, esp);
254     _redefined(Context.insert<InstFakeDef>(rsp, esp));
255     _add(rbp, r15);
256   }
257   // Keep ebp live for late-stage liveness analysis (e.g. asm-verbose mode).
258   Context.insert<InstFakeUse>(rbp);
259 }
260 
_unlink_bp()261 void TargetX8664::_unlink_bp() {
262   Variable *rsp =
263       getPhysicalRegister(Traits::RegisterSet::Reg_rsp, IceType_i64);
264   Variable *rbp =
265       getPhysicalRegister(Traits::RegisterSet::Reg_rbp, IceType_i64);
266   Variable *ebp =
267       getPhysicalRegister(Traits::RegisterSet::Reg_ebp, IceType_i32);
268   // For late-stage liveness analysis (e.g. asm-verbose mode), adding a fake
269   // use of rsp before the assignment of rsp=rbp keeps previous rsp
270   // adjustments from being dead-code eliminated.
271   Context.insert<InstFakeUse>(rsp);
272   if (!NeedSandboxing) {
273     _mov(rsp, rbp);
274     _pop(rbp);
275   } else {
276     _mov_sp(ebp);
277 
278     Variable *r15 =
279         getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
280     Variable *rcx =
281         getPhysicalRegister(Traits::RegisterSet::Reg_rcx, IceType_i64);
282     Variable *ecx =
283         getPhysicalRegister(Traits::RegisterSet::Reg_ecx, IceType_i32);
284 
285     _pop(rcx);
286     Context.insert<InstFakeDef>(ecx, rcx);
287     AutoBundle _(this);
288     _mov(ebp, ecx);
289 
290     _redefined(Context.insert<InstFakeDef>(rbp, ebp));
291     _add(rbp, r15);
292   }
293 }
294 
_push_reg(Variable * Reg)295 void TargetX8664::_push_reg(Variable *Reg) {
296   Variable *rbp =
297       getPhysicalRegister(Traits::RegisterSet::Reg_rbp, Traits::WordType);
298   if (Reg != rbp || !NeedSandboxing) {
299     _push(Reg);
300   } else {
301     _push_rbp();
302   }
303 }
304 
emitGetIP(CfgNode * Node)305 void TargetX8664::emitGetIP(CfgNode *Node) {
306   // No IP base register is needed on X86-64.
307   (void)Node;
308 }
309 
310 namespace {
isAssignedToRspOrRbp(const Variable * Var)311 bool isAssignedToRspOrRbp(const Variable *Var) {
312   if (Var == nullptr) {
313     return false;
314   }
315 
316   if (Var->isRematerializable()) {
317     return true;
318   }
319 
320   if (!Var->hasReg()) {
321     return false;
322   }
323 
324   const auto RegNum = Var->getRegNum();
325   if ((RegNum == Traits::RegisterSet::Reg_rsp) ||
326       (RegNum == Traits::RegisterSet::Reg_rbp)) {
327     return true;
328   }
329 
330   return false;
331 }
332 } // end of anonymous namespace
333 
_sandbox_mem_reference(X86OperandMem * Mem)334 Traits::X86OperandMem *TargetX8664::_sandbox_mem_reference(X86OperandMem *Mem) {
335   if (SandboxingType == ST_None) {
336     return Mem;
337   }
338 
339   if (SandboxingType == ST_Nonsfi) {
340     llvm::report_fatal_error(
341         "_sandbox_mem_reference not implemented for nonsfi");
342   }
343 
344   // In x86_64-nacl, all memory references are relative to a base register
345   // (%r15, %rsp, %rbp, or %rip).
346 
347   Variable *Base = Mem->getBase();
348   Variable *Index = Mem->getIndex();
349   uint16_t Shift = 0;
350   Variable *ZeroReg = RebasePtr;
351   Constant *Offset = Mem->getOffset();
352   Variable *T = nullptr;
353 
354   bool AbsoluteAddress = false;
355   if (Base == nullptr && Index == nullptr) {
356     if (llvm::isa<ConstantRelocatable>(Offset)) {
357       // Mem is RIP-relative. There's no need to rebase it.
358       return Mem;
359     }
360     // Offset is an absolute address, so we need to emit
361     //   Offset(%r15)
362     AbsoluteAddress = true;
363   }
364 
365   if (Mem->getIsRebased()) {
366     // If Mem.IsRebased, then we don't need to update Mem, as it's already been
367     // updated to contain a reference to one of %rsp, %rbp, or %r15.
368     // We don't return early because we still need to zero extend Index.
369     assert(ZeroReg == Base || AbsoluteAddress || isAssignedToRspOrRbp(Base));
370     if (!AbsoluteAddress) {
371       // If Mem is an absolute address, no need to update ZeroReg (which is
372       // already set to %r15.)
373       ZeroReg = Base;
374     }
375     if (Index != nullptr) {
376       T = makeReg(IceType_i32);
377       _mov(T, Index);
378       Shift = Mem->getShift();
379     }
380   } else {
381     if (Base != nullptr) {
382       // If Base is a valid base pointer we don't need to use the RebasePtr. By
383       // doing this we might save us the need to zero extend the memory operand.
384       if (isAssignedToRspOrRbp(Base)) {
385         ZeroReg = Base;
386       } else {
387         T = Base;
388       }
389     }
390 
391     if (Index != nullptr) {
392       assert(!Index->isRematerializable());
393       // If Index is not nullptr, it is mandatory that T is a nullptr.
394       // Otherwise, the lowering generated a memory operand with two registers.
395       // Note that Base might still be non-nullptr, but it must be a valid
396       // base register.
397       if (T != nullptr) {
398         llvm::report_fatal_error("memory reference contains base and index.");
399       }
400       // If the Index is not shifted, and it is a Valid Base, and the ZeroReg is
401       // still RebasePtr, then we do ZeroReg = Index, and hopefully prevent the
402       // need to zero-extend the memory operand (which may still happen -- see
403       // NeedLea below.)
404       if (Shift == 0 && isAssignedToRspOrRbp(Index) && ZeroReg == RebasePtr) {
405         ZeroReg = Index;
406       } else {
407         T = Index;
408         Shift = Mem->getShift();
409       }
410     }
411   }
412 
413   // NeedsLea is a flag indicating whether Mem needs to be materialized to a GPR
414   // prior to being used. A LEA is needed if Mem.Offset is a constant
415   // relocatable with a nonzero offset, or if Mem.Offset is a nonzero immediate;
416   // but only when the address mode contains a "user" register other than the
417   // rsp/rbp/r15 base. In both these cases, the LEA is needed to ensure the
418   // sandboxed memory operand will only use the lower 32-bits of T+Offset.
419   bool NeedsLea = false;
420   if (!Mem->getIsRebased()) {
421     bool IsOffsetZero = false;
422     if (Offset == nullptr) {
423       IsOffsetZero = true;
424     } else if (const auto *CR = llvm::dyn_cast<ConstantRelocatable>(Offset)) {
425       IsOffsetZero = (CR->getOffset() == 0);
426     } else if (const auto *Imm = llvm::dyn_cast<ConstantInteger32>(Offset)) {
427       IsOffsetZero = (Imm->getValue() == 0);
428     } else {
429       llvm::report_fatal_error("Unexpected Offset type.");
430     }
431     if (!IsOffsetZero) {
432       if (Base != nullptr && Base != ZeroReg)
433         NeedsLea = true;
434       if (Index != nullptr && Index != ZeroReg)
435         NeedsLea = true;
436     }
437   }
438 
439   RegNumT RegNum, RegNum32;
440   if (T != nullptr) {
441     if (T->hasReg()) {
442       RegNum = Traits::getGprForType(IceType_i64, T->getRegNum());
443       RegNum32 = Traits::getGprForType(IceType_i32, RegNum);
444       // At this point, if T was assigned to rsp/rbp, then we would have already
445       // made this the ZeroReg.
446       assert(RegNum != Traits::RegisterSet::Reg_rsp);
447       assert(RegNum != Traits::RegisterSet::Reg_rbp);
448     }
449 
450     switch (T->getType()) {
451     default:
452       llvm::report_fatal_error("Mem pointer should be a 32-bit GPR.");
453     case IceType_i64:
454       // Even though "default:" would also catch T.Type == IceType_i64, an
455       // explicit 'case IceType_i64' shows that memory operands are always
456       // supposed to be 32-bits.
457       llvm::report_fatal_error("Mem pointer should not be a 64-bit GPR.");
458     case IceType_i32: {
459       Variable *T64 = makeReg(IceType_i64, RegNum);
460       auto *Movzx = _movzx(T64, T);
461       if (!NeedsLea) {
462         // This movzx is only needed when Mem does not need to be lea'd into a
463         // temporary. If an lea is going to be emitted, then eliding this movzx
464         // is safe because the emitted lea will write a 32-bit result --
465         // implicitly zero-extended to 64-bit.
466         Movzx->setMustKeep();
467       }
468       T = T64;
469     } break;
470     }
471   }
472 
473   if (NeedsLea) {
474     Variable *NewT = makeReg(IceType_i32, RegNum32);
475     Variable *Base = T;
476     Variable *Index = T;
477     static constexpr bool NotRebased = false;
478     if (Shift == 0) {
479       Index = nullptr;
480     } else {
481       Base = nullptr;
482     }
483     _lea(NewT, Traits::X86OperandMem::create(
484                    Func, Mem->getType(), Base, Offset, Index, Shift,
485                    Traits::X86OperandMem::DefaultSegment, NotRebased));
486 
487     T = makeReg(IceType_i64, RegNum);
488     _movzx(T, NewT);
489     Shift = 0;
490     Offset = nullptr;
491   }
492 
493   static constexpr bool IsRebased = true;
494   return Traits::X86OperandMem::create(
495       Func, Mem->getType(), ZeroReg, Offset, T, Shift,
496       Traits::X86OperandMem::DefaultSegment, IsRebased);
497 }
498 
_sub_sp(Operand * Adjustment)499 void TargetX8664::_sub_sp(Operand *Adjustment) {
500   Variable *rsp =
501       getPhysicalRegister(Traits::RegisterSet::Reg_rsp, Traits::WordType);
502 
503   if (NeedSandboxing) {
504     Variable *esp =
505         getPhysicalRegister(Traits::RegisterSet::Reg_esp, IceType_i32);
506     Variable *r15 =
507         getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
508 
509     // .bundle_start
510     // sub Adjustment, %esp
511     // add %r15, %rsp
512     // .bundle_end
513     AutoBundle _(this);
514     _redefined(Context.insert<InstFakeDef>(esp, rsp));
515     _sub(esp, Adjustment);
516     _redefined(Context.insert<InstFakeDef>(rsp, esp));
517     _add(rsp, r15);
518   } else {
519     _sub(rsp, Adjustment);
520   }
521 
522   // Add a fake use of the stack pointer, to prevent the stack pointer adustment
523   // from being dead-code eliminated in a function that doesn't return.
524   Context.insert<InstFakeUse>(rsp);
525 }
526 
initRebasePtr()527 void TargetX8664::initRebasePtr() {
528   switch (SandboxingType) {
529   case ST_Nonsfi:
530     // Probably no implementation is needed, but error to be safe for now.
531     llvm::report_fatal_error(
532         "initRebasePtr() is not yet implemented on x32-nonsfi.");
533   case ST_NaCl:
534     RebasePtr = getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
535     break;
536   case ST_None:
537     // nothing.
538     break;
539   }
540 }
541 
initSandbox()542 void TargetX8664::initSandbox() {
543   assert(SandboxingType == ST_NaCl);
544   Context.init(Func->getEntryNode());
545   Context.setInsertPoint(Context.getCur());
546   Variable *r15 =
547       getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
548   Context.insert<InstFakeDef>(r15);
549   Context.insert<InstFakeUse>(r15);
550 }
551 
552 namespace {
isRematerializable(const Variable * Var)553 bool isRematerializable(const Variable *Var) {
554   return Var != nullptr && Var->isRematerializable();
555 }
556 } // end of anonymous namespace
557 
legalizeOptAddrForSandbox(OptAddr * Addr)558 bool TargetX8664::legalizeOptAddrForSandbox(OptAddr *Addr) {
559   if (SandboxingType == ST_Nonsfi) {
560     llvm::report_fatal_error("Nonsfi not yet implemented for x8664.");
561   }
562 
563   if (isRematerializable(Addr->Base)) {
564     if (Addr->Index == RebasePtr) {
565       Addr->Index = nullptr;
566       Addr->Shift = 0;
567     }
568     return true;
569   }
570 
571   if (isRematerializable(Addr->Index)) {
572     if (Addr->Base == RebasePtr) {
573       Addr->Base = nullptr;
574     }
575     return true;
576   }
577 
578   assert(Addr->Base != RebasePtr && Addr->Index != RebasePtr);
579 
580   if (Addr->Base == nullptr) {
581     return true;
582   }
583 
584   if (Addr->Index == nullptr) {
585     return true;
586   }
587 
588   return false;
589 }
590 
lowerIndirectJump(Variable * JumpTarget)591 void TargetX8664::lowerIndirectJump(Variable *JumpTarget) {
592   std::unique_ptr<AutoBundle> Bundler;
593 
594   if (!NeedSandboxing) {
595     if (JumpTarget->getType() != IceType_i64) {
596       Variable *T = makeReg(IceType_i64);
597       _movzx(T, JumpTarget);
598       JumpTarget = T;
599     }
600   } else {
601     Variable *T = makeReg(IceType_i32);
602     Variable *T64 = makeReg(IceType_i64);
603     Variable *r15 =
604         getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
605 
606     _mov(T, JumpTarget);
607     Bundler = makeUnique<AutoBundle>(this);
608     const SizeT BundleSize =
609         1 << Func->getAssembler<>()->getBundleAlignLog2Bytes();
610     _and(T, Ctx->getConstantInt32(~(BundleSize - 1)));
611     _movzx(T64, T);
612     _add(T64, r15);
613     JumpTarget = T64;
614   }
615 
616   _jmp(JumpTarget);
617 }
618 
emitCallToTarget(Operand * CallTarget,Variable * ReturnReg)619 Inst *TargetX8664::emitCallToTarget(Operand *CallTarget, Variable *ReturnReg) {
620   Inst *NewCall = nullptr;
621   auto *CallTargetR = llvm::dyn_cast<Variable>(CallTarget);
622   if (NeedSandboxing) {
623     // In NaCl sandbox, calls are replaced by a push/jmp pair:
624     //
625     //     push .after_call
626     //     jmp CallTarget
627     //     .align bundle_size
628     // after_call:
629     //
630     // In order to emit this sequence, we need a temporary label ("after_call",
631     // in this example.)
632     //
633     // The operand to push is a ConstantRelocatable. The easy way to implement
634     // this sequence is to create a ConstantRelocatable(0, "after_call"), but
635     // this ends up creating more relocations for the linker to resolve.
636     // Therefore, we create a ConstantRelocatable from the name of the function
637     // being compiled (i.e., ConstantRelocatable(after_call - Func, Func).
638     //
639     // By default, ConstantRelocatables are emitted (in textual output) as
640     //
641     //  ConstantName + Offset
642     //
643     // ReturnReloc has an offset that is only known during binary emission.
644     // Therefore, we set a custom emit string for ReturnReloc that will be
645     // used instead. In this particular case, the code will be emitted as
646     //
647     //  push .after_call
648     InstX86Label *ReturnAddress = InstX86Label::create(Func, this);
649     auto *ReturnRelocOffset = RelocOffset::create(Func->getAssembler());
650     ReturnAddress->setRelocOffset(ReturnRelocOffset);
651     constexpr RelocOffsetT NoFixedOffset = 0;
652     const std::string EmitString =
653         BuildDefs::dump() ? ReturnAddress->getLabelName().toString() : "";
654     auto *ReturnReloc = ConstantRelocatable::create(
655         Func->getAssembler(), IceType_i32,
656         RelocatableTuple(NoFixedOffset, {ReturnRelocOffset},
657                          Func->getFunctionName(), EmitString));
658     /* AutoBundle scoping */ {
659       std::unique_ptr<AutoBundle> Bundler;
660       if (CallTargetR == nullptr) {
661         Bundler = makeUnique<AutoBundle>(this, InstBundleLock::Opt_PadToEnd);
662         _push(ReturnReloc);
663       } else {
664         Variable *T = makeReg(IceType_i32);
665         Variable *T64 = makeReg(IceType_i64);
666         Variable *r15 =
667             getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
668 
669         _mov(T, CallTargetR);
670         Bundler = makeUnique<AutoBundle>(this, InstBundleLock::Opt_PadToEnd);
671         _push(ReturnReloc);
672         const SizeT BundleSize =
673             1 << Func->getAssembler<>()->getBundleAlignLog2Bytes();
674         _and(T, Ctx->getConstantInt32(~(BundleSize - 1)));
675         _movzx(T64, T);
676         _add(T64, r15);
677         CallTarget = T64;
678       }
679 
680       NewCall = Context.insert<Traits::Insts::Jmp>(CallTarget);
681     }
682     if (ReturnReg != nullptr) {
683       Context.insert<InstFakeDef>(ReturnReg);
684     }
685 
686     Context.insert(ReturnAddress);
687   } else {
688     if (CallTargetR != nullptr && CallTarget->getType() == IceType_i32) {
689       // x86-64 in PNaCl is ILP32. Therefore, CallTarget is i32, but the
690       // emitted call needs an i64 register (for textual asm.)
691       Variable *T = makeReg(IceType_i64);
692       _movzx(T, CallTargetR);
693       CallTarget = T;
694     } else if (llvm::isa<Constant>(CallTarget) &&
695                CallTarget->getType() == IceType_i64) {
696       // x86-64 does not support 64-bit direct calls, so write the value
697       // to a register and make an indirect call.
698       Variable *T = makeReg(IceType_i64);
699       _mov(T, CallTarget);
700       CallTarget = T;
701     }
702     NewCall = Context.insert<Traits::Insts::Call>(ReturnReg, CallTarget);
703   }
704   return NewCall;
705 }
706 
moveReturnValueToRegister(Operand * Value,Type ReturnType)707 Variable *TargetX8664::moveReturnValueToRegister(Operand *Value,
708                                                  Type ReturnType) {
709   if (isVectorType(ReturnType) || isScalarFloatingType(ReturnType)) {
710     return legalizeToReg(Value, Traits::RegisterSet::Reg_xmm0);
711   } else {
712     assert(ReturnType == IceType_i32 || ReturnType == IceType_i64);
713     Variable *Reg = nullptr;
714     _mov(Reg, Value,
715          Traits::getGprForType(ReturnType, Traits::RegisterSet::Reg_rax));
716     return Reg;
717   }
718 }
719 
emitSandboxedReturn()720 void TargetX8664::emitSandboxedReturn() {
721   Variable *T_rcx = makeReg(IceType_i64, Traits::RegisterSet::Reg_rcx);
722   Variable *T_ecx = makeReg(IceType_i32, Traits::RegisterSet::Reg_ecx);
723   _pop(T_rcx);
724   _mov(T_ecx, T_rcx);
725   // lowerIndirectJump(T_ecx);
726   Variable *r15 =
727       getPhysicalRegister(Traits::RegisterSet::Reg_r15, IceType_i64);
728 
729   /* AutoBundle scoping */ {
730     AutoBundle _(this);
731     const SizeT BundleSize =
732         1 << Func->getAssembler<>()->getBundleAlignLog2Bytes();
733     _and(T_ecx, Ctx->getConstantInt32(~(BundleSize - 1)));
734     Context.insert<InstFakeDef>(T_rcx, T_ecx);
735     _add(T_rcx, r15);
736 
737     _jmp(T_rcx);
738   }
739 }
740 
741 // In some cases, there are x-macros tables for both high-level and low-level
742 // instructions/operands that use the same enum key value. The tables are kept
743 // separate to maintain a proper separation between abstraction layers. There
744 // is a risk that the tables could get out of sync if enum values are reordered
745 // or if entries are added or deleted. The following dummy namespaces use
746 // static_asserts to ensure everything is kept in sync.
747 
748 namespace {
749 // Validate the enum values in FCMPX8664_TABLE.
750 namespace dummy1 {
751 // Define a temporary set of enum values based on low-level table entries.
752 enum _tmp_enum {
753 #define X(val, dflt, swapS, C1, C2, swapV, pred) _tmp_##val,
754   FCMPX8664_TABLE
755 #undef X
756       _num
757 };
758 // Define a set of constants based on high-level table entries.
759 #define X(tag, str) static const int _table1_##tag = InstFcmp::tag;
760 ICEINSTFCMP_TABLE
761 #undef X
762 // Define a set of constants based on low-level table entries, and ensure the
763 // table entry keys are consistent.
764 #define X(val, dflt, swapS, C1, C2, swapV, pred)                               \
765   static const int _table2_##val = _tmp_##val;                                 \
766   static_assert(                                                               \
767       _table1_##val == _table2_##val,                                          \
768       "Inconsistency between FCMPX8664_TABLE and ICEINSTFCMP_TABLE");
769 FCMPX8664_TABLE
770 #undef X
771 // Repeat the static asserts with respect to the high-level table entries in
772 // case the high-level table has extra entries.
773 #define X(tag, str)                                                            \
774   static_assert(                                                               \
775       _table1_##tag == _table2_##tag,                                          \
776       "Inconsistency between FCMPX8664_TABLE and ICEINSTFCMP_TABLE");
777 ICEINSTFCMP_TABLE
778 #undef X
779 } // end of namespace dummy1
780 
781 // Validate the enum values in ICMPX8664_TABLE.
782 namespace dummy2 {
783 // Define a temporary set of enum values based on low-level table entries.
784 enum _tmp_enum {
785 #define X(val, C_32, C1_64, C2_64, C3_64) _tmp_##val,
786   ICMPX8664_TABLE
787 #undef X
788       _num
789 };
790 // Define a set of constants based on high-level table entries.
791 #define X(tag, reverse, str) static const int _table1_##tag = InstIcmp::tag;
792 ICEINSTICMP_TABLE
793 #undef X
794 // Define a set of constants based on low-level table entries, and ensure the
795 // table entry keys are consistent.
796 #define X(val, C_32, C1_64, C2_64, C3_64)                                      \
797   static const int _table2_##val = _tmp_##val;                                 \
798   static_assert(                                                               \
799       _table1_##val == _table2_##val,                                          \
800       "Inconsistency between ICMPX8664_TABLE and ICEINSTICMP_TABLE");
801 ICMPX8664_TABLE
802 #undef X
803 // Repeat the static asserts with respect to the high-level table entries in
804 // case the high-level table has extra entries.
805 #define X(tag, reverse, str)                                                   \
806   static_assert(                                                               \
807       _table1_##tag == _table2_##tag,                                          \
808       "Inconsistency between ICMPX8664_TABLE and ICEINSTICMP_TABLE");
809 ICEINSTICMP_TABLE
810 #undef X
811 } // end of namespace dummy2
812 
813 // Validate the enum values in ICETYPEX8664_TABLE.
814 namespace dummy3 {
815 // Define a temporary set of enum values based on low-level table entries.
816 enum _tmp_enum {
817 #define X(tag, elty, cvt, sdss, pdps, spsd, int_, unpack, pack, width, fld)    \
818   _tmp_##tag,
819   ICETYPEX8664_TABLE
820 #undef X
821       _num
822 };
823 // Define a set of constants based on high-level table entries.
824 #define X(tag, sizeLog2, align, elts, elty, str, rcstr)                        \
825   static const int _table1_##tag = IceType_##tag;
826 ICETYPE_TABLE
827 #undef X
828 // Define a set of constants based on low-level table entries, and ensure the
829 // table entry keys are consistent.
830 #define X(tag, elty, cvt, sdss, pdps, spsd, int_, unpack, pack, width, fld)    \
831   static const int _table2_##tag = _tmp_##tag;                                 \
832   static_assert(_table1_##tag == _table2_##tag,                                \
833                 "Inconsistency between ICETYPEX8664_TABLE and ICETYPE_TABLE");
834 ICETYPEX8664_TABLE
835 #undef X
836 // Repeat the static asserts with respect to the high-level table entries in
837 // case the high-level table has extra entries.
838 #define X(tag, sizeLog2, align, elts, elty, str, rcstr)                        \
839   static_assert(_table1_##tag == _table2_##tag,                                \
840                 "Inconsistency between ICETYPEX8664_TABLE and ICETYPE_TABLE");
841 ICETYPE_TABLE
842 #undef X
843 } // end of namespace dummy3
844 } // end of anonymous namespace
845 
846 } // end of namespace X8664
847 } // end of namespace Ice
848