| Index: lib/Transforms/NaCl/RewriteAtomics.cpp
|
| diff --git a/lib/Transforms/NaCl/RewriteAtomics.cpp b/lib/Transforms/NaCl/RewriteAtomics.cpp
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..8aafe3d13d3f1615da60f2e26e03e3b8e7bab6cf
|
| --- /dev/null
|
| +++ b/lib/Transforms/NaCl/RewriteAtomics.cpp
|
| @@ -0,0 +1,436 @@
|
| +//===- RewriteAtomics.cpp - Stabilize instructions used for concurrency ---===//
|
| +//
|
| +// The LLVM Compiler Infrastructure
|
| +//
|
| +// This file is distributed under the University of Illinois Open Source
|
| +// License. See LICENSE.TXT for details.
|
| +//
|
| +//===----------------------------------------------------------------------===//
|
| +//
|
| +// This pass encodes atomics, volatiles and fences using NaCl intrinsics
|
| +// instead of LLVM's regular IR instructions.
|
| +//
|
| +// All of the above are transformed into one of the
|
| +// @llvm.nacl.atomic.* intrinsics.
|
| +//
|
| +//===----------------------------------------------------------------------===//
|
| +
|
| +#include "llvm/ADT/Twine.h"
|
| +#include "llvm/IR/DataLayout.h"
|
| +#include "llvm/IR/Function.h"
|
| +#include "llvm/IR/InlineAsm.h"
|
| +#include "llvm/IR/InstVisitor.h"
|
| +#include "llvm/IR/Instructions.h"
|
| +#include "llvm/IR/Intrinsics.h"
|
| +#include "llvm/IR/Module.h"
|
| +#include "llvm/IR/NaClAtomicIntrinsics.h"
|
| +#include "llvm/Pass.h"
|
| +#include "llvm/Support/CommandLine.h"
|
| +#include "llvm/Support/Compiler.h"
|
| +#include "llvm/Support/raw_ostream.h"
|
| +#include "llvm/Transforms/NaCl.h"
|
| +#include <climits>
|
| +#include <string>
|
| +
|
| +using namespace llvm;
|
| +
|
| +// TODO(jfb) Keep the default of this option to true for Chrome 42, and change
|
| +// it to false for Chrome 43. This allows the PNaCl translator to be
|
| +// updated before the SDK starts emitting atomic memory orders that
|
| +// the old translator rejected.
|
| +static cl::opt<bool> PNaClMemoryOrderSeqCstOnly(
|
| + "pnacl-memory-order-seq-cst-only",
|
| + cl::desc("PNaCl should upgrade all atomic memory orders to seq_cst"),
|
| + cl::init(true));
|
| +
|
| +namespace {
|
| +
|
| +class RewriteAtomics : public ModulePass {
|
| +public:
|
| + static char ID; // Pass identification, replacement for typeid
|
| + RewriteAtomics() : ModulePass(ID) {
|
| + // This is a module pass because it may have to introduce
|
| + // intrinsic declarations into the module and modify a global function.
|
| + initializeRewriteAtomicsPass(*PassRegistry::getPassRegistry());
|
| + }
|
| +
|
| + virtual bool runOnModule(Module &M);
|
| + virtual void getAnalysisUsage(AnalysisUsage &Info) const {
|
| + Info.addRequired<DataLayoutPass>();
|
| + }
|
| +};
|
| +
|
| +template <class T> std::string ToStr(const T &V) {
|
| + std::string S;
|
| + raw_string_ostream OS(S);
|
| + OS << const_cast<T &>(V);
|
| + return OS.str();
|
| +}
|
| +
|
| +class AtomicVisitor : public InstVisitor<AtomicVisitor> {
|
| +public:
|
| + AtomicVisitor(Module &M, Pass &P)
|
| + : M(M), C(M.getContext()),
|
| + TD(P.getAnalysis<DataLayoutPass>().getDataLayout()), AI(C),
|
| + ModifiedModule(false) {}
|
| + ~AtomicVisitor() {}
|
| + bool modifiedModule() const { return ModifiedModule; }
|
| +
|
| + void visitLoadInst(LoadInst &I);
|
| + void visitStoreInst(StoreInst &I);
|
| + void visitAtomicCmpXchgInst(AtomicCmpXchgInst &I);
|
| + void visitAtomicRMWInst(AtomicRMWInst &I);
|
| + void visitFenceInst(FenceInst &I);
|
| +
|
| +private:
|
| + Module &M;
|
| + LLVMContext &C;
|
| + const DataLayout TD;
|
| + NaCl::AtomicIntrinsics AI;
|
| + bool ModifiedModule;
|
| +
|
| + AtomicVisitor() LLVM_DELETED_FUNCTION;
|
| + AtomicVisitor(const AtomicVisitor &) LLVM_DELETED_FUNCTION;
|
| + AtomicVisitor &operator=(const AtomicVisitor &) LLVM_DELETED_FUNCTION;
|
| +
|
| + /// Create an integer constant holding a NaCl::MemoryOrder that can be
|
| + /// passed as an argument to one of the @llvm.nacl.atomic.*
|
| + /// intrinsics. This function may strengthen the ordering initially
|
| + /// specified by the instruction \p I for stability purpose.
|
| + template <class Instruction>
|
| + ConstantInt *freezeMemoryOrder(const Instruction &I, AtomicOrdering O) const;
|
| + std::pair<ConstantInt *, ConstantInt *>
|
| + freezeMemoryOrder(const AtomicCmpXchgInst &I, AtomicOrdering S,
|
| + AtomicOrdering F) const;
|
| +
|
| + /// Sanity-check that instruction \p I which has pointer and value
|
| + /// parameters have matching sizes \p BitSize for the type-pointed-to
|
| + /// and the value's type \p T.
|
| + void checkSizeMatchesType(const Instruction &I, unsigned BitSize,
|
| + const Type *T) const;
|
| +
|
| + /// Verify that loads and stores are at least naturally aligned. Use
|
| + /// byte alignment because converting to bits could truncate the
|
| + /// value.
|
| + void checkAlignment(const Instruction &I, unsigned ByteAlignment,
|
| + unsigned ByteSize) const;
|
| +
|
| + /// Create a cast before Instruction \p I from \p Src to \p Dst with \p Name.
|
| + CastInst *createCast(Instruction &I, Value *Src, Type *Dst, Twine Name) const;
|
| +
|
| + /// Try to find the atomic intrinsic of with its \p ID and \OverloadedType.
|
| + /// Report fatal error on failure.
|
| + const NaCl::AtomicIntrinsics::AtomicIntrinsic *
|
| + findAtomicIntrinsic(const Instruction &I, Intrinsic::ID ID,
|
| + Type *OverloadedType) const;
|
| +
|
| + /// Helper function which rewrites a single instruction \p I to a
|
| + /// particular \p intrinsic with overloaded type \p OverloadedType,
|
| + /// and argument list \p Args. Will perform a bitcast to the proper \p
|
| + /// DstType, if different from \p OverloadedType.
|
| + void replaceInstructionWithIntrinsicCall(
|
| + Instruction &I, const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic,
|
| + Type *DstType, Type *OverloadedType, ArrayRef<Value *> Args);
|
| +
|
| + /// Most atomics instructions deal with at least one pointer, this
|
| + /// struct automates some of this and has generic sanity checks.
|
| + template <class Instruction> struct PointerHelper {
|
| + Value *P;
|
| + Type *OriginalPET;
|
| + Type *PET;
|
| + unsigned BitSize;
|
| + PointerHelper(const AtomicVisitor &AV, Instruction &I)
|
| + : P(I.getPointerOperand()) {
|
| + if (I.getPointerAddressSpace() != 0)
|
| + report_fatal_error("unhandled pointer address space " +
|
| + Twine(I.getPointerAddressSpace()) + " for atomic: " +
|
| + ToStr(I));
|
| + assert(P->getType()->isPointerTy() && "expected a pointer");
|
| + PET = OriginalPET = P->getType()->getPointerElementType();
|
| + BitSize = AV.TD.getTypeSizeInBits(OriginalPET);
|
| + if (!OriginalPET->isIntegerTy()) {
|
| + // The pointer wasn't to an integer type. We define atomics in
|
| + // terms of integers, so bitcast the pointer to an integer of
|
| + // the proper width.
|
| + Type *IntNPtr = Type::getIntNPtrTy(AV.C, BitSize);
|
| + P = AV.createCast(I, P, IntNPtr, P->getName() + ".cast");
|
| + PET = P->getType()->getPointerElementType();
|
| + }
|
| + AV.checkSizeMatchesType(I, BitSize, PET);
|
| + }
|
| + };
|
| +};
|
| +}
|
| +
|
| +char RewriteAtomics::ID = 0;
|
| +INITIALIZE_PASS(RewriteAtomics, "nacl-rewrite-atomics",
|
| + "rewrite atomics, volatiles and fences into stable "
|
| + "@llvm.nacl.atomics.* intrinsics",
|
| + false, false)
|
| +
|
| +bool RewriteAtomics::runOnModule(Module &M) {
|
| + AtomicVisitor AV(M, *this);
|
| + AV.visit(M);
|
| + return AV.modifiedModule();
|
| +}
|
| +
|
| +template <class Instruction>
|
| +ConstantInt *AtomicVisitor::freezeMemoryOrder(const Instruction &I,
|
| + AtomicOrdering O) const {
|
| + NaCl::MemoryOrder AO = NaCl::MemoryOrderInvalid;
|
| +
|
| + // TODO Volatile load/store are promoted to sequentially consistent
|
| + // for now. We could do something weaker.
|
| + if (const LoadInst *L = dyn_cast<LoadInst>(&I)) {
|
| + if (L->isVolatile())
|
| + AO = NaCl::MemoryOrderSequentiallyConsistent;
|
| + } else if (const StoreInst *S = dyn_cast<StoreInst>(&I)) {
|
| + if (S->isVolatile())
|
| + AO = NaCl::MemoryOrderSequentiallyConsistent;
|
| + }
|
| +
|
| + if (AO == NaCl::MemoryOrderInvalid) {
|
| + switch (O) {
|
| + case NotAtomic: llvm_unreachable("unexpected memory order");
|
| + // Monotonic is a strict superset of Unordered. Both can therefore
|
| + // map to Relaxed ordering, which is in the C11/C++11 standard.
|
| + case Unordered: AO = NaCl::MemoryOrderRelaxed; break;
|
| + case Monotonic: AO = NaCl::MemoryOrderRelaxed; break;
|
| + // TODO Consume is currently unspecified by LLVM's internal IR.
|
| + case Acquire: AO = NaCl::MemoryOrderAcquire; break;
|
| + case Release: AO = NaCl::MemoryOrderRelease; break;
|
| + case AcquireRelease: AO = NaCl::MemoryOrderAcquireRelease; break;
|
| + case SequentiallyConsistent:
|
| + AO = NaCl::MemoryOrderSequentiallyConsistent; break;
|
| + }
|
| + }
|
| +
|
| + // TODO For now only acquire/release/acq_rel/seq_cst are allowed.
|
| + if (PNaClMemoryOrderSeqCstOnly || AO == NaCl::MemoryOrderRelaxed)
|
| + AO = NaCl::MemoryOrderSequentiallyConsistent;
|
| +
|
| + return ConstantInt::get(Type::getInt32Ty(C), AO);
|
| +}
|
| +
|
| +std::pair<ConstantInt *, ConstantInt *>
|
| +AtomicVisitor::freezeMemoryOrder(const AtomicCmpXchgInst &I, AtomicOrdering S,
|
| + AtomicOrdering F) const {
|
| + if (S == Release || (S == AcquireRelease && F != Acquire))
|
| + // According to C++11's [atomics.types.operations.req], cmpxchg with release
|
| + // success memory ordering must have relaxed failure memory ordering, which
|
| + // PNaCl currently disallows. The next-strongest ordering is acq_rel which
|
| + // is also an invalid failure ordering, we therefore have to change the
|
| + // success ordering to seq_cst, which can then fail as seq_cst.
|
| + S = F = SequentiallyConsistent;
|
| + if (F == Unordered || F == Monotonic) // Both are treated as relaxed.
|
| + F = AtomicCmpXchgInst::getStrongestFailureOrdering(S);
|
| + return std::make_pair(freezeMemoryOrder(I, S), freezeMemoryOrder(I, F));
|
| +}
|
| +
|
| +void AtomicVisitor::checkSizeMatchesType(const Instruction &I, unsigned BitSize,
|
| + const Type *T) const {
|
| + Type *IntType = Type::getIntNTy(C, BitSize);
|
| + if (IntType && T == IntType)
|
| + return;
|
| + report_fatal_error("unsupported atomic type " + ToStr(*T) + " of size " +
|
| + Twine(BitSize) + " bits in: " + ToStr(I));
|
| +}
|
| +
|
| +void AtomicVisitor::checkAlignment(const Instruction &I, unsigned ByteAlignment,
|
| + unsigned ByteSize) const {
|
| + if (ByteAlignment < ByteSize)
|
| + report_fatal_error("atomic load/store must be at least naturally aligned, "
|
| + "got " +
|
| + Twine(ByteAlignment) + ", bytes expected at least " +
|
| + Twine(ByteSize) + " bytes, in: " + ToStr(I));
|
| +}
|
| +
|
| +CastInst *AtomicVisitor::createCast(Instruction &I, Value *Src, Type *Dst,
|
| + Twine Name) const {
|
| + Type *SrcT = Src->getType();
|
| + Instruction::CastOps Op = SrcT->isIntegerTy() && Dst->isPointerTy()
|
| + ? Instruction::IntToPtr
|
| + : SrcT->isPointerTy() && Dst->isIntegerTy()
|
| + ? Instruction::PtrToInt
|
| + : Instruction::BitCast;
|
| + if (!CastInst::castIsValid(Op, Src, Dst))
|
| + report_fatal_error("cannot emit atomic instruction while converting type " +
|
| + ToStr(*SrcT) + " to " + ToStr(*Dst) + " for " + Name +
|
| + " in " + ToStr(I));
|
| + return CastInst::Create(Op, Src, Dst, Name, &I);
|
| +}
|
| +
|
| +const NaCl::AtomicIntrinsics::AtomicIntrinsic *
|
| +AtomicVisitor::findAtomicIntrinsic(const Instruction &I, Intrinsic::ID ID,
|
| + Type *OverloadedType) const {
|
| + if (const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic =
|
| + AI.find(ID, OverloadedType))
|
| + return Intrinsic;
|
| + report_fatal_error("unsupported atomic instruction: " + ToStr(I));
|
| +}
|
| +
|
| +void AtomicVisitor::replaceInstructionWithIntrinsicCall(
|
| + Instruction &I, const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic,
|
| + Type *DstType, Type *OverloadedType, ArrayRef<Value *> Args) {
|
| + std::string Name(I.getName());
|
| + Function *F = Intrinsic->getDeclaration(&M);
|
| + CallInst *Call = CallInst::Create(F, Args, "", &I);
|
| + Call->setDebugLoc(I.getDebugLoc());
|
| + Instruction *Res = Call;
|
| +
|
| + assert((I.getType()->isStructTy() == isa<AtomicCmpXchgInst>(&I)) &&
|
| + "cmpxchg returns a struct, and other instructions don't");
|
| + if (auto S = dyn_cast<StructType>(I.getType())) {
|
| + assert(S->getNumElements() == 2 &&
|
| + "cmpxchg returns a struct with two elements");
|
| + assert(S->getElementType(0) == DstType &&
|
| + "cmpxchg struct's first member should be the value type");
|
| + assert(S->getElementType(1) == Type::getInt1Ty(C) &&
|
| + "cmpxchg struct's second member should be the success flag");
|
| + // Recreate struct { T value, i1 success } after the call.
|
| + auto Success = CmpInst::Create(
|
| + Instruction::ICmp, CmpInst::ICMP_EQ, Res,
|
| + cast<AtomicCmpXchgInst>(&I)->getCompareOperand(), "success", &I);
|
| + Res = InsertValueInst::Create(
|
| + InsertValueInst::Create(UndefValue::get(S), Res, 0,
|
| + Name + ".insert.value", &I),
|
| + Success, 1, Name + ".insert.success", &I);
|
| + } else if (!Call->getType()->isVoidTy() && DstType != OverloadedType) {
|
| + // The call returns a value which needs to be cast to a non-integer.
|
| + Res = createCast(I, Call, DstType, Name + ".cast");
|
| + Res->setDebugLoc(I.getDebugLoc());
|
| + }
|
| +
|
| + I.replaceAllUsesWith(Res);
|
| + I.eraseFromParent();
|
| + Call->setName(Name);
|
| + ModifiedModule = true;
|
| +}
|
| +
|
| +/// %res = load {atomic|volatile} T* %ptr memory_order, align sizeof(T)
|
| +/// becomes:
|
| +/// %res = call T @llvm.nacl.atomic.load.i<size>(%ptr, memory_order)
|
| +void AtomicVisitor::visitLoadInst(LoadInst &I) {
|
| + if (I.isSimple())
|
| + return;
|
| + PointerHelper<LoadInst> PH(*this, I);
|
| + const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic =
|
| + findAtomicIntrinsic(I, Intrinsic::nacl_atomic_load, PH.PET);
|
| + checkAlignment(I, I.getAlignment(), PH.BitSize / CHAR_BIT);
|
| + Value *Args[] = {PH.P, freezeMemoryOrder(I, I.getOrdering())};
|
| + replaceInstructionWithIntrinsicCall(I, Intrinsic, PH.OriginalPET, PH.PET,
|
| + Args);
|
| +}
|
| +
|
| +/// store {atomic|volatile} T %val, T* %ptr memory_order, align sizeof(T)
|
| +/// becomes:
|
| +/// call void @llvm.nacl.atomic.store.i<size>(%val, %ptr, memory_order)
|
| +void AtomicVisitor::visitStoreInst(StoreInst &I) {
|
| + if (I.isSimple())
|
| + return;
|
| + PointerHelper<StoreInst> PH(*this, I);
|
| + const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic =
|
| + findAtomicIntrinsic(I, Intrinsic::nacl_atomic_store, PH.PET);
|
| + checkAlignment(I, I.getAlignment(), PH.BitSize / CHAR_BIT);
|
| + Value *V = I.getValueOperand();
|
| + if (!V->getType()->isIntegerTy()) {
|
| + // The store isn't of an integer type. We define atomics in terms of
|
| + // integers, so bitcast the value to store to an integer of the
|
| + // proper width.
|
| + CastInst *Cast = createCast(I, V, Type::getIntNTy(C, PH.BitSize),
|
| + V->getName() + ".cast");
|
| + Cast->setDebugLoc(I.getDebugLoc());
|
| + V = Cast;
|
| + }
|
| + checkSizeMatchesType(I, PH.BitSize, V->getType());
|
| + Value *Args[] = {V, PH.P, freezeMemoryOrder(I, I.getOrdering())};
|
| + replaceInstructionWithIntrinsicCall(I, Intrinsic, PH.OriginalPET, PH.PET,
|
| + Args);
|
| +}
|
| +
|
| +/// %res = atomicrmw OP T* %ptr, T %val memory_order
|
| +/// becomes:
|
| +/// %res = call T @llvm.nacl.atomic.rmw.i<size>(OP, %ptr, %val, memory_order)
|
| +void AtomicVisitor::visitAtomicRMWInst(AtomicRMWInst &I) {
|
| + NaCl::AtomicRMWOperation Op;
|
| + switch (I.getOperation()) {
|
| + default: report_fatal_error("unsupported atomicrmw operation: " + ToStr(I));
|
| + case AtomicRMWInst::Add: Op = NaCl::AtomicAdd; break;
|
| + case AtomicRMWInst::Sub: Op = NaCl::AtomicSub; break;
|
| + case AtomicRMWInst::And: Op = NaCl::AtomicAnd; break;
|
| + case AtomicRMWInst::Or: Op = NaCl::AtomicOr; break;
|
| + case AtomicRMWInst::Xor: Op = NaCl::AtomicXor; break;
|
| + case AtomicRMWInst::Xchg: Op = NaCl::AtomicExchange; break;
|
| + }
|
| + PointerHelper<AtomicRMWInst> PH(*this, I);
|
| + const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic =
|
| + findAtomicIntrinsic(I, Intrinsic::nacl_atomic_rmw, PH.PET);
|
| + checkSizeMatchesType(I, PH.BitSize, I.getValOperand()->getType());
|
| + Value *Args[] = {ConstantInt::get(Type::getInt32Ty(C), Op), PH.P,
|
| + I.getValOperand(), freezeMemoryOrder(I, I.getOrdering())};
|
| + replaceInstructionWithIntrinsicCall(I, Intrinsic, PH.OriginalPET, PH.PET,
|
| + Args);
|
| +}
|
| +
|
| +/// %res = cmpxchg [weak] T* %ptr, T %old, T %new, memory_order_success
|
| +/// memory_order_failure
|
| +/// %val = extractvalue { T, i1 } %res, 0
|
| +/// %success = extractvalue { T, i1 } %res, 1
|
| +/// becomes:
|
| +/// %val = call T @llvm.nacl.atomic.cmpxchg.i<size>(
|
| +/// %object, %expected, %desired, memory_order_success,
|
| +/// memory_order_failure)
|
| +/// %success = icmp eq %old, %val
|
| +/// Note: weak is currently dropped if present, the cmpxchg is always strong.
|
| +void AtomicVisitor::visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) {
|
| + PointerHelper<AtomicCmpXchgInst> PH(*this, I);
|
| + const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic =
|
| + findAtomicIntrinsic(I, Intrinsic::nacl_atomic_cmpxchg, PH.PET);
|
| + checkSizeMatchesType(I, PH.BitSize, I.getCompareOperand()->getType());
|
| + checkSizeMatchesType(I, PH.BitSize, I.getNewValOperand()->getType());
|
| + auto Order =
|
| + freezeMemoryOrder(I, I.getSuccessOrdering(), I.getFailureOrdering());
|
| + Value *Args[] = {PH.P, I.getCompareOperand(), I.getNewValOperand(),
|
| + Order.first, Order.second};
|
| + replaceInstructionWithIntrinsicCall(I, Intrinsic, PH.OriginalPET, PH.PET,
|
| + Args);
|
| +}
|
| +
|
| +/// fence memory_order
|
| +/// becomes:
|
| +/// call void @llvm.nacl.atomic.fence(memory_order)
|
| +/// and
|
| +/// call void asm sideeffect "", "~{memory}"()
|
| +/// fence seq_cst
|
| +/// call void asm sideeffect "", "~{memory}"()
|
| +/// becomes:
|
| +/// call void asm sideeffect "", "~{memory}"()
|
| +/// call void @llvm.nacl.atomic.fence.all()
|
| +/// call void asm sideeffect "", "~{memory}"()
|
| +/// Note that the assembly gets eliminated by the -remove-asm-memory pass.
|
| +void AtomicVisitor::visitFenceInst(FenceInst &I) {
|
| + Type *T = Type::getInt32Ty(C); // Fences aren't overloaded on type.
|
| + BasicBlock::InstListType &IL(I.getParent()->getInstList());
|
| + bool isFirst = IL.empty() || &*I.getParent()->getInstList().begin() == &I;
|
| + bool isLast = IL.empty() || &*I.getParent()->getInstList().rbegin() == &I;
|
| + CallInst *PrevC = isFirst ? 0 : dyn_cast<CallInst>(I.getPrevNode());
|
| + CallInst *NextC = isLast ? 0 : dyn_cast<CallInst>(I.getNextNode());
|
| +
|
| + if ((PrevC && PrevC->isInlineAsm() &&
|
| + cast<InlineAsm>(PrevC->getCalledValue())->isAsmMemory()) &&
|
| + (NextC && NextC->isInlineAsm() &&
|
| + cast<InlineAsm>(NextC->getCalledValue())->isAsmMemory()) &&
|
| + I.getOrdering() == SequentiallyConsistent) {
|
| + const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic =
|
| + findAtomicIntrinsic(I, Intrinsic::nacl_atomic_fence_all, T);
|
| + replaceInstructionWithIntrinsicCall(I, Intrinsic, T, T,
|
| + ArrayRef<Value *>());
|
| + } else {
|
| + const NaCl::AtomicIntrinsics::AtomicIntrinsic *Intrinsic =
|
| + findAtomicIntrinsic(I, Intrinsic::nacl_atomic_fence, T);
|
| + Value *Args[] = {freezeMemoryOrder(I, I.getOrdering())};
|
| + replaceInstructionWithIntrinsicCall(I, Intrinsic, T, T, Args);
|
| + }
|
| +}
|
| +
|
| +ModulePass *llvm::createRewriteAtomicsPass() { return new RewriteAtomics(); }
|
|
|