Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(522)

Unified Diff: lib/Transforms/NaCl/RewriteAtomics.cpp

Issue 17777004: Concurrency support for PNaCl ABI (Closed) Base URL: http://git.chromium.org/native_client/pnacl-llvm.git@master
Patch Set: Missed one cleanup file. Created 7 years, 6 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: lib/Transforms/NaCl/RewriteAtomics.cpp
diff --git a/lib/Transforms/NaCl/RewriteAtomics.cpp b/lib/Transforms/NaCl/RewriteAtomics.cpp
new file mode 100644
index 0000000000000000000000000000000000000000..2e86dd09f8f7b7cfbca8f31a0f56f9af5a6c7f1d
--- /dev/null
+++ b/lib/Transforms/NaCl/RewriteAtomics.cpp
@@ -0,0 +1,259 @@
+//===- RewriteAtomics.cpp - Stabilize instructions used for concurrency ---===//
+//
+// The LLVM Compiler Infrastructure
+//
+// This file is distributed under the University of Illinois Open Source
+// License. See LICENSE.TXT for details.
+//
+//===----------------------------------------------------------------------===//
+//
+// This pass encodes atomics, volatiles and fences using NaCl intrinsics
+// instead of LLVM's regular IR instructions.
+//
+// All of the above are transformed into one of the
+// @llvm.nacl.atomic.* intrinsics.
+//
+//===----------------------------------------------------------------------===//
+
+#include "llvm/ADT/Twine.h"
+#include "llvm/IR/Function.h"
+#include "llvm/IR/Instructions.h"
+#include "llvm/IR/Intrinsics.h"
+#include "llvm/IR/Module.h"
+#include "llvm/IR/NaClIntrinsics.h"
+#include "llvm/InstVisitor.h"
+#include "llvm/Pass.h"
+#include "llvm/Support/Compiler.h"
+#include "llvm/Support/raw_ostream.h"
+#include "llvm/Support/raw_ostream.h"
+#include "llvm/Transforms/NaCl.h"
+
+using namespace llvm;
+
+namespace {
+class RewriteAtomics : public ModulePass {
+public:
+ static char ID; // Pass identification, replacement for typeid
+ RewriteAtomics() : ModulePass(ID) {
+ // This is a module pass because it may have to introduce
+ // intrinsic declarations into the module and modify a global function.
+ initializeRewriteAtomicsPass(*PassRegistry::getPassRegistry());
+ }
+
+ virtual bool runOnModule(Module &M);
+};
+
+template <class T> Twine ToTwine(const T &V) {
+ std::string S;
+ raw_string_ostream OS(S);
+ OS << const_cast<T &>(V);
+ return OS.str();
+}
+
+class AtomicVisitor : public InstVisitor<AtomicVisitor> {
+ Module &M;
+ LLVMContext &C;
+ NaCl::AtomicIntrinsics AI;
+ bool ModifiedModule;
+
+ AtomicVisitor() LLVM_DELETED_FUNCTION;
+ AtomicVisitor(const AtomicVisitor &) LLVM_DELETED_FUNCTION;
+ AtomicVisitor &operator=(const AtomicVisitor &) LLVM_DELETED_FUNCTION;
+
+ template <class Instruction>
+ ConstantInt *freezeMemoryOrder(const Instruction &I) const;
+ void checkSizeMatchesType(const Instruction &I, unsigned S,
+ const Type *T) const;
+ void checkAlignment(const Instruction &I, unsigned Alignment,
+ unsigned Size) const;
+ void replaceInstructionWithIntrinsicCall(Instruction &I, Intrinsic::ID ID,
+ Type *OverloadedType,
+ ArrayRef<Value *> Args);
+
+ // Most atomics instructions deal with at least one pointer, this
+ // struct automates some of this and has generic sanity checks.
+ template <class Instruction> struct PointerHelper {
+ Value *P;
+ Type *PET;
+ unsigned Size;
+ PointerHelper(const AtomicVisitor &AV, Instruction &I)
+ : P(I.getPointerOperand()) {
+ if (I.getPointerAddressSpace() != 0)
+ report_fatal_error("unhandled pointer address space " +
+ Twine(I.getPointerAddressSpace()) + " for atomic: " +
+ ToTwine(I));
+ assert(P->getType()->isPointerTy() && "expected a pointer");
+ PET = P->getType()->getPointerElementType();
+ Size = PET->getIntegerBitWidth();
+ AV.checkSizeMatchesType(I, Size, PET);
+ }
+ };
+
+public:
+ AtomicVisitor(Module &M)
jvoung (off chromium) 2013/07/03 20:14:20 nit: Might be more useful to have public API be li
JF 2013/07/03 22:28:30 Done.
+ : M(M), C(M.getContext()), AI(C), ModifiedModule(false) {}
+ ~AtomicVisitor() {}
+ bool modifiedModule() const { return ModifiedModule; }
+
+ void visitLoadInst(LoadInst &I);
+ void visitStoreInst(StoreInst &I);
+ void visitAtomicCmpXchgInst(AtomicCmpXchgInst &I);
+ void visitAtomicRMWInst(AtomicRMWInst &I);
+ void visitFenceInst(FenceInst &I);
+};
+}
+
+char RewriteAtomics::ID = 0;
+INITIALIZE_PASS(RewriteAtomics, "nacl-rewrite-atomics",
+ "rewrite atomics, volatiles and fences into stable "
+ "@llvm.nacl.atomics.* intrinsics",
+ false, false)
+
+bool RewriteAtomics::runOnModule(Module &M) {
+ AtomicVisitor AV(M);
+ AV.visit(M);
+ return AV.modifiedModule();
+}
+
+template <class Instruction>
+ConstantInt *AtomicVisitor::freezeMemoryOrder(const Instruction &I) const {
+ NaCl::MemoryOrder AO = NaCl::MemoryOrderInvalid;
+
+ // TODO Volatile load/store are promoted to sequentially consistent
+ // for now. We could do something weaker.
+ if (const LoadInst *L = dyn_cast<LoadInst>(&I)) {
+ if (L->isVolatile())
+ AO = NaCl::MemoryOrderSequentiallyConsistent;
+ } else if (const StoreInst *S = dyn_cast<StoreInst>(&I)) {
+ if (S->isVolatile())
+ AO = NaCl::MemoryOrderSequentiallyConsistent;
+ }
+
+ if (AO == NaCl::MemoryOrderInvalid)
jvoung (off chromium) 2013/07/03 20:14:20 Put curlies around the if's then-statement since i
eliben 2013/07/03 20:24:21 Inspired by Steve's stories from yesterday, JF ;-)
JF 2013/07/03 22:28:30 Done.
+ switch (I.getOrdering()) {
+ default:
+ case NotAtomic: llvm_unreachable("unexpected memory order");
+ // Monotonic is a strict superset of Unordered. Both can therefore
+ // map to Relaxed ordering, which is in the C11/C++11 standard.
+ case Unordered: AO = NaCl::MemoryOrderRelaxed; break;
+ case Monotonic: AO = NaCl::MemoryOrderRelaxed; break;
+ // TODO Consume is currently unspecified by LLVM's internal IR.
+ case Acquire: AO = NaCl::MemoryOrderAcquire; break;
+ case Release: AO = NaCl::MemoryOrderRelease; break;
+ case AcquireRelease: AO = NaCl::MemoryOrderAcquireRelease; break;
+ case SequentiallyConsistent:
+ AO = NaCl::MemoryOrderSequentiallyConsistent; break;
+ }
+
+ // TODO For now only sequential consistency is allowed.
+ AO = NaCl::MemoryOrderSequentiallyConsistent;
+
+ return ConstantInt::get(Type::getInt32Ty(C), AO);
+}
+
+void AtomicVisitor::checkSizeMatchesType(const Instruction &I, unsigned S,
+ const Type *T) const {
+ Type *IntType = Type::getIntNTy(C, S);
+ if (IntType && T == IntType)
+ return;
+ report_fatal_error("unsupported atomic type " + ToTwine(*T) + " of size " +
+ Twine(S) + " in: " + ToTwine(I));
+}
+
+void AtomicVisitor::checkAlignment(const Instruction &I, unsigned Alignment,
jvoung (off chromium) 2013/07/03 17:50:26 nit: Perhaps name variables AlignmentBits, SizeBit
JF 2013/07/03 22:28:30 Done.
+ unsigned Size) const {
+ if (Alignment < Size)
+ report_fatal_error("atomic load/store must be at least naturally aligned, "
+ "got " + Twine(Alignment) + ", expected at least " +
+ Twine(Size) + ", in: " + ToTwine(I));
+}
+
+void AtomicVisitor::replaceInstructionWithIntrinsicCall(
+ Instruction &I, Intrinsic::ID ID, Type *OverloadedType,
+ ArrayRef<Value *> Args) {
+ Function *F = AI.find(ID, OverloadedType)->getDeclaration(&M);
+ CallInst *Call = CallInst::Create(F, Args, "", &I);
+ Call->setDebugLoc(I.getDebugLoc());
+ I.replaceAllUsesWith(Call);
+ I.eraseFromParent();
+ ModifiedModule = true;
+}
+
+// %res = load {atomic|volatile} T* %ptr memory_order, align sizeof(T)
+// becomes:
+// %res = call T @llvm.nacl.atomic.load.i<size>(%ptr, memory_order)
+void AtomicVisitor::visitLoadInst(LoadInst &I) {
+ if (I.isSimple())
+ return;
+ PointerHelper<LoadInst> PH(*this, I);
+ checkAlignment(I, I.getAlignment() * 8, PH.Size);
+ Value *Args[] = { PH.P, freezeMemoryOrder(I) };
+ replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_load, PH.PET,
+ Args);
+}
+
+// store {atomic|volatile} T %val, T* %ptr memory_order, align sizeof(T)
+// becomes:
+// call void @llvm.nacl.atomic.store.i<size>(%val, %ptr, memory_order)
+void AtomicVisitor::visitStoreInst(StoreInst &I) {
+ if (I.isSimple())
+ return;
+ PointerHelper<StoreInst> PH(*this, I);
+ checkAlignment(I, I.getAlignment() * 8, PH.Size);
+ checkSizeMatchesType(I, PH.Size, I.getValueOperand()->getType());
+ Value *Args[] = { I.getValueOperand(), PH.P, freezeMemoryOrder(I) };
+ replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_store, PH.PET,
+ Args);
+}
+
+// %res = atomicrmw OP T* %ptr, T %val memory_order
+// becomes:
+// %res = call T @llvm.nacl.atomic.rmw.i<size>(OP, %ptr, %val, memory_order)
+void AtomicVisitor::visitAtomicRMWInst(AtomicRMWInst &I) {
+ NaCl::AtomicRMWOperation Op;
+ switch (I.getOperation()) {
+ default:
+ report_fatal_error("unsupported atomicrmw operation: " + ToTwine(I));
+ case AtomicRMWInst::Add: Op = NaCl::AtomicAdd; break;
+ case AtomicRMWInst::Sub: Op = NaCl::AtomicSub; break;
+ case AtomicRMWInst::And: Op = NaCl::AtomicAnd; break;
+ case AtomicRMWInst::Or: Op = NaCl::AtomicOr; break;
+ case AtomicRMWInst::Xor: Op = NaCl::AtomicXor; break;
+ case AtomicRMWInst::Xchg: Op = NaCl::AtomicExchange; break;
+ }
+ PointerHelper<AtomicRMWInst> PH(*this, I);
+ checkSizeMatchesType(I, PH.Size, I.getValOperand()->getType());
+ Value *Args[] = { ConstantInt::get(Type::getInt32Ty(C), Op), PH.P,
+ I.getValOperand(), freezeMemoryOrder(I) };
+ replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_rmw, PH.PET,
+ Args);
+}
+
+// %res = cmpxchg T* %ptr, T %old, T %new memory_order
+// becomes:
+// %res = call T @llvm.nacl.atomic.cmpxchg.i<size>(
+// %object, %expected, %desired, memory_order_success, memory_order_failure)
+void AtomicVisitor::visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) {
+ PointerHelper<AtomicCmpXchgInst> PH(*this, I);
+ checkSizeMatchesType(I, PH.Size, I.getCompareOperand()->getType());
+ checkSizeMatchesType(I, PH.Size, I.getNewValOperand()->getType());
+ // TODO LLVM currently doesn't support specifying separate memory
+ // orders for compare exchange's success and failure cases: LLVM
+ // IR implicitly drops the Release part of the specified memory
+ // order on failure.
+ Value *Args[] = { PH.P, I.getCompareOperand(), I.getNewValOperand(),
+ freezeMemoryOrder(I), freezeMemoryOrder(I) };
+ replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_cmpxchg, PH.PET,
+ Args);
+}
+
+// fence memory_order
+// becomes:
+// call void @llvm.nacl.atomic.fence(memory_order)
+void AtomicVisitor::visitFenceInst(FenceInst &I) {
+ Type *T = Type::getInt32Ty(C); // Fences aren't overloaded on type.
+ Value *Args[] = { freezeMemoryOrder(I) };
+ replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_fence, T, Args);
+}
+
+ModulePass *llvm::createRewriteAtomicsPass() { return new RewriteAtomics(); }

Powered by Google App Engine
This is Rietveld 408576698