Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(43)

Unified Diff: src/arm64/simulator-arm64.cc

Issue 2720133004: Revert "Revert of [Atomics] Implement ldaxr/stlxr instructions in ARM64 simulator (patchset #8 id:1… (Closed)
Patch Set: fix compile issue Created 3 years, 10 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/arm64/simulator-arm64.h ('k') | test/cctest/BUILD.gn » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/arm64/simulator-arm64.cc
diff --git a/src/arm64/simulator-arm64.cc b/src/arm64/simulator-arm64.cc
index 83b4cf7ee8954f0838a8389eb3a5c971558eb5a2..3c49b21e52d274373ce07b2c548162aa4d7f5981 100644
--- a/src/arm64/simulator-arm64.cc
+++ b/src/arm64/simulator-arm64.cc
@@ -55,6 +55,9 @@ TEXT_COLOUR clr_debug_number = FLAG_log_colour ? COLOUR_BOLD(YELLOW) : "";
TEXT_COLOUR clr_debug_message = FLAG_log_colour ? COLOUR(YELLOW) : "";
TEXT_COLOUR clr_printf = FLAG_log_colour ? COLOUR(GREEN) : "";
+// static
+base::LazyInstance<Simulator::GlobalMonitor>::type Simulator::global_monitor_ =
+ LAZY_INSTANCE_INITIALIZER;
// This is basically the same as PrintF, with a guard for FLAG_trace_sim.
void Simulator::TraceSim(const char* format, ...) {
@@ -429,6 +432,7 @@ void Simulator::ResetState() {
Simulator::~Simulator() {
+ global_monitor_.Pointer()->RemoveProcessor(&global_monitor_processor_);
delete[] reinterpret_cast<byte*>(stack_);
if (FLAG_log_instruction_stats) {
delete instrument_;
@@ -1628,6 +1632,15 @@ void Simulator::LoadStoreHelper(Instruction* instr,
uintptr_t address = LoadStoreAddress(addr_reg, offset, addrmode);
uintptr_t stack = 0;
+ base::LockGuard<base::Mutex> lock_guard(&global_monitor_.Pointer()->mutex);
+ if (instr->IsLoad()) {
+ local_monitor_.NotifyLoad(address);
+ } else {
+ local_monitor_.NotifyStore(address);
+ global_monitor_.Pointer()->NotifyStore_Locked(address,
+ &global_monitor_processor_);
+ }
+
// Handle the writeback for stores before the store. On a CPU the writeback
// and the store are atomic, but when running on the simulator it is possible
// to be interrupted in between. The simulator is not thread safe and V8 does
@@ -1730,6 +1743,19 @@ void Simulator::LoadStorePairHelper(Instruction* instr,
uintptr_t address2 = address + access_size;
uintptr_t stack = 0;
+ base::LockGuard<base::Mutex> lock_guard(&global_monitor_.Pointer()->mutex);
+ if (instr->IsLoad()) {
+ local_monitor_.NotifyLoad(address);
+ local_monitor_.NotifyLoad(address2);
+ } else {
+ local_monitor_.NotifyStore(address);
+ local_monitor_.NotifyStore(address2);
+ global_monitor_.Pointer()->NotifyStore_Locked(address,
+ &global_monitor_processor_);
+ global_monitor_.Pointer()->NotifyStore_Locked(address2,
+ &global_monitor_processor_);
+ }
+
// Handle the writeback for stores before the store. On a CPU the writeback
// and the store are atomic, but when running on the simulator it is possible
// to be interrupted in between. The simulator is not thread safe and V8 does
@@ -1853,6 +1879,9 @@ void Simulator::VisitLoadLiteral(Instruction* instr) {
uintptr_t address = instr->LiteralAddress();
unsigned rt = instr->Rt();
+ base::LockGuard<base::Mutex> lock_guard(&global_monitor_.Pointer()->mutex);
+ local_monitor_.NotifyLoad(address);
+
switch (instr->Mask(LoadLiteralMask)) {
// Use _no_log variants to suppress the register trace (LOG_REGS,
// LOG_FP_REGS), then print a more detailed log.
@@ -1906,8 +1935,81 @@ void Simulator::LoadStoreWriteBack(unsigned addr_reg,
}
}
+Simulator::TransactionSize Simulator::get_transaction_size(unsigned size) {
+ switch (size) {
+ case 0:
+ return TransactionSize::None;
+ case 1:
+ return TransactionSize::Byte;
+ case 2:
+ return TransactionSize::HalfWord;
+ case 4:
+ return TransactionSize::Word;
+ default:
+ UNREACHABLE();
+ }
+ return TransactionSize::None;
+}
+
void Simulator::VisitLoadStoreAcquireRelease(Instruction* instr) {
- // TODO(binji)
+ unsigned rs = instr->Rs();
+ unsigned rt = instr->Rt();
+ unsigned rn = instr->Rn();
+ LoadStoreAcquireReleaseOp op = static_cast<LoadStoreAcquireReleaseOp>(
+ instr->Mask(LoadStoreAcquireReleaseMask));
+ int32_t is_acquire_release = instr->LoadStoreXAcquireRelease();
+ int32_t is_not_exclusive = instr->LoadStoreXNotExclusive();
+ int32_t is_load = instr->LoadStoreXLoad();
+ int32_t is_pair = instr->LoadStoreXPair();
+ DCHECK_NE(is_acquire_release, 0);
+ DCHECK_EQ(is_not_exclusive, 0); // Non exclusive unimplemented.
+ DCHECK_EQ(is_pair, 0); // Pair unimplemented.
+ unsigned access_size = 1 << instr->LoadStoreXSizeLog2();
+ uintptr_t address = LoadStoreAddress(rn, 0, AddrMode::Offset);
+ DCHECK(address % access_size == 0);
+ base::LockGuard<base::Mutex> lock_guard(&global_monitor_.Pointer()->mutex);
+ if (is_load != 0) {
+ local_monitor_.NotifyLoadExcl(address, get_transaction_size(access_size));
+ global_monitor_.Pointer()->NotifyLoadExcl_Locked(
+ address, &global_monitor_processor_);
+ switch (op) {
+ case LDAXR_b:
+ set_wreg_no_log(rt, MemoryRead<uint8_t>(address));
+ break;
+ case LDAXR_h:
+ set_wreg_no_log(rt, MemoryRead<uint16_t>(address));
+ break;
+ case LDAXR_w:
+ set_wreg_no_log(rt, MemoryRead<uint32_t>(address));
+ break;
+ default:
+ UNIMPLEMENTED();
+ }
+ LogRead(address, access_size, rt);
+ } else {
+ if (local_monitor_.NotifyStoreExcl(address,
+ get_transaction_size(access_size)) &&
+ global_monitor_.Pointer()->NotifyStoreExcl_Locked(
+ address, &global_monitor_processor_)) {
+ switch (op) {
+ case STLXR_b:
+ MemoryWrite<uint8_t>(address, wreg(rt));
+ break;
+ case STLXR_h:
+ MemoryWrite<uint16_t>(address, wreg(rt));
+ break;
+ case STLXR_w:
+ MemoryWrite<uint32_t>(address, wreg(rt));
+ break;
+ default:
+ UNIMPLEMENTED();
+ }
+ LogWrite(address, access_size, rt);
+ set_wreg(rs, 0);
+ } else {
+ set_wreg(rs, 1);
+ }
+ }
}
void Simulator::CheckMemoryAccess(uintptr_t address, uintptr_t stack) {
@@ -3877,6 +3979,186 @@ void Simulator::DoPrintf(Instruction* instr) {
delete[] format;
}
+Simulator::LocalMonitor::LocalMonitor()
+ : access_state_(MonitorAccess::Open),
+ tagged_addr_(0),
+ size_(TransactionSize::None) {}
+
+void Simulator::LocalMonitor::Clear() {
+ access_state_ = MonitorAccess::Open;
+ tagged_addr_ = 0;
+ size_ = TransactionSize::None;
+}
+
+void Simulator::LocalMonitor::NotifyLoad(uintptr_t addr) {
+ if (access_state_ == MonitorAccess::Exclusive) {
+ // A non exclusive load could clear the local monitor. As a result, it's
+ // most strict to unconditionally clear the local monitor on load.
+ Clear();
+ }
+}
+
+void Simulator::LocalMonitor::NotifyLoadExcl(uintptr_t addr,
+ TransactionSize size) {
+ access_state_ = MonitorAccess::Exclusive;
+ tagged_addr_ = addr;
+ size_ = size;
+}
+
+void Simulator::LocalMonitor::NotifyStore(uintptr_t addr) {
+ if (access_state_ == MonitorAccess::Exclusive) {
+ // A non exclusive store could clear the local monitor. As a result, it's
+ // most strict to unconditionally clear the local monitor on store.
+ Clear();
+ }
+}
+
+bool Simulator::LocalMonitor::NotifyStoreExcl(uintptr_t addr,
+ TransactionSize size) {
+ if (access_state_ == MonitorAccess::Exclusive) {
+ // It is allowed for a processor to require that the address matches
+ // exactly (B2.10.1), so this comparison does not mask addr.
+ if (addr == tagged_addr_ && size_ == size) {
+ Clear();
+ return true;
+ } else {
+ // It is implementation-defined whether an exclusive store to a
+ // non-tagged address will update memory. As a result, it's most strict
+ // to unconditionally clear the local monitor.
+ Clear();
+ return false;
+ }
+ } else {
+ DCHECK(access_state_ == MonitorAccess::Open);
+ return false;
+ }
+}
+
+Simulator::GlobalMonitor::Processor::Processor()
+ : access_state_(MonitorAccess::Open),
+ tagged_addr_(0),
+ next_(nullptr),
+ prev_(nullptr),
+ failure_counter_(0) {}
+
+void Simulator::GlobalMonitor::Processor::Clear_Locked() {
+ access_state_ = MonitorAccess::Open;
+ tagged_addr_ = 0;
+}
+
+void Simulator::GlobalMonitor::Processor::NotifyLoadExcl_Locked(
+ uintptr_t addr) {
+ access_state_ = MonitorAccess::Exclusive;
+ tagged_addr_ = addr;
+}
+
+void Simulator::GlobalMonitor::Processor::NotifyStore_Locked(
+ uintptr_t addr, bool is_requesting_processor) {
+ if (access_state_ == MonitorAccess::Exclusive) {
+ // A non exclusive store could clear the global monitor. As a result, it's
+ // most strict to unconditionally clear global monitors on store.
+ Clear_Locked();
+ }
+}
+
+bool Simulator::GlobalMonitor::Processor::NotifyStoreExcl_Locked(
+ uintptr_t addr, bool is_requesting_processor) {
+ if (access_state_ == MonitorAccess::Exclusive) {
+ if (is_requesting_processor) {
+ // It is allowed for a processor to require that the address matches
+ // exactly (B2.10.2), so this comparison does not mask addr.
+ if (addr == tagged_addr_) {
+ Clear_Locked();
+ // Introduce occasional stxr failures. This is to simulate the
+ // behavior of hardware, which can randomly fail due to background
+ // cache evictions.
+ if (failure_counter_++ >= kMaxFailureCounter) {
+ failure_counter_ = 0;
+ return false;
+ } else {
+ return true;
+ }
+ }
+ } else if ((addr & kExclusiveTaggedAddrMask) ==
+ (tagged_addr_ & kExclusiveTaggedAddrMask)) {
+ // Check the masked addresses when responding to a successful lock by
+ // another processor so the implementation is more conservative (i.e. the
+ // granularity of locking is as large as possible.)
+ Clear_Locked();
+ return false;
+ }
+ }
+ return false;
+}
+
+Simulator::GlobalMonitor::GlobalMonitor() : head_(nullptr) {}
+
+void Simulator::GlobalMonitor::NotifyLoadExcl_Locked(uintptr_t addr,
+ Processor* processor) {
+ processor->NotifyLoadExcl_Locked(addr);
+ PrependProcessor_Locked(processor);
+}
+
+void Simulator::GlobalMonitor::NotifyStore_Locked(uintptr_t addr,
+ Processor* processor) {
+ // Notify each processor of the store operation.
+ for (Processor* iter = head_; iter; iter = iter->next_) {
+ bool is_requesting_processor = iter == processor;
+ iter->NotifyStore_Locked(addr, is_requesting_processor);
+ }
+}
+
+bool Simulator::GlobalMonitor::NotifyStoreExcl_Locked(uintptr_t addr,
+ Processor* processor) {
+ DCHECK(IsProcessorInLinkedList_Locked(processor));
+ if (processor->NotifyStoreExcl_Locked(addr, true)) {
+ // Notify the other processors that this StoreExcl succeeded.
+ for (Processor* iter = head_; iter; iter = iter->next_) {
+ if (iter != processor) {
+ iter->NotifyStoreExcl_Locked(addr, false);
+ }
+ }
+ return true;
+ } else {
+ return false;
+ }
+}
+
+bool Simulator::GlobalMonitor::IsProcessorInLinkedList_Locked(
+ Processor* processor) const {
+ return head_ == processor || processor->next_ || processor->prev_;
+}
+
+void Simulator::GlobalMonitor::PrependProcessor_Locked(Processor* processor) {
+ if (IsProcessorInLinkedList_Locked(processor)) {
+ return;
+ }
+
+ if (head_) {
+ head_->prev_ = processor;
+ }
+ processor->prev_ = nullptr;
+ processor->next_ = head_;
+ head_ = processor;
+}
+
+void Simulator::GlobalMonitor::RemoveProcessor(Processor* processor) {
+ base::LockGuard<base::Mutex> lock_guard(&mutex);
+ if (!IsProcessorInLinkedList_Locked(processor)) {
+ return;
+ }
+
+ if (processor->prev_) {
+ processor->prev_->next_ = processor->next_;
+ } else {
+ head_ = processor->next_;
+ }
+ if (processor->next_) {
+ processor->next_->prev_ = processor->prev_;
+ }
+ processor->prev_ = nullptr;
+ processor->next_ = nullptr;
+}
#endif // USE_SIMULATOR
« no previous file with comments | « src/arm64/simulator-arm64.h ('k') | test/cctest/BUILD.gn » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698